About 25 results
Open links in new tab
  1. How to run gpt-oss with Transformers

    Aug 5, 2025 · This guide will walk you through running OpenAI gpt-oss-20b or OpenAI gpt-oss-120b using Transformers, either with a high-level pipeline or via low-level generate calls with raw token IDs.

  2. Fine-tuning with gpt-oss and Hugging Face Transformers

    Next, install the remaining dependencies: %pip install "trl>=0.20.0" "peft>=0.17.0" "transformers>=4.55.0" trackio

  3. Introducing gpt-oss | OpenAI

    Aug 5, 2025 · We’re releasing gpt-oss-120b and gpt-oss-20b—two state-of-the-art open-weight language models that deliver strong real-world performance at low cost. Available under the flexible …

  4. User guide for gpt-oss-safeguard - developers.openai.com

    Oct 29, 2025 · This guide takes you through running OpenAI gpt-oss models using Transformers, either with a high-level pipeline or via low-level generate calls with raw token IDs.

  5. How to run gpt-oss-20b on Google Colab

    Aug 6, 2025 · Since support for mxfp4 in transformers is bleeding edge, we need a recent version of PyTorch and CUDA, in order to be able to install the mxfp4 triton kernels.

  6. gpt-oss | OpenAI Developers

    Fine-tuning with gpt-oss and Hugging Face Transformers Authored by: Edward Beeching, Quentin Gallouédec, and Lewis Tunstall Large reasoning models like OpenAI o3 generate a chain-of-thought …

  7. Transformers.js + WebGPU: Run a local LLM in your browser ...

    Dec 23, 2025 · <!doctype html> <html lang="en"> <head> <meta charset="utf-8" /> <meta name="viewport" content="width=device-width, initial-scale=1" /> <title>Transformers.js Browser …