
gpt-oss-120b is an open-weight, 117B-parameter Mixture-of-Experts (MoE) model from OpenAI, built for advanced reasoning, agentic behavior, and versatile production use cases. It activates 5.1B parameters per forward pass and is optimized to run efficiently on a single H100 GPU with native MXFP4 quantization. The model supports configurable reasoning depth, full chain-of-thought access, and native tool capabilities such as function calling, web browsing, and structured output generation.
| Creator | OpenAI |
| Release Date | August, 2025 |
| License | Apache 2.0 |
| Context Window | 131,000 |
| Image Input Support | No |
| Open Source (Weights) | Yes |
| Parameters | 117B, 5.1B active at inference time |
| Model Weights | Click here |

