OpenAI: gpt-oss-120b (free)Model25/100 via “mixture-of-experts reasoning and task decomposition”
gpt-oss-120b is an open-weight, 117B-parameter Mixture-of-Experts (MoE) language model from OpenAI designed for high-reasoning, agentic, and general-purpose production use cases. It activates 5.1B parameters per forward pass and is optimized...
Unique: Uses a learned Mixture-of-Experts routing mechanism with 5.1B active parameters per token, enabling 40-50% inference cost reduction compared to dense 120B models while maintaining reasoning quality through specialized expert networks trained for different reasoning modalities
vs others: More efficient than dense models like GPT-4 (175B+) for reasoning tasks due to sparse activation, while maintaining comparable reasoning depth through expert specialization; faster inference than dense 120B alternatives with similar capability