Comparisons of Medium Open Source AI Models (40B-150B)
Open source AI models with between 40B to 150B parameters. Models are considered Open Source (also commonly referred to as open weights) where their weights are accessible to download. This allows self-hosting on your own infrastructure and enables customizing the model such as through fine-tuning. Click on any model to see detailed metrics. For more details including relating to our methodology, see our FAQs.
Qwen3.5 122B A10B and
NVIDIA Nemotron 3 Super are the highest intelligence Medium open source models, defined as those with 40B-150B parameters, followed by
Qwen3.5 122B A10B &
gpt-oss-120B (high).
Intelligence
Artificial Analysis Intelligence Index; Higher is better
Total Parameters
Trainable parameters in billions
Navigation
Openness
Artificial Analysis Openness Index: Results
Openness Index assesses model openness on a 0 to 100 normalized scale (higher is more open)
Intelligence
Artificial Analysis Intelligence Index
Artificial Analysis Intelligence Index v4.0 incorporates 10 evaluations: GDPval-AA, 𝜏²-Bench Telecom, Terminal-Bench Hard, SciCode, AA-LCR, AA-Omniscience, IFBench, Humanity's Last Exam, GPQA Diamond, CritPt
Reasoning models are indicated by a lightbulb icon.
Intelligence Evaluations
Intelligence evaluations measured independently by Artificial Analysis; Higher is better
Results claimed by AI Lab (not yet independently verified)
GDPval-AA (Agentic Real-World Work Tasks, (ELO-500)/2000)
Terminal-Bench Hard (Agentic Coding & Terminal Use)
𝜏²-Bench Telecom (Agentic Tool Use)
AA-LCR (Long Context Reasoning)
AA-Omniscience Accuracy (Knowledge)
AA-Omniscience Non-Hallucination Rate (1 - Hallucination Rate)
Humanity's Last Exam (Reasoning & Knowledge)
GPQA Diamond (Scientific Reasoning)
SciCode (Coding)
IFBench (Instruction Following)
CritPt (Physics Reasoning)
MMMU Pro (Visual Reasoning)
Reasoning models are indicated by a lightbulb icon.
Size
Model Size: Total and Active Parameters
Comparison between total model parameters and parameters active during inference
Active Parameters
Passive Parameters
Reasoning models are indicated by a lightbulb icon.
Intelligence vs. Active Parameters
Active Parameters at Inference Time; Artificial Analysis Intelligence Index
Most attractive quadrant
Alibaba
LongCat
MBZUAI Institute of Foundation Models
Meta
Mistral
NVIDIA
OpenAI
Prime Intellect
Z AI
Reasoning models are indicated by a lightbulb icon.
Intelligence vs. Total Parameters
Artificial Analysis Intelligence Index; Size in Parameters (Billions)
Most attractive quadrant
Alibaba
LongCat
MBZUAI Institute of Foundation Models
Meta
Mistral
NVIDIA
OpenAI
Prime Intellect
Z AI
Reasoning models are indicated by a lightbulb icon.
Context Window
Context Window
Context Window: Tokens Limit; Higher is better
Reasoning models are indicated by a lightbulb icon.
Further details
| Weights | Provider Benchmarks | |||||||
|---|---|---|---|---|---|---|---|---|
Qwen3.5 122B A10B (Reasoning) Alibaba | 42 | 125B (10B active at inference time) | 262k | $1.1 | 132 | 🤗 | View | |
NVIDIA Nemotron 3 Super 120B A12B (Reasoning) NVIDIA | 36 | 120.6B (12.7B active at inference time) | 1.00M | $0.4 | 458 | Not available | +2 more | View |
Qwen3.5 122B A10B (Non-reasoning) Alibaba | 36 | 125B (10B active at inference time) | 262k | $1.1 | 126 | 🤗 | View | |
gpt-oss-120B (high) OpenAI | 33 | 117B (5.1B active at inference time) | 131k | $0.3 | 280 | 🤗 | +22 more | View |
Qwen3 Coder Next Alibaba | 28 | 79.7B (3B active at inference time) | 256k | $0.6 | 137 | 🤗 | +1 more | View |
Mistral Small 4 (Reasoning) Mistral | 27 | 119B (6.5B active at inference time) | 256k | $0.3 | 153 | 🤗 | View | |
Qwen3 Next 80B A3B (Reasoning) Alibaba | 27 | 80B (3B active at inference time) | 262k | $1.9 | 155 | 🤗 | +4 more | View |
gpt-oss-120B (low) OpenAI | 24 | 117B (5.1B active at inference time) | 131k | $0.3 | 288 | 🤗 | +18 more | View |
K2 Think V2 MBZUAI Institute of Foundation Models | 24 | 70B | 262k | - | - | Not available | - | View |
LongCat Flash Lite LongCat | 24 | 68.5B (3B active at inference time) | 256k | - | 101 | 🤗 | View | |
GLM-4.6V (Reasoning) Z AI | 23 | 108B | 128k | $0.5 | 31 | 🤗 | View | |
INTELLECT-3 Prime Intellect | 22 | 107B | 131k | - | - | 🤗 | - | View |
Devstral 2 Mistral | 22 | 125B | 256k | - | 82 | 🤗 | View | |
K2-V2 (high) MBZUAI Institute of Foundation Models | 21 | 70B | 512k | - | - | 🤗 | - | View |
Qwen3 Next 80B A3B Instruct Alibaba | 20 | 80B (3B active at inference time) | 262k | $0.9 | 149 | 🤗 | +4 more | View |
Llama Nemotron Super 49B v1.5 (Reasoning) NVIDIA | 19 | 49B | 128k | $0.2 | 81 | 🤗 | View | |
K2-V2 (medium) MBZUAI Institute of Foundation Models | 19 | 70B | 512k | - | - | 🤗 | - | View |
Mistral Small 4 (Non-reasoning) Mistral | 19 | 119B (6.5B active at inference time) | 256k | $0.3 | 130 | 🤗 | View | |
Llama 3.3 Nemotron Super 49B v1 (Reasoning) NVIDIA | 18 | 49B | 128k | - | - | 🤗 | - | View |
Sarvam 105B (Reasoning) Sarvam | 18 | 106B (10.3B active at inference time) | 65.5k | - | 78 | 🤗 | View | |
GLM-4.6V (Non-reasoning) Z AI | 17 | 108B | 128k | $0.5 | 21 | 🤗 | View | |
Hermes 4 - Llama-3.1 70B (Reasoning) Nous Research | 16 | 70.6B | 128k | $0.2 | 77 | 🤗 | View | |
DeepSeek R1 Distill Llama 70B DeepSeek | 16 | 70B | 128k | $0.9 | 53 | 🤗 | View | |
Ling-flash-2.0 InclusionAI | 16 | 103B (6.1B active at inference time) | 128k | $0.2 | 58 | 🤗 | View | |
Llama Nemotron Super 49B v1.5 (Non-reasoning) NVIDIA | 15 | 49B | 128k | $0.2 | 81 | 🤗 | View | |
Llama 3.3 Instruct 70B Meta | 14 | 70B | 128k | $0.7 | 81 | 🤗 | +19 more | View |
K2-V2 (low) MBZUAI Institute of Foundation Models | 14 | 70B | 512k | - | - | 🤗 | - | View |
Kimi Linear 48B A3B Instruct Kimi | 14 | 49.1B (3B active at inference time) | 1.00M | - | - | 🤗 | - | View |
Llama 3.3 Nemotron Super 49B v1 (Non-reasoning) NVIDIA | 14 | 49B | 128k | - | - | 🤗 | - | View |
Ring-flash-2.0 InclusionAI | 14 | 103B (6.1B active at inference time) | 128k | $0.2 | 78 | 🤗 | View | |
Llama 4 Scout Meta | 14 | 109B (17B active at inference time) | 10.0M | $0.3 | 127 | 🤗 | +7 more | View |
Command A Cohere | 13 | 111B | 256k | $4.4 | 46 | 🤗 | View | |
Llama 3.1 Nemotron Instruct 70B NVIDIA | 13 | 70B | 128k | $1.2 | 36 | 🤗 | View | |
Hermes 4 - Llama-3.1 70B (Non-reasoning) Nous Research | 13 | 70.6B | 128k | $0.2 | 79 | 🤗 | View | |
Llama 3.2 Instruct 90B (Vision) Meta | 12 | 90B | 128k | $0.7 | 56 | 🤗 | +1 more | View |
Jamba 1.7 Mini AI21 Labs | 8 | 52B (12B active at inference time) | 258k | - | - | 🤗 | - | View |