Comparisons of Large Open Source AI Models (>150B)
Open source AI models with over 150B parameters. Models are considered Open Source (also commonly referred to as open weights) where their weights are accessible to download. This allows self-hosting on your own infrastructure and enables customizing the model such as through fine-tuning. Click on any model to see detailed metrics. For more details including relating to our methodology, see our FAQs.
GLM-5.1 and
GLM-5 are the highest intelligence Large open source models, defined as those with >150B parameters, followed by
Kimi K2.5 & Qwen3.5 397B A17B.
Intelligence
Artificial Analysis Intelligence Index; Higher is better
Total Parameters
Trainable parameters in billions
Navigation
Openness
Artificial Analysis Openness Index: Results
Openness Index assesses model openness on a 0 to 100 normalized scale (higher is more open)
Intelligence
Artificial Analysis Intelligence Index
Artificial Analysis Intelligence Index v4.0 incorporates 10 evaluations: GDPval-AA, 𝜏²-Bench Telecom, Terminal-Bench Hard, SciCode, AA-LCR, AA-Omniscience, IFBench, Humanity's Last Exam, GPQA Diamond, CritPt
Reasoning models are indicated by a lightbulb icon.
Intelligence Evaluations
Intelligence evaluations measured independently by Artificial Analysis; Higher is better
Results claimed by AI Lab (not yet independently verified)
GDPval-AA
Terminal-Bench Hard
𝜏²-Bench Telecom
AA-LCR
AA-Omniscience Accuracy
AA-Omniscience Non-Hallucination Rate
Humanity's Last Exam
GPQA Diamond
SciCode
IFBench
CritPt
APEX-Agents-AA
No data available
MMMU-Pro
Reasoning models are indicated by a lightbulb icon.
Size
Model Size: Total and Active Parameters
Comparison between total model parameters and parameters active during inference
Active Parameters
Passive Parameters
Reasoning models are indicated by a lightbulb icon.
Intelligence vs. Active Parameters
Active Parameters at Inference Time; Artificial Analysis Intelligence Index
Most attractive quadrant
Alibaba
Arcee AI
DeepSeek
Kimi
LG AI Research
Meta
StepFun
Xiaomi
Z AI
Reasoning models are indicated by a lightbulb icon.
Intelligence vs. Total Parameters
Artificial Analysis Intelligence Index; Size in Parameters (Billions)
Most attractive quadrant
Alibaba
Arcee AI
DeepSeek
Kimi
LG AI Research
Meta
StepFun
Xiaomi
Z AI
Reasoning models are indicated by a lightbulb icon.
Context Window
Context Window
Context Window: Tokens Limit; Higher is better
Reasoning models are indicated by a lightbulb icon.
Further details
| Weights | Provider Benchmarks | |||||||
|---|---|---|---|---|---|---|---|---|
GLM-5.1 (Reasoning) Z AI | 51 | 754B | 200k | $2.1 | 70 | 🤗 | +4 more | View |
GLM-5 (Reasoning) Z AI | 50 | 744B (40B active at inference time) | 200k | $1.6 | 75 | 🤗 | +11 more | View |
Kimi K2.5 (Reasoning) Kimi | 47 | 1.0KB (32B active at inference time) | 256k | $1.2 | 36 | 🤗 | +14 more | View |
Qwen3.5 397B A17B (Reasoning) Alibaba | 45 | 397B (17B active at inference time) | 262k | $1.4 | 89 | 🤗 | +6 more | View |
DeepSeek V3.2 (Reasoning) DeepSeek | 42 | 685B (37B active at inference time) | 128k | $0.3 | 47 | 🤗 | +8 more | View |
MiMo-V2-Flash (Feb 2026) Xiaomi | 41 | 309B (15B active at inference time) | 256k | $0.1 | 130 | 🤗 | View | |
GLM-5 (Non-reasoning) Z AI | 41 | 744B (40B active at inference time) | 200k | $1.6 | 48 | 🤗 | +4 more | View |
Qwen3.5 397B A17B (Non-reasoning) Alibaba | 40 | 397B (17B active at inference time) | 262k | $1.4 | 85 | 🤗 | +3 more | View |
Step 3.5 Flash StepFun | 38 | 196B (11B active at inference time) | 256k | $0.1 | 90 | 🤗 | View | |
Kimi K2.5 (Non-reasoning) Kimi | 37 | 1.0KB (32B active at inference time) | 256k | $1.2 | 34 | 🤗 | +6 more | View |
K-EXAONE (Reasoning) LG AI Research | 32 | 236B (23B active at inference time) | 256k | - | - | 🤗 | - | View |
DeepSeek V3.2 (Non-reasoning) DeepSeek | 32 | 685B (37B active at inference time) | 128k | $0.3 | 46 | 🤗 | +11 more | View |
Trinity Large Thinking Arcee AI | 32 | 399B (13B active at inference time) | 512k | $0.4 | 101 | 🤗 | View | |
MiMo-V2-Flash (Non-reasoning) Xiaomi | 30 | 309B (15B active at inference time) | 256k | $0.1 | 130 | 🤗 | View | |
DeepSeek V3.2 Speciale DeepSeek | 29 | 685B (37B active at inference time) | 128k | - | - | 🤗 | - | View |
DeepSeek R1 0528 (May '25) DeepSeek | 27 | 685B (37B active at inference time) | 128k | $2.4 | - | 🤗 | +6 more | View |
Qwen3 Coder 480B A35B Instruct Alibaba | 25 | 480B (35B active at inference time) | 262k | $3.0 | 61 | 🤗 | +8 more | View |
K-EXAONE (Non-reasoning) LG AI Research | 23 | 236B (23B active at inference time) | 256k | - | - | 🤗 | - | View |
Mistral Large 3 Mistral | 23 | 675B (41B active at inference time) | 256k | $0.8 | 40 | 🤗 | View | |
Ring-1T InclusionAI | 23 | 1.0KB (50B active at inference time) | 128k | - | - | 🤗 | - | View |
Ling-1T InclusionAI | 19 | 1.0KB (50B active at inference time) | 128k | - | - | 🤗 | - | View |
Hermes 4 - Llama-3.1 405B (Reasoning) Nous Research | 19 | 406B | 128k | $1.5 | 32 | 🤗 | View | |
Llama 4 Maverick Meta | 18 | 402B (17B active at inference time) | 1.00M | $0.5 | 111 | 🤗 | +9 more | View |
Hermes 4 - Llama-3.1 405B (Non-reasoning) Nous Research | 18 | 406B | 128k | $1.5 | 33 | 🤗 | View | |
Llama 3.1 Instruct 405B Meta | 17 | 405B | 128k | $3.7 | 29 | 🤗 | +1 more | View |
Llama 3.1 Nemotron Ultra 253B v1 (Reasoning) NVIDIA | 15 | 253B | 128k | $0.9 | 41 | 🤗 | View | |
ERNIE 4.5 300B A47B Baidu | 15 | 300B (47B active at inference time) | 131k | $0.5 | 24 | 🤗 | View | |
R1 1776 Perplexity | 12 | 671B (37B active at inference time) | 128k | - | - | 🤗 | - | View |
Jamba 1.7 Large AI21 Labs | 11 | 398B (94B active at inference time) | 256k | $3.5 | 58 | 🤗 | View | |
Cogito v2.1 (Reasoning) Deep Cogito | - | 671B (37B active at inference time) | 128k | $1.3 | 89 | 🤗 | View |