LLM Leaderboard - Comparison of GPT-4o, Llama 3, Mistral, Gemini and over 30 models
Comparison and ranking the performance of over 30 AI models (LLMs) across key metrics including quality, price, performance and speed (output speed - tokens per second & latency - TTFT), context window & others. For more details including relating to our methodology, see our FAQs.
HIGHLIGHTS






Features | Intelligence | Price | Output tokens/s | Latency | |||
---|---|---|---|---|---|---|---|
Further Analysis | |||||||
o4-mini (high) | 200k | 70 | $1.93 | 127.7 | 49.52 | ||
Gemini 2.5 Pro | 1m | 69 | $3.44 | 147.8 | 45.75 | ||
o3 | 128k | 67 | $17.50 | 166.6 | 17.18 | ||
Grok 3 mini Reasoning (high) | 1m | 67 | $0.35 | 60.0 | 0.31 | ||
o3-mini (high) | 200k | 66 | $1.93 | 171.2 | 39.91 | ||
Gemini 2.5 Flash (Reasoning) | 1m | 65 | $0.99 | 331.6 | 15.52 | ||
o3-mini | 200k | 63 | $1.93 | 166.6 | 14.08 | ||
Qwen3 235B (Reasoning) | 128k | 62 | $2.63 | 69.5 | 1.19 | ||
o1 | 200k | 62 | $26.25 | 131.1 | 24.42 | ||
Llama Nemotron Ultra Reasoning | 128k | 61 | $0.90 | 41.9 | 0.64 | ||
Claude 4 Sonnet Thinking | 200k | 61 | $6.00 | 72.6 | 1.50 | ||
Gemini 2.5 Flash (April '25) (Reasoning) | 1m | 60 | $0.99 | 381.9 | 7.79 | ||
DeepSeek R1 | ![]() | 128k | 60 | $2.68 | |||
o1-preview | 128k | 60 | $26.25 | 163.3 | 18.35 | ||
Qwen3 32B (Reasoning) | 128k | 59 | $2.63 | 63.1 | 1.14 | ||
QwQ-32B | 131k | 58 | $0.47 | 99.4 | 0.44 | ||
Claude 4 Opus | 200k | 58 | $30.00 | 53.7 | 2.55 | ||
Claude 3.7 Sonnet Thinking | 200k | 57 | $6.00 | 86.4 | 1.95 | ||
o1-pro | 200k | 56 | $262.50 | ||||
Grok 3 Reasoning Beta | 1m | 56 | $0.00 | ||||
Qwen3 14B (Reasoning) | 128k | 56 | $1.31 | 63.4 | 1.08 | ||
Qwen3 30B A3B (Reasoning) | 128k | 56 | $0.75 | 91.9 | 1.04 | ||
o1-mini | 128k | 54 | $1.93 | 211.9 | 9.89 | ||
Gemini 2.5 Flash | 1m | 53 | $0.26 | 262.4 | 0.26 | ||
DeepSeek V3 | ![]() | 128k | 53 | $0.48 | 25.4 | 3.61 | |
Claude 4 Sonnet | 200k | 53 | $6.00 | 75.7 | 1.64 | ||
GPT-4.5 (Preview) | 128k | 53 | $93.75 | 70.1 | 1.11 | ||
GPT-4.1 mini | 1m | 53 | $0.70 | 74.9 | 0.66 | ||
GPT-4.1 | 1m | 53 | $3.50 | 119.1 | 0.65 | ||
Gemini 2.0 Flash Thinking exp. (Jan '25) | 1m | 52 | $0.00 | ||||
DeepSeek R1 Distill Qwen 32B | ![]() | 128k | 52 | $0.22 | 41.5 | 0.65 | |
Qwen3 8B (Reasoning) | 128k | 51 | $0.66 | 94.0 | 1.02 | ||
Llama 3.3 Nemotron Super 49B Reasoning | 128k | 51 | $0.00 | ||||
Grok 3 | 1m | 51 | $6.00 | 75.3 | 0.43 | ||
Llama 4 Maverick | 1m | 51 | $0.37 | 165.0 | 0.33 | ||
GPT-4o (March 2025) | 128k | 50 | $7.50 | 183.6 | 0.51 | ||
Gemini 2.0 Pro Experimental | 2m | 49 | $0.00 | 68.5 | 17.54 | ||
DeepSeek R1 Distill Qwen 14B | ![]() | 128k | 49 | $0.88 | 103.5 | 0.56 | |
Mistral Medium 3 | ![]() | 128k | 49 | $0.80 | 81.9 | 0.42 | |
Sonar Reasoning | 127k | 49 | $2.00 | 75.2 | 1.71 | ||
Gemini 2.5 Flash | 1m | 49 | $0.26 | 304.5 | 0.36 | ||
DeepSeek R1 Distill Llama 70B | ![]() | 128k | 48 | $0.81 | 67.1 | 0.45 | |
Claude 3.7 Sonnet | 200k | 48 | $6.00 | 77.8 | 1.83 | ||
Gemini 2.0 Flash | 1m | 48 | $0.17 | 221.7 | 0.36 | ||
Qwen3 4B (Reasoning) | 32k | 47 | $0.40 | 100.3 | 1.02 | ||
Reka Flash 3 | ![]() | 128k | 47 | $0.35 | 56.3 | 0.91 | |
Qwen3 235B | 128k | 47 | $1.23 | 70.6 | 1.11 | ||
Gemini 2.0 Flash (exp) | 1m | 46 | $0.00 | 222.7 | 0.24 | ||
DeepSeek V3 (Dec '24) | ![]() | 128k | 46 | $0.48 | |||
Qwen2.5 Max | 32k | 45 | $2.80 | 40.9 | 1.26 | ||
Gemini 1.5 Pro (Sep) | 2m | 45 | $2.19 | 92.6 | 0.39 | ||
Claude 3.5 Sonnet (Oct) | 200k | 44 | $6.00 | 78.8 | 0.87 | ||
Qwen3 32B | 128k | 44 | $1.23 | 63.3 | 1.04 | ||
Sonar | 127k | 43 | $1.00 | 143.0 | 1.81 | ||
Llama 4 Scout | 10m | 43 | $0.27 | 123.2 | 0.36 | ||
Sonar Pro | 200k | 43 | $6.00 | 82.0 | 2.88 | ||
QwQ 32B-Preview | 33k | 43 | $0.17 | 65.4 | 0.44 | ||
Nova Premier | ![]() | 1m | 43 | $5.00 | 61.8 | 0.83 | |
Qwen3 30B A3B | 128k | 43 | $0.35 | 92.4 | 1.10 | ||
GPT-4o (Nov '24) | 128k | 41 | $4.38 | 110.4 | 0.57 | ||
Gemini 2.0 Flash-Lite (Feb '25) | 1m | 41 | $0.13 | 212.6 | 0.25 | ||
Llama 3.3 70B | 128k | 41 | $0.60 | 121.7 | 0.45 | ||
GPT-4.1 nano | 1m | 41 | $0.17 | 111.1 | 0.43 | ||
Qwen3 14B | 128k | 41 | $0.61 | 64.3 | 1.03 | ||
GPT-4o (May '24) | 128k | 41 | $7.50 | 93.3 | 0.64 | ||
Gemini 2.0 Flash-Lite (Preview) | 1m | 41 | $0.13 | 214.3 | 0.27 | ||
GPT-4o (Aug '24) | 128k | 41 | $4.38 | 82.3 | 0.70 | ||
Llama 3.1 405B | 128k | 40 | $3.50 | 32.8 | 0.77 | ||
Qwen2.5 72B | 131k | 40 | $0.00 | 58.0 | 1.25 | ||
MiniMax-Text-01 | ![]() | 4m | 40 | $0.42 | 33.1 | 0.82 | |
Phi-4 | ![]() | 16k | 40 | $0.22 | 38.5 | 0.43 | |
Claude 3.5 Sonnet (June) | 200k | 40 | $6.00 | 80.6 | 1.03 | ||
Command A | 256k | 40 | $4.38 | 94.8 | 0.22 | ||
Tulu3 405B | 128k | 40 | $0.00 | ||||
GPT-4o (ChatGPT) | 128k | 40 | $7.50 | ||||
Llama 3.3 Nemotron Super 49B v1 | 128k | 39 | $0.00 | ||||
Grok 2 | 131k | 39 | $0.00 | ||||
Gemini 1.5 Flash (Sep) | 1m | 39 | $0.13 | 185.7 | 0.18 | ||
GPT-4 Turbo | 128k | 39 | $15.00 | 46.6 | 0.84 | ||
Mistral Large 2 (Nov '24) | ![]() | 128k | 38 | $3.00 | 70.1 | 0.44 | |
Qwen3 1.7B (Reasoning) | 32k | 38 | $0.40 | 130.3 | 1.02 | ||
Gemma 3 27B | 128k | 38 | $0.00 | 48.8 | 0.61 | ||
Grok Beta | 128k | 38 | $7.50 | 66.7 | 0.28 | ||
Pixtral Large | ![]() | 128k | 37 | $3.00 | 80.9 | 0.38 | |
Qwen2.5 Instruct 32B | 128k | 37 | $0.15 | ||||
Llama 3.1 Nemotron 70B | 128k | 37 | $0.18 | 43.9 | 0.55 | ||
Nova Pro | ![]() | 300k | 37 | $1.40 | |||
Qwen3 8B | 128k | 37 | $0.31 | 95.1 | 0.96 | ||
Mistral Large 2 (Jul '24) | ![]() | 128k | 37 | $3.00 | 38.2 | 0.48 | |
Qwen2.5 Coder 32B | 131k | 36 | $0.15 | 51.6 | 0.36 | ||
GPT-4 | 8k | 36 | $37.50 | 25.5 | 0.77 | ||
GPT-4o mini | 128k | 36 | $0.26 | 74.8 | 0.55 | ||
Llama 3.1 70B | 128k | 35 | $0.72 | 58.1 | 0.47 | ||
Mistral Small 3.1 | ![]() | 128k | 35 | $0.15 | 97.5 | 0.27 | |
Mistral Small 3 | ![]() | 32k | 35 | $0.15 | 151.8 | 0.31 | |
DeepSeek-V2.5 (Dec '24) | ![]() | 128k | 35 | $0.17 | |||
Qwen3 4B | 32k | 35 | $0.19 | 102.6 | 1.02 | ||
Claude 3 Opus | 200k | 35 | $30.00 | 27.5 | 1.07 | ||
Claude 3.5 Haiku | 200k | 35 | $1.60 | 64.8 | 0.89 | ||
Gemini 2.0 Flash Thinking exp. (Dec '24) | 2m | 35 | $0.00 | ||||
DeepSeek-V2.5 | ![]() | 128k | 35 | $0.17 | |||
Snowflake Llama 3.3 70B | 128k | 34 | $0.00 | ||||
Devstral | ![]() | 256k | 34 | $0.15 | 128.6 | 0.30 | |
Mistral Saba | ![]() | 32k | 34 | $0.30 | 92.3 | 0.30 | |
DeepSeek R1 Distill Llama 8B | ![]() | 128k | 34 | $0.04 | 53.5 | 0.66 | |
Reka Core | ![]() | 128k | 34 | $2.00 | 27.4 | 0.85 | |
Gemma 3 12B | 128k | 34 | $0.06 | 20.5 | 0.99 | ||
Gemini 1.5 Pro (May) | 2m | 34 | $2.19 | 66.5 | 0.33 | ||
R1 1776 | 128k | 34 | $3.50 | ||||
Qwen2.5 Turbo | 1m | 34 | $0.09 | 107.3 | 1.05 | ||
Reka Flash | ![]() | 128k | 34 | $0.35 | 46.2 | 0.89 | |
Gemma 3 1B | 32k | 33 | $0.00 | ||||
Llama 3.2 90B (Vision) | 128k | 33 | $0.72 | 32.3 | 0.40 | ||
Solar Mini | ![]() | 4k | 33 | $0.15 | 85.0 | 1.04 | |
Reka Flash (Feb '24) | ![]() | 128k | 33 | $0.35 | 45.8 | 0.85 | |
Reka Edge | ![]() | 128k | 33 | $0.10 | 85.7 | 0.82 | |
Qwen2 72B | 131k | 33 | $0.00 | 30.9 | 1.35 | ||
Nova Lite | ![]() | 300k | 33 | $0.10 | 276.8 | 0.30 | |
Gemini 1.5 Flash-8B | 1m | 31 | $0.07 | 280.3 | 0.19 | ||
DeepHermes 3 - Mistral 24B | ![]() | 32k | 30 | $0.00 | |||
Jamba 1.5 Large | 256k | 29 | $3.50 | ||||
Hermes 3 - Llama-3.1 70B | ![]() | 128k | 29 | $0.00 | |||
DeepSeek-Coder-V2 | ![]() | 128k | 29 | $0.17 | |||
Jamba 1.6 Large | 256k | 29 | $3.50 | 52.2 | 0.62 | ||
Gemini 1.5 Flash (May) | 1m | 28 | $0.13 | 333.1 | 0.26 | ||
Nova Micro | ![]() | 130k | 28 | $0.06 | 311.9 | 0.29 | |
Yi-Large | ![]() | 32k | 28 | $3.00 | 65.9 | 0.36 | |
Claude 3 Sonnet | 200k | 28 | $6.00 | 60.5 | 0.77 | ||
Codestral (Jan '25) | ![]() | 256k | 28 | $0.45 | 112.4 | 0.29 | |
Llama 3 70B | 8k | 27 | $0.84 | 47.3 | 0.57 | ||
Mistral Small (Sep '24) | ![]() | 33k | 27 | $0.30 | 83.3 | 0.30 | |
Gemini 1.0 Ultra | 33k | 27 | $0.00 | ||||
Phi-4 Multimodal | ![]() | 128k | 27 | $0.00 | 17.1 | 0.35 | |
Qwen2.5 Coder 7B | 131k | 27 | $0.03 | 193.3 | 0.47 | ||
Mistral Large (Feb '24) | ![]() | 33k | 26 | $6.00 | 28.5 | 0.62 | |
Jamba Instruct | 256k | 26 | $0.00 | ||||
Mixtral 8x22B | ![]() | 65k | 26 | $3.00 | 62.4 | 0.32 | |
Phi-4 Mini | ![]() | 128k | 26 | $0.00 | 56.6 | 0.32 | |
Gemma 2 27B | 8k | 26 | $0.26 | ||||
Llama 3.2 11B (Vision) | 128k | 25 | $0.16 | 116.8 | 0.23 | ||
Qwen3 1.7B | 32k | 25 | $0.19 | 133.9 | 0.94 | ||
Qwen1.5 Chat 110B | 32k | 25 | $0.00 | 23.7 | 1.61 | ||
Phi-3 Medium 14B | ![]() | 128k | 25 | $0.30 | 53.2 | 0.40 | |
Gemma 3 4B | 128k | 24 | $0.03 | 88.8 | 0.26 | ||
Claude 2.1 | 200k | 24 | $12.00 | 14.0 | 0.89 | ||
Claude 3 Haiku | 200k | 24 | $0.50 | 145.8 | 0.86 | ||
Llama 3.1 8B | 128k | 24 | $0.10 | 215.1 | 0.30 | ||
Pixtral 12B | ![]() | 128k | 23 | $0.15 | 84.9 | 0.30 | |
Qwen3 0.6B (Reasoning) | 32k | 23 | $0.40 | 211.0 | 0.92 | ||
Claude 2.0 | 100k | 23 | $12.00 | 30.8 | 0.93 | ||
DeepSeek-V2 | ![]() | 128k | 23 | $0.17 | |||
Mistral Small (Feb '24) | ![]() | 33k | 23 | $1.50 | 139.5 | 0.28 | |
Mistral Medium | ![]() | 33k | 23 | $4.09 | 69.8 | 0.54 | |
GPT-3.5 Turbo | 4k | 23 | $0.75 | 150.2 | 0.39 | ||
Ministral 8B | ![]() | 128k | 22 | $0.10 | 128.5 | 0.29 | |
Gemma 2 9B | 8k | 22 | $0.04 | ||||
Phi-3 Mini | ![]() | 4k | 22 | $0.00 | |||
Arctic | 4k | 22 | $0.00 | ||||
Qwen Chat 72B | 34k | 22 | $1.00 | ||||
LFM 40B | 32k | 22 | $0.15 | 160.8 | 0.15 | ||
Command-R+ | 128k | 21 | $4.38 | 47.4 | 0.27 | ||
Llama 3 8B | 8k | 21 | $0.09 | 103.7 | 0.37 | ||
PALM-2 | 8k | 21 | $0.00 | ||||
Gemini 1.0 Pro | 33k | 21 | $0.75 | ||||
DeepSeek Coder V2 Lite | ![]() | 128k | 20 | $0.09 | 107.8 | 0.50 | |
Codestral (May '24) | ![]() | 33k | 20 | $0.30 | 113.7 | 0.32 | |
Aya Expanse 32B | 128k | 20 | $0.75 | 121.0 | 0.18 | ||
Llama 2 Chat 70B | 4k | 20 | $0.00 | ||||
DeepSeek LLM 67B (V1) | ![]() | 4k | 20 | $0.00 | |||
Llama 2 Chat 13B | 4k | 20 | $0.00 | ||||
Command-R+ (Apr '24) | 128k | 20 | $6.00 | 56.1 | 0.23 | ||
OpenChat 3.5 | ![]() | 8k | 20 | $0.05 | 50.6 | 0.49 | |
DBRX | 33k | 20 | $0.00 | ||||
Ministral 3B | ![]() | 128k | 20 | $0.04 | 225.2 | 0.26 | |
Mistral NeMo | ![]() | 128k | 20 | $0.15 | 140.7 | 0.29 | |
Llama 3.2 3B | 128k | 20 | $0.04 | 122.2 | 0.39 | ||
DeepSeek R1 Distill Qwen 1.5B | ![]() | 128k | 19 | $0.18 | 384.1 | 0.23 | |
Jamba 1.5 Mini | 256k | 18 | $0.25 | ||||
Jamba 1.6 Mini | 256k | 18 | $0.25 | 178.7 | 0.36 | ||
Mixtral 8x7B | ![]() | 33k | 17 | $0.70 | 81.9 | 0.29 | |
Qwen3 0.6B | 32k | 17 | $0.19 | 216.1 | 0.98 | ||
DeepHermes 3 - Llama-3.1 8B | ![]() | 128k | 16 | $0.00 | |||
Aya Expanse 8B | 8k | 16 | $0.75 | 165.1 | 0.14 | ||
Command-R | 128k | 15 | $0.26 | 58.7 | 0.20 | ||
Command-R (Mar '24) | 128k | 15 | $0.75 | 161.3 | 0.16 | ||
Qwen Chat 14B | 8k | 14 | $0.00 | ||||
Claude Instant | 100k | 14 | $1.20 | 62.9 | 0.52 | ||
Codestral-Mamba | ![]() | 256k | 14 | $0.25 | 94.3 | 0.42 | |
Llama 65B | 2k | 11 | $0.00 | ||||
Mistral 7B | ![]() | 8k | 10 | $0.25 | 104.3 | 0.29 | |
Llama 3.2 1B | 128k | 10 | $0.03 | 131.1 | 0.34 | ||
Llama 2 Chat 7B | 4k | 8 | $0.10 | 131.8 | 0.48 | ||
GPT-4o mini Realtime (Dec '24) | 128k | $0.00 | |||||
GPT-4o Realtime (Dec '24) | 128k | $0.00 | |||||
Claude 4 Opus Thinking | 200k | $30.00 | 55.6 | 3.04 | |||
Sonar Reasoning Pro | 127k | $0.00 | |||||
Grok 3 mini Reasoning (low) | 1m | $0.35 | 125.9 | 0.31 | |||
GPT-3.5 Turbo (0613) | 4k | $0.00 |
Key definitions
Models compared: OpenAI: GPT 4o Audio, GPT 4o Realtime, GPT 4o Speech Pipeline, GPT-3.5 Turbo, GPT-3.5 Turbo (0125), GPT-3.5 Turbo (0301), GPT-3.5 Turbo (0613), GPT-3.5 Turbo (1106), GPT-3.5 Turbo Instruct, GPT-4, GPT-4 Turbo, GPT-4 Turbo (0125), GPT-4 Turbo (1106), GPT-4 Vision, GPT-4.1, GPT-4.1 mini, GPT-4.1 nano, GPT-4.5 (Preview), GPT-4o (April 2025), GPT-4o (Aug '24), GPT-4o (ChatGPT), GPT-4o (March 2025), GPT-4o (May '24), GPT-4o (Nov '24), GPT-4o Realtime (Dec '24), GPT-4o mini, GPT-4o mini Realtime (Dec '24), o1, o1-mini, o1-preview, o1-pro, o3, o3-mini, o3-mini (high), and o4-mini (high), Meta: Code Llama 70B, Llama 2 Chat 13B, Llama 2 Chat 70B, Llama 2 Chat 7B, Llama 3 70B, Llama 3 8B, Llama 3.1 405B, Llama 3.1 70B, Llama 3.1 8B, Llama 3.2 11B (Vision), Llama 3.2 1B, Llama 3.2 3B, Llama 3.2 90B (Vision), Llama 3.3 70B, Llama 4 Behemoth, Llama 4 Maverick, Llama 4 Scout, and Llama 65B, Google: Gemini 1.0 Pro, Gemini 1.0 Ultra, Gemini 1.5 Flash (May), Gemini 1.5 Flash (Sep), Gemini 1.5 Flash-8B, Gemini 1.5 Pro (May), Gemini 1.5 Pro (Sep), Gemini 2.0 Flash, Gemini 2.0 Flash (exp), Gemini 2.0 Flash Thinking exp. (Dec '24), Gemini 2.0 Flash Thinking exp. (Jan '25), Gemini 2.0 Flash-Lite (Feb '25), Gemini 2.0 Flash-Lite (Preview), Gemini 2.0 Pro Experimental, Gemini 2.5 Flash, Gemini 2.5 Flash (Reasoning), Gemini 2.5 Flash (April '25) (Reasoning), Gemini 2.5 Pro, Gemini 2.5 Pro Preview (May' 25), Gemini Experimental (Nov), Gemma 2 27B, Gemma 2 9B, Gemma 3 12B, Gemma 3 1B, Gemma 3 27B, Gemma 3 4B, Gemma 3n E4B, Gemma 7B, and PALM-2, Anthropic: Claude 2.0, Claude 2.1, Claude 3 Haiku, Claude 3 Opus, Claude 3 Sonnet, Claude 3.5 Haiku, Claude 3.5 Sonnet (June), Claude 3.5 Sonnet (Oct), Claude 3.7 Sonnet Thinking, Claude 3.7 Sonnet, Claude 4 Opus, Claude 4 Opus Thinking, Claude 4 Sonnet, Claude 4 Sonnet Thinking, and Claude Instant, Mistral: Codestral (Jan '25), Codestral (May '24), Codestral-Mamba, Devstral, Ministral 3B, Ministral 8B, Mistral 7B, Mistral Large (Feb '24), Mistral Large 2 (Jul '24), Mistral Large 2 (Nov '24), Mistral Medium, Mistral Medium 3, Mistral NeMo, Mistral Saba, Mistral Small (Feb '24), Mistral Small (Sep '24), Mistral Small 3, Mistral Small 3.1, Mixtral 8x22B, Mixtral 8x7B, Pixtral 12B, and Pixtral Large, DeepSeek: DeepSeek Coder V2 Lite, DeepSeek LLM 67B (V1), DeepSeek Prover V2 671B, DeepSeek R1, DeepSeek R1 (FP4), DeepSeek R1 (May '25), DeepSeek R1 Distill Llama 70B, DeepSeek R1 Distill Llama 8B, DeepSeek R1 Distill Qwen 1.5B, DeepSeek R1 Distill Qwen 14B, DeepSeek R1 Distill Qwen 32B, DeepSeek V3 (Dec '24), DeepSeek V3, DeepSeek-Coder-V2, DeepSeek-V2, DeepSeek-V2.5, DeepSeek-V2.5 (Dec '24), DeepSeek-VL2, and Janus Pro 7B, Perplexity: PPLX-70B Online, PPLX-7B-Online, R1 1776, Sonar, Sonar 3.1 Huge, Sonar 3.1 Large, Sonar 3.1 Small , Sonar Large, Sonar Pro, Sonar Reasoning, Sonar Reasoning Pro, and Sonar Small, xAI: Grok 2, Grok 3, Grok 3 Reasoning Beta, Grok 3 mini, Grok 3 mini Reasoning (low), Grok 3 mini Reasoning (high), Grok Beta, and Grok-1, OpenChat: OpenChat 3.5, Amazon: Nova Lite, Nova Micro, Nova Premier, and Nova Pro, Microsoft Azure: Phi-3 Medium 14B, Phi-3 Mini, Phi-4, Phi-4 Mini, Phi-4 Multimodal, Phi-4 mini reasoning, Phi-4 reasoning, and Phi-4 reasoning plus, Liquid AI: LFM 1.3B, LFM 3B, and LFM 40B, Upstage: Solar Mini, Solar Pro, and Solar Pro (Nov '24), Databricks: DBRX, MiniMax: MiniMax-Text-01, NVIDIA: Cosmos Nemotron 34B, Llama 3.1 Nemotron 70B, Llama 3.1 Nemotron Nano 8B, Llama 3.3 Nemotron Nano 8B v1 (Reasoning), Llama Nemotron Ultra Reasoning, Llama 3.3 Nemotron Super 49B v1, and Llama 3.3 Nemotron Super 49B Reasoning, IBM: Granite 3.0 2B, OpenVoice: Granite 3.0 8B, Inceptionlabs: Mercury Coder Mini, Mercury Coder Small, Mercury Instruct, and Mercury Small, Reka AI: Reka Core, Reka Edge, Reka Flash (Feb '24), Reka Flash, and Reka Flash 3, Xiaomi: MiMo 7B RL, Baichuan: Baichuan 4 and Baichuan M1 (Preview), vercel: v0-1.0-md, Other: LLaVA-v1.5-7B, Cohere: Aya Expanse 32B, Aya Expanse 8B, Command, Command A, Command Light, Command R7B, Command-R, Command-R (Mar '24), Command-R+ (Apr '24), and Command-R+, Bytedance: Duobao 1.5 Pro, Seed-Thinking-v1.5, Skylark Lite, and Skylark Pro, AI21 Labs: Jamba 1.5 Large, Jamba 1.5 Large (Feb '25), Jamba 1.5 Mini, Jamba 1.5 Mini (Feb 2025), Jamba 1.6 Large, Jamba 1.6 Mini, and Jamba Instruct, Snowflake: Arctic and Snowflake Llama 3.3 70B, Alibaba: QwQ-32B, QwQ 32B-Preview, Qwen Chat 14B, Qwen Chat 72B, Qwen Chat 7B, Qwen1.5 Chat 110B, Qwen1.5 Chat 14B, Qwen1.5 Chat 32B, Qwen1.5 Chat 72B, Qwen1.5 Chat 7B, Qwen2 72B, Qwen2 Instruct 7B, Qwen2 Instruct A14B 57B, Qwen2-VL 72B, Qwen2.5 Coder 32B, Qwen2.5 Coder 7B , Qwen2.5 Instruct 14B, Qwen2.5 Instruct 32B, Qwen2.5 72B, Qwen2.5 Instruct 7B, Qwen2.5 Max, Qwen2.5 Max 01-29, Qwen2.5 Omni 7B, Qwen2.5 Plus, Qwen2.5 Turbo, Qwen2.5 VL 72B, Qwen2.5 VL 7B, Qwen3 0.6B, Qwen3 0.6B (Reasoning), Qwen3 1.7B, Qwen3 1.7B (Reasoning), Qwen3 14B, Qwen3 14B (Reasoning), Qwen3 235B, Qwen3 235B (Reasoning), Qwen3 30B A3B, Qwen3 30B A3B (Reasoning), Qwen3 32B, Qwen3 32B (Reasoning), Qwen3 4B, Qwen3 4B (Reasoning), Qwen3 8B, and Qwen3 8B (Reasoning), and 01.AI: Yi-Large and Yi-Lightning.