Comparisons of Medium Open Source AI Models (40B-150B)
Open source AI models with between 40 to 150 billion parameters. Models are considered Open Source (also commonly referred to as open weights) where their weights are accessible to download. This allows self-hosting on your own infrastructure and enables customizing the model such as through fine-tuning. Click on any model to see detailed metrics. For more details including relating to our methodology, see our FAQs.
Llama 3.3 Nemotron Super 49B Reasoning and
DeepSeek R1 Distill Llama 70B are the highest intelligence Medium open source models, defined as those with 40B-150B parameters, followed by
Llama 4 Scout &
Llama 3.3 70B.
Highlights
Intelligence
Artificial Analysis Intelligence Index; Higher is better
Total Parameters
Trainable parameters in billions
Further details
Weights | Provider Benchmarks | |||||||
---|---|---|---|---|---|---|---|---|
Llama 3.3 Nemotron Super 49B v1 (Reasoning) NVIDIA | 51 | 49B | 128k | - | - | 🤗 | - | View |
DeepSeek R1 Distill Llama 70B DeepSeek | 48 | 70B | 128k | $0.8 | 65 | 🤗 | ![]() +6 more | View |
Llama 4 Scout Meta | 43 | 109B (17B active at inference time) | 10.0M | $0.3 | 122 | 🤗 | ![]() ![]() +12 more | View |
Llama 3.3 Instruct 70B Meta | 41 | 70B | 128k | $0.6 | 110 | 🤗 | ![]() ![]() +19 more | View |
Qwen2.5 Instruct 72B Alibaba | 40 | 72B | 131k | - | 58 | 🤗 | ![]() ![]() ![]() +4 more | View |
![]() Command A Cohere | 40 | 111B | 256k | $4.4 | 167 | 🤗 | ![]() | View |
Llama 3.3 Nemotron Super 49B v1 NVIDIA | 39 | 49B | 128k | - | - | 🤗 | - | View |
![]() Mistral Large 2 (Nov '24) Mistral | 38 | 123B | 128k | $3.0 | 87 | 🤗 | ![]() | View |
![]() Pixtral Large Mistral | 37 | 124B | 128k | $3.0 | 91 | 🤗 | ![]() | View |
Llama 3.1 Nemotron Instruct 70B NVIDIA | 37 | 70B | 128k | $0.2 | 41 | 🤗 | ![]() | View |
![]() Mistral Large 2 (Jul '24) Mistral | 37 | 123B | 128k | $3.0 | 105 | 🤗 | ![]() | View |
Llama 3.1 Instruct 70B Meta | 35 | 70B | 128k | $0.8 | 64 | 🤗 | ![]() ![]() +10 more | View |
Llama 3.2 Instruct 90B (Vision) Meta | 33 | 90B | 128k | $0.5 | 38 | 🤗 | ![]() | View |
Qwen2 Instruct 72B Alibaba | 33 | 72B | 131k | - | 31 | 🤗 | View | |
![]() Hermes 3 - Llama-3.1 70B Nous Research | 29 | 70.6B | 128k | - | - | 🤗 | - | View |
Llama 3 Instruct 70B Meta | 27 | 70B | 8.19k | $0.8 | 47 | 🤗 | +7 more | View |
![]() Mixtral 8x22B Instruct Mistral | 26 | 141B (39B active at inference time) | 65.4k | $3.0 | 55 | 🤗 | ![]() ![]() | View |
Qwen1.5 Chat 110B Alibaba | 25 | 110B | 32.0k | - | 24 | 🤗 | View | |
Qwen Chat 72B Alibaba | 22 | 72B | 33.8k | $1.0 | - | 🤗 | View | |
![]() Command-R+ (Aug '24) Cohere | 21 | 104B | 128k | $4.4 | 51 | 🤗 | ![]() | View |
Llama 2 Chat 70B Meta | 20 | 70B | 4.10k | - | - | 🤗 | - | View |
![]() Command-R+ (Apr '24) Cohere | 20 | 104B | 128k | $6.0 | 60 | 🤗 | ![]() | View |
![]() DBRX Instruct Databricks | 20 | 132B (36B active at inference time) | 32.8k | - | - | 🤗 | - | View |
![]() Jamba 1.5 Mini AI21 Labs | 18 | 52B (12B active at inference time) | 256k | $0.3 | - | 🤗 | View | |
![]() Jamba 1.6 Mini AI21 Labs | 18 | 52B (12B active at inference time) | 256k | $0.3 | 187 | 🤗 | ![]() | View |
![]() Mixtral 8x7B Instruct Mistral | 17 | 46.7B (12.9B active at inference time) | 32.8k | $0.7 | 84 | 🤗 | ![]() ![]() +2 more | View |