DeepSeek V3 0324 vs. Gemini 1.5 Flash-8B
Comparison between DeepSeek V3 0324 and Gemini 1.5 Flash-8B across intelligence, price, speed, context window and more.
For details relating to our methodology, see our Methodology page.
Model Comparison
| Metric | Analysis | ||
|---|---|---|---|
| Creator | |||
| Context Window | 128k tokens (~192 A4 pages of size 12 Arial font) | 1049k tokens (~1573 A4 pages of size 12 Arial font) | DeepSeek V3 0324 is smaller than Gemini 1.5 Flash-8B |
| Release Date | March, 2025 | October, 2024 | DeepSeek V3 0324 has a more recent release date than Gemini 1.5 Flash-8B |
| Parameters | 671B, 37B active at inference time | 8B | DeepSeek V3 0324 is larger than Gemini 1.5 Flash-8B |
| Image Input Support | No | Yes | Gemini 1.5 Flash-8B has image input support while DeepSeek V3 0324 does not |
| Open Source (Weights) | No | DeepSeek V3 0324 is open source while Gemini 1.5 Flash-8B is proprietary |
Intelligence
Artificial Analysis Intelligence Index
Artificial Analysis Intelligence Index by Open Weights / Proprietary
Intelligence Evaluations
Openness
Artificial Analysis Openness Index: Results
Intelligence Index Comparisons
Intelligence vs. Price
Intelligence Index Token Use & Cost
Output Tokens Used to Run Artificial Analysis Intelligence Index
Cost to Run Artificial Analysis Intelligence Index
Context Window
Context Window
Pricing
Pricing: Input and Output Prices
Intelligence vs. Price (Log Scale)
Speed
Measured by Output Speed (tokens per second)
Output Speed
Output Speed vs. Price
Latency
Measured by Time (seconds) to First Token
Latency: Time To First Answer Token
End-to-End Response Time
Seconds to output 500 Tokens, calculated based on time to first token, 'thinking' time for reasoning models, and output speed
End-to-End Response Time
Model Size (Open Weights Models Only)
Model Size: Total and Active Parameters
Comparisons to DeepSeek V3 0324
DeepSeek V3 0324
gpt-oss-120B (high)
GPT-5.4 (xhigh)
gpt-oss-20B (high)
GPT-5.4 Pro (xhigh)
GPT-5.2 (xhigh)
GPT-5.3 Codex (xhigh)
Llama 4 Maverick
Gemini 3.1 Pro Preview
Gemini 3 Flash
Gemini 3.1 Flash-Lite Preview
Claude 4.5 Haiku
Claude Sonnet 4.6 (max)
Claude Opus 4.6 (max)
Mistral Large 3DeepSeek V3.2
Grok 4.20 Beta 0309
Grok 4.1 Fast
Nova 2.0 Pro Preview (medium)
MiniMax-M2.5
NVIDIA Nemotron 3 Nano
NVIDIA Nemotron 3 Super
Kimi K2.5
K-EXAONEMiMo-V2-Flash (Feb 2026)
K2 Think V2
Mi:dm K 2.5 ProGLM-5
Qwen3.5 397B A17B
Comparisons to Gemini 1.5 Flash-8B
Gemini 1.5 Flash-8B
gpt-oss-120B (high)
GPT-5.4 (xhigh)
gpt-oss-20B (high)
GPT-5.4 Pro (xhigh)
GPT-5.2 (xhigh)
GPT-5.3 Codex (xhigh)
Llama 4 Maverick
Gemini 3.1 Pro Preview
Gemini 3 Flash
Gemini 3.1 Flash-Lite Preview
Claude 4.5 Haiku
Claude Sonnet 4.6 (max)
Claude Opus 4.6 (max)
Mistral Large 3DeepSeek V3.2
Grok 4.20 Beta 0309
Grok 4.1 Fast
Nova 2.0 Pro Preview (medium)
MiniMax-M2.5
NVIDIA Nemotron 3 Nano
NVIDIA Nemotron 3 Super
Kimi K2.5
K-EXAONEMiMo-V2-Flash (Feb 2026)
K2 Think V2
Mi:dm K 2.5 ProGLM-5
Qwen3.5 397B A17B
Frequently Asked Questions
Gemini 3.1 Pro Preview currently leads the Artificial Analysis Intelligence Index with a score of 57, out of 295 models evaluated.
The top AI models by Intelligence Index are: 1. Gemini 3.1 Pro Preview (57), 2. GPT-5.4 (xhigh) (57), 3. GPT-5.3 Codex (xhigh) (54), 4. Claude Opus 4.6 (Adaptive Reasoning, Max Effort) (53), 5. Claude Sonnet 4.6 (Adaptive Reasoning, Max Effort) (52).
Mercury 2 is the fastest at 756.1 tokens per second, followed by NVIDIA Nemotron 3 Super 120B A12B (Reasoning) (419.9 t/s) and Granite 4.0 H Small (390.8 t/s).
Gemma 3n E4B Instruct is the most affordable at $0.03 per 1M tokens (blended), followed by LFM2 24B A2B ($0.05) and Nova Micro ($0.06).
Gemini 2.5 Flash-Lite (Non-reasoning) has the lowest time to first token at 0.34s, followed by Apriel-v1.5-15B-Thinker (0.37s) and LFM2 24B A2B (0.41s).
GLM-5 (Reasoning) is the highest-ranked open weights model with an Intelligence Index score of 50. There are 193 open weights models out of 295 total evaluated.
The top open weights AI models by Intelligence Index are: 1. GLM-5 (Reasoning) (50), 2. Kimi K2.5 (Reasoning) (47), 3. Qwen3.5 397B A17B (Reasoning) (45).
Gemini 3.1 Pro Preview leads among 146 reasoning models with an Intelligence Index score of 57. Reasoning models use extended thinking to work through complex problems before providing answers.
Models are compared across multiple dimensions including intelligence (quality), pricing, output speed (tokens per second), latency (time to first token), end-to-end response time, and context window size. Performance metrics are measured directly using standardized prompts across 410 models.
Click on any model name or row in the charts to view its dedicated page with detailed metrics and direct comparisons against similar models. You can also use the model selector to customize which models appear in each chart. View the leaderboard