Stay connected with us on X, Discord, and LinkedIn to stay up to date with future analysis

AI Hardware Benchmarking & Performance Analysis

Comprehensive benchmarking of AI accelerator systems for language model inference. We measure how performance scales with concurrent load on NVIDIA 8×H100, 8×H200, and 8×B200 systems using Deepseek R1, Llama 4 Maverick, Llama 3.3 70B, and GPT-OSS 120B.

For details regarding the methodology, see our methodology section. Benchmarks are conducted periodically, at least once per quarter, and benchmark specifications are shared in the System & Benchmark Specifications section below.

For model benchmarks, see our LLM model comparison.

Coming soon:Model Deployment Report

Highlights

Peak System Output Throughput, Llama 3.3 70B
Total System Output Tokens per Second; Higher is better
Peak Output Speed per Query, Llama 3.3 70B
Output Tokens per Second per Query; Higher is better
Rental Price (ON-DEMAND)
Minimum Rental Price per GPU per Hour, USD; Lower is better

Price per GPU Hour

Price per GPU Hour

Leading cloud hyperscaler endpoints; Price in USD
Minimum
Runpod
Crusoe Cloud
Digitalocean
Amazon Web Services
Google Cloud
Nebius
Microsoft Azure
Lambda
Coreweave

Instance configurations may differ slightly by provider, even when using the same GPU model. Factors like memory size, interconnect bandwidth, and system architecture can impact pricing.

Prices are based on the US Central region, or the closest available equivalent if not directly listed (e.g., US East - Ohio - for AWS).

Pricing reflects on-demand hourly rates as listed by each cloud provider.

Prices shown are updated at the start of each month. While prices may fluctuate daily, we maintain a monthly update schedule to provide consistent comparisons.

System Output Throughput at 100 tokens/s Per Query Output Speed

gpt-oss-120B (high) | System Output Throughput (Tokens per Second) at 100 tokens/s Output Speed

The total number of tokens that can be processed per second across all concurrent requests. This metric measures the overall system capacity and efficiency, taking into account both the per-query output speed and the system's ability to handle multiple concurrent requests.

  • Max Throughput: Optimized for the highest sustained request volume.
  • Minimum Latency: Tuned to deliver the fastest response times.
  • Optimal: Combines max throughput and minimum latency configurations by selecting the configuration with higher system throughput at a given concurrency level.

Peak Output Speed per Query

gpt-oss-120B (high) | Peak Output Speed per Query (Tokens per Second)

Tokens per second received by each individual query after the first chunk has been received. Represented as the median result of all queries in each concurrency phase.

  • Max Throughput: Optimized for the highest sustained request volume.
  • Minimum Latency: Tuned to deliver the fastest response times.
  • Optimal: Combines max throughput and minimum latency configurations by selecting the configuration with higher system throughput at a given concurrency level.

System Output Throughput vs Output Speed per Query

gpt-oss-120B (high) | System Output Throughput (Tokens per Second) vs Output Speed per Query (Tokens per Second)
8xH100 - vLLM
8xH200 - vLLM
8xB200 - TensorRT-LLM - Optimal
8xMI300X - vLLM

The total number of tokens that can be processed per second across all concurrent requests. This metric measures the overall system capacity and efficiency, taking into account both the per-query output speed and the system's ability to handle multiple concurrent requests.

Tokens per second received by each individual query after the first chunk has been received. Represented as the median result of all queries in each concurrency phase.

  • Max Throughput: Optimized for the highest sustained request volume.
  • Minimum Latency: Tuned to deliver the fastest response times.
  • Optimal: Combines max throughput and minimum latency configurations by selecting the configuration with higher system throughput at a given concurrency level.

System Output Throughput & Output Speed per Query vs. Concurrency

gpt-oss-120B (high) | System Output Throughput (Tokens per Second) & Output Speed per Query (Tokens per Second)
8xH100 - vLLM
8xH200 - vLLM
8xB200 - TensorRT-LLM - Optimal
8xMI300X - vLLM
Throughput
Speed

Tokens per second received by each individual query after the first chunk has been received. Represented as the median result of all queries in each concurrency phase.

The total number of tokens that can be processed per second across all concurrent requests. This metric measures the overall system capacity and efficiency, taking into account both the per-query output speed and the system's ability to handle multiple concurrent requests.

The number of simultaneous requests that can be processed by the API at any given time. Higher concurrency enables better handling of parallel requests, which is essential for applications requiring high throughput and scalability.

  • Max Throughput: Optimized for the highest sustained request volume.
  • Minimum Latency: Tuned to deliver the fastest response times.
  • Optimal: Combines max throughput and minimum latency configurations by selecting the configuration with higher system throughput at a given concurrency level.

Cost per Million Input and Output Tokens at 100 tokens/s Per Query Output Speed

gpt-oss-120B (high) | Cost per One Million Input and One Million Output Tokens (USD) at 100 tokens/s Output Speed

The cost per million input and output tokens is calculated using the average price per GPU per hour and the system output throughput, assuming 1k input and 1k output tokens per request. The formula is:

average price per GPU per hour×number of GPUs×1,000,000system output throughput×3600\frac{\text{average price per GPU per hour} \times \text{number of GPUs} \times 1,000,000}{\text{system output throughput} \times 3600}

Prices shown are updated at the start of each month. While prices may fluctuate daily, we maintain a monthly update schedule to provide consistent comparisons.

  • Max Throughput: Optimized for the highest sustained request volume.
  • Minimum Latency: Tuned to deliver the fastest response times.
  • Optimal: Combines max throughput and minimum latency configurations by selecting the configuration with higher system throughput at a given concurrency level.

End-to-End Latency vs. Concurrency

gpt-oss-120B (high) | End-to-End Latency (s) vs. Concurrency
8xH100 - vLLM
8xH200 - vLLM
8xB200 - TensorRT-LLM - Optimal
8xMI300X - vLLM

The time it takes for a request to be processed and the response to be returned.

The number of simultaneous requests that can be processed by the API at any given time. Higher concurrency enables better handling of parallel requests, which is essential for applications requiring high throughput and scalability.

  • Max Throughput: Optimized for the highest sustained request volume.
  • Minimum Latency: Tuned to deliver the fastest response times.
  • Optimal: Combines max throughput and minimum latency configurations by selecting the configuration with higher system throughput at a given concurrency level.

System & Benchmark Specifications

Model NameSystemProviderPrecisionTensor ParallelExpert ParallelData ParallelInference FrameworkKernel VersionConfigurationConfiguration ReferenceDate
Llama 4 Maverick8xB200 (SXM)googlefp8811vLLM 0.9.0.1CUDA 12.8Default4 June 2025
Llama 4 Maverick8xH100 (SXM)googlefp8811vLLM 0.9.2CUDA 12.8Default4 August 2025
Llama 4 Maverick8xH200 (SXM)googlefp8811vLLM 0.9.2CUDA 12.8Default4 August 2025
Llama 4 Maverick8xB200 (SXM)googlefp8881TensorRT-LLM 1.2.0rc3CUDA 13.0Max throughputLink19 November 2025
Llama 4 Maverick8xMI300Xrunpodfp8811vLLM 0.11.1rc2ROCm 7.0DefaultLink19 November 2025
Llama 4 Maverick8xB200 (SXM)googlefp8821TensorRT-LLM 1.2.0rc1CUDA 13.0BalancedLink19 November 2025
Llama 4 Maverick8xB200 (SXM)googlefp8811TensorRT-LLM 1.2.0rc3CUDA 13.0Min latencyLink19 November 2025
Llama 3.3 Instruct 70B8xB200 (SXM)googlebf1681vLLM 0.9.2CUDA 12.8Default4 August 2025
Llama 3.3 Instruct 70B8xH100 (SXM)googlebf1681vLLM 0.9.2CUDA 12.8Default4 August 2025
Llama 3.3 Instruct 70B8xH200 (SXM)googlebf1681vLLM 0.9.2CUDA 12.8Default4 August 2025
Llama 3.3 Instruct 70B8xMI300Xrunpodbf1681vLLM 0.11.1rc2ROCm 7.0DefaultLink19 November 2025
Llama 3.3 Instruct 70B8xB200 (SXM)googlebf1681TensorRT-LLM 1.2.0rc1CUDA 12.9DefaultLink21 November 2025
Llama 3.3 Instruct 70BTPU v6e-8googlebf1681vLLM vllm-tpu:nightly-20251129-28607fc-39e63deXLA [v2-alpha-tpuv6e]DefaultLink9 December 2025
gpt-oss-120B (high)8xB200 (SXM)nebiusmxfp4881TensorRT-LLM 1.1.0rc1CUDA 12.9Max throughputLink29 September 2025
gpt-oss-120B (high)8xH100 (SXM)googlemxfp4811vLLM 0.10.0CUDA 12.8Default8 August 2025
gpt-oss-120B (high)8xH200 (SXM)googlemxfp4811vLLM 0.10.0CUDA 12.8Default8 August 2025
gpt-oss-120B (high)8xB200 (SXM)nebiusmxfp4811TensorRT-LLM 1.1.0rc1CUDA 12.9Min latencyLink29 September 2025
gpt-oss-120B (high)8xB200 (SXM)nebiusmxfp481TensorRT-LLM 1.1.0rc1CUDA 12.9Optimal15 October 2025
gpt-oss-120B (high)8xMI300Xrunpodmxfp4881vLLM 0.10.1ROCm 7.0DefaultLink19 November 2025
DeepSeek R1 0528 (May '25)8xH200 (SXM)googlefp8811vLLM 0.9.0.1CUDA 12.8Default5 June 2025
DeepSeek R1 0528 (May '25)8xB200 (SXM)googlefp8811vLLM 0.9.0.1CUDA 12.8Default4 June 2025
DeepSeek R1 0528 (May '25)8xB200 (SXM)googlefp4881TensorRT-LLM 1.2.0rc2CUDA 13.0Optimal19 November 2025
DeepSeek R1 0528 (May '25)8xH200 (SXM)googlefp8881TensorRT-LLM 1.0.0rc2CUDA 12.9Max throughputLink31 July 2025
DeepSeek R1 0528 (May '25)8xB200 (SXM)googlefp8881TensorRT-LLM 1.1.0rc2.post1CUDA 12.9Max throughputLink8 September 2025
DeepSeek R1 0528 (May '25)8xMI300Xrunpodfp8811SGLang 0.5.2rc2ROCm 7.0DefaultLink19 November 2025
DeepSeek R1 0528 (May '25)8xB200 (SXM)googlefp8881TensorRT-LLM 1.2.0rc2CUDA 13.0Min latencyLink19 November 2025
DeepSeek R1 0528 (May '25)8xB200 (SXM)googlefp4881TensorRT-LLM 1.2.0rc3CUDA 13.0Max throughputLink19 November 2025
DeepSeek R1 0528 (May '25)SN40L-16sambanovafp8111na-Optimal12 January 2026
DeepSeek R1 0528 (May '25)8xB200 (SXM)googlefp881TensorRT-LLM 1.2.0rc2CUDA 13.0Optimal19 November 2025
DeepSeek R1 0528 (May '25)8xB200 (SXM)googlefp4881TensorRT-LLM 1.2.0rc2CUDA 13.0Min latencyLink19 November 2025