H100$6.39/hr 1.2% 7d
A100 80GB$2.45/hr 0.5% 7d
H200$10.29/hr 0.8% 7d
L40S$1.28/hr 0.3% 7d
T4$0.24/hr 0.6% 7d
L4$0.45/hr 1.1% 7d
H100$6.39/hr 1.2% 7d
A100 80GB$2.45/hr 0.5% 7d
H200$10.29/hr 0.8% 7d
L40S$1.28/hr 0.3% 7d
T4$0.24/hr 0.6% 7d
L4$0.45/hr 1.1% 7d
Monthly Report

GPU Market Report — March 2026

Signwl ResearchApril 1, 20265 min read

Executive Summary

  • Market overview: Blended GPU compute pricing increased in March 2026, with an average month-over-month change of +13.7% across 39 tracked accelerator types. Training GPUs moved +0.4% while inference GPUs moved +0.4%.
  • Biggest movers: Largest price declines: V100 32GB (-6.7% MoM). Largest increase: ALVEO_U30 (+27.9% MoM). H100 pricing held firm as Blackwell supply continues to expand.
  • Spread: The training-to-inference pricing spread held stable at 8.3x.

Training GPU Pricing — March 2026

Source: Signwl · Blended pricing (spot + on-demand + reserved)

Pricing Trends

Blended GPU compute pricing increased in March 2026. Across 39 tracked accelerator types, the average month-over-month price change was +13.7%, reflecting sustained demand for AI compute capacity.

The month was characterised by broadly consistent movement across GPU tiers. Training-class GPUs — the workhorses of AI model development — averaged $4.30/hr (+0.4% MoM), while inference-class GPUs — used for serving models in production — averaged $0.52/hr (+0.4% MoM).

Training GPUs

GPUMarch PriceMoM ChangeRegions
GB200$14.95/hr-0.0%99
H200$10.18/hr+0.0%146
B200$7.35/hr+1.0%15
H100$6.73/hr+1.4%144
MI300X$3.39/hr+0.0%81
A100 80GB$2.41/hr+1.6%144
A100 40GB$1.76/hr+3.3%127

The H100 — still the most widely deployed training GPU — showed pricing resilience at $6.73/hr (+1.4% MoM).

AMD MI300X at $3.39/hr (+0.0% MoM) continues to gain traction as a competitive alternative to NVIDIA's Hopper generation, available across 81 regions. The price differential between MI300X and H100 stands at $3.34/hr, or 50% — a significant gap that may drive further AMD adoption.

Inference GPUs

Inference GPU Pricing — March 2026

Source: Signwl · Blended pricing (spot + on-demand + reserved)
GPUMarch PriceMoM ChangeRegions
L40S$1.32/hr+3.7%55
L4$0.51/hr+1.4%124
A10G$0.71/hr+0.8%99
A10$0.92/hr+0.3%166
T4$0.24/hr+0.2%152

The inference tier showed more movement than training this month. Budget inference options like the T4 ($0.24/hr) remain the most cost-effective entry point for production AI deployment.

Training vs Inference Gap

The training-to-inference pricing ratio stands at 8.3x — stable compared to last month.

This ratio is a key structural indicator for the AI compute market. A widening spread suggests strong demand for frontier model training is outpacing inference deployment. A narrowing spread would signal inference demand growing relative to training.

Training GPUs averaged $4.30/hr (+0.4% MoM) while inference GPUs averaged $0.52/hr (+0.4% MoM). The gap remains elevated by historical standards, suggesting the market has not yet reached equilibrium on training compute pricing.

Regional Dynamics

Regional GPU Pricing — March 2026

Source: Signwl · Blended pricing across all GPU types
RegionAvg Blended PriceGPUs Available
Europe$2.24/hr34
Asia Pacific$2.26/hr31
North America$2.32/hr38
Oceania$3.20/hr20
South America$3.31/hr19
Africa$3.82/hr7
Middle East$4.09/hr13

Europe offers the lowest average blended GPU pricing at $2.24/hr, while Middle East is the most expensive at $4.09/hr — a 83% premium.

This regional spread reflects differences in data centre maturity, energy costs, and provider competition. For training workloads where latency is not critical, deploying in the cheapest region can save 83% on compute costs. For inference serving local users, the cost premium of deploying in-region is typically offset by latency improvements.

Detailed regional analysis and comparisons →

Hardware Transition Watch

Blackwell Generation

GPUMarch PriceMoM ChangeRegions
B200$7.35/hr+1.0%15
B300$10.31/hr+0.0%4
GB200$14.95/hr-0.0%99
GB300$12.32/hr+7.1%14

Blackwell-generation GPUs are now available across 132 region instances — a significant expansion from last month.

The H100 at $6.73/hr (+1.4% MoM) is showing surprising resilience despite increasing Blackwell availability. This suggests Hopper-generation demand remains robust — likely driven by the H100's mature software ecosystem and deep integration into existing training pipelines. NVIDIA's revenue from H100 may hold longer than some analysts expect.

Implications

For GPU investors: Rising rental rates are improving returns, making this a favourable window for GPU acquisition. Model scenarios with the GPU ROI Calculator →

For cloud buyers: Europe continues to offer the best value for cost-sensitive workloads. The 8.3x training-to-inference spread means organisations can still deploy inference at a fraction of training costs. Compare regional pricing →

For semiconductor analysts: H100 pricing resilience (+1.4% MoM) supports the thesis that NVIDIA's data centre revenue remains well-supported despite the Blackwell ramp. Watch for signals on H200 adoption rates. AMD MI300X pricing (+0.0% MoM) suggests growing competitive traction. View all GPU profiles →

Source: Signwl data for March 2026 (2026-03-01 to 2026-03-31). All prices are GPU-only (excluding CPU, RAM, networking), blended across spot, on-demand, and 1-year reserved rates from major cloud providers. Month-over-month changes compare end-of-March to end-of-previous-month blended averages.

monthly reportGPU pricingmarket analysisAI infrastructuremarch 2026
Get Started

Explore Signwl's GPU Data

Live pricing, regional analysis, and comparisons for 39 GPU and AI accelerator types.