The price changes: side-by-side breakdown
EC2 Capacity Blocks for ML allow pre-booking accelerator capacity for fixed periods, ideal for scheduled ML training. The recent update hit p5e and p5en instances (8x Nvidia H200 GPUs each), with most regions seeing jumps from prior rates.| Instance type | Key regions | Old hourly rate | New hourly rate | % Increase |
|---|---|---|---|---|
| p5e.48xlarge | Global (Most) | ~$34.61 | $39.80 | ~15% |
| p5e.48xlarge | US West (N. CA) | ~$43.26 | $49.75 | ~15% |
| p5en.48xlarge | Global (Most) | ~$36.18 | $41.61 | ~15% |
| p5en.48xlarge | US West (N. CA) | Higher baseline | $52.02 | ~15%+ |
Why now? Supply, demand, and AI economics
H200 GPUs power large-scale inference and fine-tuning, but Nvidia supply constraints have created bottlenecks. AWS’s move reflects surging demand outpacing even their massive investments in Trainium/Inferentia alternatives.
- Public list prices anchor discounts: Even if you have a private pricing agreement, a 15% list hike can inflate your effective costs if your discount is calculated as a percentage off the public rate.
- Capacity scarcity: While AWS claims On-Demand prices won’t rise, the reality is that “On-Demand” availability for H200s is rare, effectively forcing enterprises into the more expensive Capacity Block model to ensure project timelines.
Enterprise risks and finops impact
Beyond the numbers, this sets precedents:
- Vendor dependency amplified. Heavy AWS AI reliance means indirect exposure to Nvidia pricing power via hyperscalers.
- Budget creep in stealth mode. Changes buried in pricing pages evade standard change alerts; quarterly reviews often miss them until the bill arrives.
- Contract vulnerabilities. Many Enterprise Discount Programs (EDPs) are exposed to list-price volatility on “dynamic” SKUs.
Get ready for what’s next with insights and breakthrough topics in cloud, AI, and innovation. Join our newsletter for curated topics delivered straight to your inbox.
By signing up, you agree to Cloud Latitude’s Privacy Policy and Terms of Use.
5 strategies to mitigate GPU pricing risk
Cloud Latitude recommends these steps for immediate resilience:
- Diversify your accelerator mix. Blend H100/H200 workloads with AWS Trainium (trn2.48xlarge) or explore multi-cloud alternatives like Azure ND H200 v5 to benchmark costs.
- Lock in rates via EDP renewals. Negotiate clauses that cap list-price exposure on dynamic SKUs during your next contract renewal.
- Model “worst-case” scenarios. Use the AWS Pricing Calculator to run 15–25% cost stress tests on upcoming AI roadmaps.
- Optimize reservation types. Favor Savings Plans for steady-state production; use Capacity Blocks strictly for time-sensitive, high-burst training windows.
- Automate price tracking. Use Lambda and CloudWatch to scrape pricing pages or alert FinOps teams to “dynamic” price updates before they impact the monthly budget.
Cloud Latitude’s takeaway
This hike isn’t panic-worthy, but it’s a pivot point: AI infrastructure pricing is maturing into a two-way street. Enterprises succeeding here will treat GPUs like commodities—hedged, benchmarked, and negotiated aggressively.
Need a custom GPU TCO audit or multi-cloud migration roadmap?
Contact Cloud Latitude for a no-obligation strategy session.


