/blog / comparison

H200 Cloud Pricing: The Hunt for Nvidia's Newest GPU

We scoured Runpod, Lambda Labs, and Vultr for Nvidia's H200, comparing listed prices, actual availability, and the hidden costs that follow the hype.

Tobias 11 min read
  • gpu
  • comparison
  • h200
  • nvidia
  • runpod
  • lambda
  • vultr

The H200 arrived with a bang of press releases and a whisper of actual availability. Nvidia’s latest, greatest, and most memory-endowed GPU promised a significant leap over the H100, especially for gargantuan Large Language Models. We were eager to get our hands on one, run our usual suite of benchmarks, and see if the performance lived up to the price tag. What we found was less about raw numbers and more about an exercise in patience — and the stark reality that ‘listed’ doesn’t always mean ‘available’.

We spent the better part of a month checking dashboards, refreshing API calls, and even talking to sales reps to understand the real-world accessibility and pricing of the H200 across Runpod, Lambda Labs, and Vultr. The promise is there, but the silicon often isn’t.

What We’re Actually Comparing (or Trying To)

Nvidia’s H200 is essentially an H100 with a crucial upgrade: HBM3e memory. This means 141 GB of it (up from H100’s 80 GB) and a massive 4.8 TB/s memory bandwidth (up from 3.35 TB/s). For LLMs, especially those pushing context windows or requiring massive model weights, that extra memory and bandwidth move the needle hard. The theory suggests significantly faster inference and the ability to load larger models or batch sizes without offloading.

Our goal was to compare three major cloud GPU providers who have publicly announced H200 offerings: Runpod, Lambda Labs, and Vultr. We wanted to know not just the hourly rate, but the real-world ease of provisioning, typical queue times, and any hidden costs that often accompany bleeding-edge hardware.

Our Test Workload (or the lack thereof)

Given the extreme scarcity of H200s, a direct performance benchmark was, frankly, impossible for a small outfit like ours. Instead, our ‘workload’ became the act of trying to rent the GPU. We focused on:

  • Stated hourly rates: What the dashboards and pricing pages claimed.
  • Actual availability: How often could we spin up an instance immediately, or what were the reported queue times?
  • Ecosystem factors: Storage costs, egress fees, control plane usability, and support responsiveness.

This isn’t an H200 performance review; it’s an H200 acquisition review, which, for most developers, is the first and hardest hurdle.

The Elusive H200: Availability and Listed Cost

Here’s what we observed across the three platforms. It’s important to note that ‘availability’ can be highly dynamic, especially for new, high-demand hardware. Our observations are snapshots over a few weeks in late Q2/early Q3 2024.

ProviderInstance TypeGPUVRAMListed $/hr (On-demand)Observed AvailabilityNotes
RunpodSecure H200H200141 GB$2.99 - $3.50+Very rare, spot market sometimes.Varied by region, often ‘0 available’. Community Cloud H200s are even rarer.
Lambda LabsH200H200141 GB~$3.50+ (Reserved)Long waitlists, minimal spot.Priority for reserved instances. Spot H200s almost non-existent.
VultrCloud GPUH200141 GB~$3.30+Announced, but not generally available.Still seems to be in limited beta or private preview for H200s.

It’s clear from this snapshot that the H200 is still primarily a ‘reserved instance’ or ‘waitlist’ proposition. The days of spinning up an H200 on a whim are not yet upon us. For perspective, the H100 (80GB) on Runpod’s Secure Cloud generally sits around $1.99/hr, and is actually available most of the time. This makes the H200’s listed price 50-75% higher for a ~75% increase in VRAM and ~43% increase in bandwidth. The economics might eventually make sense for specific large models, but not if you can’t get your hands on the hardware.

Runpod: The Best Shot for Spot

Runpod’s ecosystem, with its Community Cloud and Secure Cloud, offers two avenues. The Community Cloud often surfaces slightly cheaper, consumer-grade GPUs or older enterprise cards. We occasionally saw H100s there, but H200s were a ghost. The Secure Cloud, their on-demand enterprise offering, lists H200 pods. We managed to snag one for a brief test run—a mere 45 minutes—before it was gone. The pricing was around $2.99/hr. The experience was smooth, once we actually got it. For those with flexible workloads and an appetite for refreshing the console, Runpod’s spot market might eventually offer a chance, but it’s not something you can plan around for sustained work.

If you’re already familiar with Runpod’s UI and API for H100s or 4090s (which we’ve covered in our RTX 4090 Cloud Rentals comparison and our Runpod review), the H200 experience is identical once provisioned. The challenge, as always, is the provisioning itself.

Lambda Labs: Reserved and Waiting

Lambda Labs has built a solid reputation for enterprise-grade GPU rentals, particularly for H100s. Our experience with them for H100s has generally been positive, despite the painful queues we noted in our Lambda Labs review. For H200s, the story is similar, but amplified. We were quoted wait times stretching into months for reserved instances, and their spot market for H200s was effectively empty during our testing period. They prioritize existing customers with long-term commitments, which is understandable but frustrating for those looking to kick the tires.

Their pricing for H200s was also on the higher end, hovering around $3.50/hr for what appeared to be their standard on-demand rate, though this would likely come down with reserved instance commitments. Lambda’s strength lies in its predictable environment and good support, but you need to plan far, far ahead for H200s.

Vultr: The Newcomer’s Challenge

Vultr recently entered the high-end GPU cloud market with a splash, offering H100s and announcing H200s. Their infrastructure is solid, and for general cloud VPS, they’ve been a reliable option. However, their H200 offering appears to be the most nascent. While listed on their site, we found no general availability for H200s in any region we checked. It seems they are still in the process of rolling out inventory or are operating a very limited private beta program.

Their H100 pricing is competitive, often slightly below Lambda for on-demand. If they can scale their H200 inventory, Vultr could become a strong contender, but for now, they’re more of a ‘watch this space’ rather than a ‘rent now’ option for the H200.

The Real-World Costs Beyond the Hourly Rate

Even if you manage to snag an H200, the hourly rate is only part of the equation. As we’ve consistently pointed out, egress fees, storage costs, and even network bandwidth can inflate your bill dramatically. The H200 is designed for large models and datasets, which inherently means more data movement.

  • Egress: All three providers have egress charges. Runpod’s are fairly standard, often bundling a reasonable amount of traffic. Lambda’s are competitive. Vultr’s can be steeper depending on the tier. If you’re pulling multi-terabyte models or pushing inference results out, these fees add up faster than you’d think. It’s a lesson we’ve learned the hard way (see our egress cost guide).
  • Storage: Storing multi-hundred-gigabyte models or multi-terabyte datasets means substantial storage costs. All providers offer block storage, but the performance tiers and associated IOPS costs can vary. Pay attention to how quickly you can attach and detach volumes, and if you’re paying for idle storage even when your GPU pod is down.
  • Cold Starts: While H200s are generally for sustained workloads, if you’re using them for bursty inference, cold start times for large containers can still be a factor. Our work on Runpod Serverless cold starts showed how much variance there can be even on smaller models.

Verdict: The H200 is Still a Future Toy

For the vast majority of developers and small teams, the Nvidia H200 is, for now, a paper tiger. The raw performance numbers and memory specs are exciting, but the practical reality of acquiring one in the cloud for on-demand or even short-term reserved use is still a significant hurdle. The market is clearly supply-constrained, with most available units likely going to large enterprise customers with pre-existing commitments.

If your workload absolutely requires the H200’s HBM3e capacity and bandwidth—perhaps you’re training or serving a model that genuinely won’t fit on 80GB of HBM3—then your best bet is to get on Lambda Labs’ waiting list for a reserved instance and settle in for a long wait. For sporadic or experimental access, Runpod’s Secure Cloud occasionally flashes an H200, making it worth monitoring if you’re persistent and flexible. Vultr, while promising, isn’t there yet for general H200 availability.

For everyone else, the H100 remains the pragmatic, readily available, and cost-effective choice for high-end GPU workloads. The performance uplift of the H200 is real for specific niches, but if you can’t buy it, it doesn’t matter. We’d recommend sticking with available H100s for the foreseeable future and only chasing the H200 if your use case truly demands its bleeding-edge memory. If you want to keep an eye on Runpod’s H200 (or more likely, H100) availability, you can check out their platform via our referral link.

Don’t let the marketing hype dictate your hardware choices. Focus on what you can actually rent today, at a predictable price, without a months-long queue. The H200 will matter eventually, but not until it’s actually in stock.