/blog / review / runpod
Runpod review: bare-metal H100s without the enterprise tax
Six weeks on Runpod across Community Cloud, Secure Cloud, and Serverless. The benchmarks, the bills, and where it falls short of AWS and Lambda.
- gpu
- review
- runpod
- h100
- serverless
We rented Runpod for six weeks across all three of their tiers — Community Cloud, Secure Cloud, and Serverless — and ran the same benchmark suite we use for every GPU provider. Below is what we found, what to watch out for, and where the marketing pages tell a slightly nicer story than the reality.
If you want to skip ahead and try them, our referral is runpod.io/?ref=8vbo5oc9 — you get a random credit bonus between $5 and $500 when you spend your first $10. Yes, $500 is real, just unlikely.
Why Runpod, why now
The hyperscalers price H100s like enterprise software. You pay for the GPU, the node, the networking, the egress, and the privilege. Runpod’s pitch is that for 80% of workloads — fine-tuning, batch inference, prototyping — you don’t need any of that. You just need the GPU.
Community Cloud is the cheapest tier, where individual hosts list spare capacity. Secure Cloud is Runpod’s own datacenter footprint with stronger networking and SLAs. Serverless is exactly what it sounds like — pay-per-second, scale-to-zero containers.
The numbers
We ran a fixed harness against four providers. Higher is better unless noted.
- $/hr H100 80GB (PCIe) $ 1.992.491.793.09
- Llama-3 70B tok/s (4-bit) 142138121109
- SDXL imgs/min 38353128
- fio rand-read MB/s 3850410022002900
- Cold start (s) s 2.56.111.48.7
- Egress $/GB $ 000.010.02
A few things stand out. Vast.ai undercuts Runpod on raw $/hr — but the Vast box was a Community-tier listing and the throughput numbers reflect that. Runpod’s Community Cloud was more consistent across our 7-day window than Vast’s, which is the reverse of what you’d expect.
Cold start is the headline metric. Runpod Serverless reliably hit our 2.5-second target across 1,200 invocations. Lambda’s slowest tier was nearly 3× that. If your workload is spiky, this is the difference between affording GPU inference and not.
Pricing in real terms
These are the plans we benchmarked. The setup column is “time from clicking deploy to running your first command.” Egress is per-GB outbound.
- Community Cloud H100 80GB $1.99 ~90s $0
- Secure Cloud H100 80GB $2.79 ~120s $0
- Serverless H100 (active) $0.00074/s ~2.5s cold $0
- RTX 4090 (Community) $0.34 ~60s $0
A 24-hour H100 fine-tune on Community Cloud is about $48. The same job on AWS p5.48xlarge (prorated) is north of $300 once you factor in storage and egress. Even comparing apples to apples — Lambda 1xH100 on-demand at ~$2.49/hr — Runpod is meaningfully cheaper at the going rate, with the caveat that “going rate” on Community Cloud fluctuates.
What’s actually good
The web console is fast. Spawning a pod genuinely takes ~90 seconds from first click. SSH
keys propagate in seconds. The Jupyter integration just works. Volumes are stable across
restarts. The CLI (runpodctl) is a thin but coherent wrapper — nothing fancy, but it
doesn’t fight you.
Serverless is the killer feature. We routed an LLM proxy through Runpod Serverless for a week and the p99 cold start was 2.8 seconds. The p50 with a warm worker was 110ms.
What’s not
Documentation is patchy. Community Cloud capacity disclosures are vague — you’ll hit “no available hosts” warnings on busy days, especially for H200s. Networking between pods is fine, not great; you wouldn’t run a multi-node training job here without a hard look at NCCL throughput.
Their support channel is Discord-first. That’s fine if you live in Discord. Less fine if you’re trying to file a billing question on a Tuesday.
Verdict
Verdict
8.7
/ 10
Pros
- + Best $/hr H100 across our tested providers
- + Genuinely zero egress fees
- + Serverless cold start is real (~2.5s)
- + Fast pod spin-up, clean console
- + Works for hobbyists and production both
Cons
- − Community Cloud variance is real (mostly fine, occasionally not)
- − Docs lag behind product
- − Multi-node networking isn't a strength
- − Discord-first support
Buy it. If you’re choosing your first GPU rental, choose Runpod. If you’re already on Lambda or Paperspace and your monthly bill is under $1k, the move probably isn’t worth it. Above that, run our benchmark suite on Runpod yourself — that 20-30% bill reduction materialises faster than you’d think.
Sign up via our referral if you want the random $5–$500 credit bonus on your first $10 spend. We’ll get a small kickback. Thanks if you do.
review · runpod
Runpod Serverless deep-dive: cold starts, queueing, billing edges
We pushed Runpod Serverless to its limits over a 30-day production deployment. The good and the gotchas.
11 min
review
Runpod Serverless Cold Starts: A Thousand Invocations, Three Weeks Later
We measured cold start latency for a common PyTorch model across 1,000 invocations on Runpod Serverless.
10 min
review · lambda
Lambda Labs review: pretty UI, predictable bill, painful queue
Lambda is the boring-good option for ML teams that hate surprises. Here's what we'd warn you about.
10 min