← Hardware · RTX 5090 vs RTX 4090Head-to-head · matched quantisationIssue: April 22, 2026
Matched precision · same workloads · April 2026

RTX 5090 vs RTX 4090.

Blackwell vs Ada. The 5090 lifts FP16 by 1.27×, VRAM from 24 → 32 GB, and bandwidth from 1,008 → 1,792 GB/s — for $400 more MSRP and 125 W more under load. The case for upgrading is mostly about the extra 8 GB.

§ 01 · Specs

Side by side, on paper.

Datasheet specs only. Throughput on real workloads follows in §02 — the gap there is often smaller than the FP16 number suggests, because most ML workloads are memory-bound.

SpecRTX 5090RTX 4090
VendorNVIDIANVIDIA
TierConsumerConsumer
GenerationBlackwellAda Lovelace
VRAM32 GB GDDR724 GB GDDR6X
Bandwidth1,792 GB/s1,008 GB/s
FP16 dense209.5 TFLOPS165.2 TFLOPS
TDP575 W450 W
Released20252022
StatusAvailableAvailable
Price$1,999 MSRP$1,599 MSRP
Fig 2 · Spec deltas. Copper dot marks the column with the bigger number for that axis (lower W is better; otherwise higher).
§ 02 · Benchmarks

On real workloads.

Same model revision, same quantisation, same batch size on both cards. Where one side has no measurement we leave the cell empty rather than extrapolate.

Methodology: how we test.

CategoryWorkloadMetricRTX 5090RTX 4090Δ
LLM InferenceLlama 3.1 8Btok/s140950.68×
LLM InferenceLlama 3.1 70B · 4-bittok/s38220.58×
LLM InferenceQwen 2.5 32B · 4-bittok/s48300.63×
LLM InferenceMistral 7Btok/s1651100.67×
Image GenerationSDXL 1024×1024it/s6.54.20.65×
Image GenerationFlux.1 Devit/s3.42.10.62×
TrainingFine-tune Llama 3.1 8B LoRAsamples/s12.57.80.62×
TrainingResNet-50 · ImageNetimg/s2,8001,9500.70×
Computer VisionYOLOv8x · inferenceFPS3202100.66×
Computer VisionSAM ViT-Hmasks/s9.25.80.63×
Audio/VideoWhisper Large v3× RT28180.64×
Fig 3 · Δ column shows RTX 4090 ÷ RTX 5090 on the same workload. Copper dot marks the winner per row.
§ 03 · Verdict

When each one wins.

The right card is the one whose envelope covers your worst-case workload — not the one with the bigger TFLOPS number.

Pick the RTX 5090

When the RTX 5090 wins.

When your worst-case workload needs more than 24 GB. 32 GB unlocks a 70B at INT4 with 32k context, comfortable LoRA on 13B models, and sustained Flux generation without OOMs.

Pick the RTX 4090

When the RTX 4090 wins.

On price/perf. $1,599 vs $1,999 buys you 79% of the FP16 and 75% of the VRAM. If your worst-case workload fits in 24 GB, the 4090 stays the value pick — especially with $0.29/hr aggregator clouds for spillover.

Bottom line. Buy the 5090 if you need 32 GB locally; otherwise the 4090 + occasional cloud H200 hours is still the sharpest workstation strategy.

§ 04 · More head-to-heads

Other matchups, same format.

/hardware/h200-vs-b200

H200 vs B200

Hopper’s last word against Blackwell’s first. 4.5× the FP16, almost 50% more VRAM bandwidth.

/hardware/h100-vs-h200

H100 SXM vs H200

Same FP16 ceiling (989), but H200 nearly doubles VRAM (80 → 141 GB) and 1.4× bandwidth.

/hardware/rtx-5090-vs-h200

RTX 5090 vs H200

The biggest consumer card vs a real datacenter accelerator. When does the 5090 actually catch up?

Read next

Three places to go from here.

Per-chip page
RTX 5090
First consumer card with 32 GB. The ceiling for a single-PSU workstation.
Per-chip page
RTX 4090
Still the workhorse: 24 GB GDDR6X, $0.29/hr on Vast.ai spot.
Hub
Hardware register
Every accelerator on the leaderboard with FP16 TFLOPS, VRAM, $/hr, and energy cost.