B200 on AWS: Pricing, Specs & How to Rent

Deploybase · March 24, 2026 · GPU Pricing

Contents

B200 Hardware Specifications

NVIDIA B200: 192GB HBM3e, 20 petaFLOPS tensor perf. 16,896 CUDA cores. 8TB/s memory bandwidth. Faster than H100 for LLM training and inference.

Blackwell adds sparsity support and dynamic precision. BF16, FP32. Memory bandwidth: 8TB/s (HBM3e). Runs trillion-parameter models efficiently.

Single and dual-GPU variants available. Scale from research to production clusters. Early adopters see 2-3x speedup over H100.

B200 Availability on AWS

AWS launched B200 in Q1 2026. Limited availability in us-east-1 and us-west-2. Capacity-constrained. Expect multi-week waits if not an early-access customer.

No reserved instances yet. On-demand and spot only. Spot costs 40-50% less but interruption risk is real.

Pricing and Rental Options

B200 on AWS carries premium pricing reflecting both hardware cost and supply constraints. Single B200 GPU instances cost approximately $5.50-$6.50 per hour on-demand, as of March 2026. Dual-GPU B200 instances cost roughly $12-$14 per hour. These rates exceed H100 pricing by 80-100%.

AWS bundles B200s in the p5e instance family for large-scale deployments. Eight-GPU configurations cost approximately $113.93 per hour. This translates to roughly $14.24 per GPU when purchased in bundles. Dual-GPU instances provide the best value for mid-size workloads.

Spot instances reduce B200 costs by 40-50% but carry interruption risk. Production workloads should use on-demand or reserved capacity. Teams benchmarking Blackwell can optimize costs through spot instances during evaluation phases.

B200 vs Other Providers

Lambda Labs charges $6.08 per hour for B200 SXM configuration, undercutting AWS on single-GPU pricing. CoreWeave bundles eight B200s at $68.80 per hour, translating to $8.60 per GPU. RunPod offers B200 at $5.98 per hour, providing the lowest published rate among major providers.

Availability differs significantly across providers. RunPod and CoreWeave report better immediate availability, though supply remains constrained globally. AWS promises higher SLA commitments and dedicated support, which may justify premium pricing for mission-critical workloads.

For teams building production AI infrastructure, AWS integration with VPC, IAM, and S3 provides substantial operational advantages. For price-sensitive research and development, specialist providers like RunPod win on hourly economics.

Getting Started with B200 on AWS

Request B200 access through AWS support. Quota limits restrict B200 capacity, and AWS processes requests manually to manage supply. Expected wait time ranges from days to weeks depending on region selection and account history.

Once quota approves, launch B200 instances through the EC2 console. Select the g6e instance family for single-GPU or dual-GPU configurations. Configure security groups, storage, and networking. Download the deep learning AMI for B200 support, which includes pre-configured CUDA 12.x and NVIDIA driver packages.

After launch, verify GPU availability with nvidia-smi. The B200 should display as a Blackwell architecture GPU with 192GB memory. Install PyTorch with 2.x or newer, TensorFlow 2.17+, or NVIDIA's CUDA C++ libraries. Test basic tensor operations before running production workloads.

FAQ

Q: Is B200 available on AWS? A: Yes, but with limited regional availability. B200 instances exist in select availability zones in us-east-1 and us-west-2. Request quota increases through AWS support.

Q: What does B200 cost per hour on AWS? A: AWS 8xB200 instances cost approximately $113.93 per hour on-demand as of March 2026, or ~$14.24 per GPU.

Q: How much faster is B200 compared to H100? A: Early benchmarks show 2-3x improvements on LLM training workloads. Inference speedups vary by precision and batch size, ranging from 1.5-4x depending on workload characteristics.

Q: Can I use spot instances with B200? A: Yes. Spot instances reduce costs by 40-50% but risk interruption. Use spot for development and evaluation, on-demand for production.

Q: Which provider offers the cheapest B200 pricing? A: RunPod's B200 costs $5.98 per hour. Lambda Labs charges $6.08. AWS on-demand pricing starts at $5.50-$6.50 depending on configuration.

Sources