AI Inference Platform Cost Calculator: Production Pricing Guide

Deploybase · February 3, 2026 · LLM Pricing

Contents

Estimate the inference hosting costs. LLaMA, Mistral, custom models: if developers are serving thousands of queries daily, platform cost beats model quality.

What Drives Inference Costs?

Four things: hardware rental, data transfer, request processing, supporting infrastructure.

Hardware is king. H100 on RunPod: $1.99/hour ($2.69 SXM). 20 hours daily: $1,200-1,600/month.

Data transfer adds 5-15%. Egress: $0.05-0.10 per GB.

Request overhead (gateways, load balancing, monitoring) gets bundled into hourly costs.

Redundancy across availability zones costs more than a single GPU.

Platform-by-Platform Cost Breakdown

RunPod offers transparent per-hour pricing. An H100 PCIe GPU costs $1.99 per GPU-hour, or $1,433 per month assuming continuous operation. For intermittent workloads, costs scale with actual usage.

Lambda Labs pricing is comparable. H100 PCIe at $2.86 per hour costs $2,057 monthly for 24/7 operation. Lambda charges higher per-hour rates but includes more support. See Lambda Labs GPU pricing for current rates.

AWS offers on-demand and reserved instance pricing. H100 instances (p5.48xlarge, 8x H100 SXM) cost approximately $55.04/hour on demand, or $39,629 monthly for continuous operation. Reserved instances reduce costs further with 1-year commitments. Note: P4d instances have A100 GPUs, not H100. See AWS GPU pricing for exact rates.

CoreWeave specializes in large multi-GPU clusters. An 8xH100 pod costs $49.24 per hour, or $35,452 monthly. This configuration runs production workloads serving thousands of concurrent users.

Calculating Infrastructure Costs for Common Scenarios

Scenario 1: Text Generation Service (5,000 daily users)

  • Hardware: RTX 4090 ($0.34/hr on RunPod)
  • Operating hours: 16 hours daily
  • Monthly cost: 16 * 30 * $0.34 = $163.20

Add 10% for data transfer ($16) and API overhead ($50) for total monthly spend of $229.

Cost per user per month: $229 / 5,000 = $0.046

Scenario 2: Image Generation Platform (100 concurrent users)

  • Hardware: A100 GPU (80GB, $1.39/hr on RunPod)
  • Operating hours: 24 hours daily
  • Monthly cost: 24 * 30 * $1.39 = $1,000.80

With infrastructure overhead and redundancy (2 GPUs for failover): $2,002 monthly

Cost per concurrent user: $20 monthly / $240 annually

Scenario 3: Production LLM API (100,000 monthly tokens)

  • Hardware: H100 SXM ($2.69/hr on RunPod)
  • Average inference time: 3 seconds per 100 tokens
  • Monthly compute: 100,000 * 3 = 300,000 seconds = 83.3 hours
  • Hardware cost: 83.3 * $2.69 = $224

Add 15% for data transfer, caching, and redundancy: $258 monthly

Cost per million tokens: $2.58

Right-Sizing Hardware for Workloads

Matching hardware to workload prevents overpaying. A chatbot serving 1,000 queries daily with 2-second response latency doesn't need H100 hardware. RTX 4090 ($0.34/hr) handles the load at a fraction of the cost.

Measure actual inference latency on target hardware before committing. A model requiring 4 seconds on H100 might need 12 seconds on RTX 4090. If response latency matters for user experience, the H100's 3x cost premium is justified.

For throughput-constrained workloads, measure tokens per second instead of latency. A model generating 100 tokens per second on H100 versus 30 on RTX 4090 represents a 3.3x efficiency gain, justifying the cost difference if serving high volumes.

Memory requirements constrain hardware selection. A 70B parameter model needs 140GB VRAM in full precision, requiring A100 (80GB won't fit) or H100 (80GB won't fit either). Quantization to 4-bit reduces memory to 17.5GB, fitting on RTX 4090.

Cost vs performance tradeoffs require benchmarking. Some models degrade gracefully on lower-tier hardware. Others fail at quantization. Measure quality metrics (BLEU scores, user ratings) on candidate hardware before deploying.

Scaling Costs as Traffic Grows

Traffic scaling requires understanding hardware saturation. An RTX 4090 serves approximately 2-4 concurrent inference requests before performance degradation. At 1,000 concurrent users, infrastructure requires 250-500 GPUs.

This scaling introduces non-linear cost increases. Managing 500 GPUs across 50 pods costs significantly more than managing 10 pods due to orchestration complexity and cross-region failover requirements.

Distributed inference (splitting models across multiple GPUs) adds complexity and cost. A 70B model running on a single GPU costs less per request than the same model split across two 35B shards due to orchestration overhead.

Reserved capacity helps with predictable traffic. RunPod and Lambda offer bulk discount pricing for committed usage. A customer committing to 1,000 GPU-hours monthly might negotiate 20-30% discounts versus spot pricing.

API Hosting vs Self-Hosted Comparison

Hosting inference on API providers (Together AI, Anyscale, Modal) costs more per inference but includes operational overhead. A Together AI inference request to LLaMA 2 70B costs roughly $0.001 per 1K tokens. Processing 100M tokens monthly costs $100.

Self-hosting the same model on an A100 GPU ($1.39/hour) for 24/7 operation costs $1,000 monthly. Self-hosting breaks even at 10B monthly tokens, making it economical for heavy usage.

The calculation shifts with redundancy requirements. Self-hosting at scale (10,000 QPS) requires geographic distribution, failover pairs, and monitoring infrastructure. Total operational cost often exceeds direct compute costs by 3-5x.

Provider Comparison Matrix

ProviderH100 CostRTX 4090 CostA100 CostEgress CostMinimum Commitment
RunPod$1.99/hr (PCIe)$0.34/hr$1.39/hr~$0.08/GBNone
Lambda Labs$2.86/hrN/A$1.48/hr$0.08/GBNone
AWS$40/hr$6/hr$14/hr$0.05/GB (same region)None
CoreWeave$49.24/hr (8xH100)N/AN/A$0.08/GBNone

RunPod offers the lowest cost structure with no commitments. Lambda provides better support but higher costs. AWS offers production features but premium pricing. CoreWeave targets large-scale deployments.

Optimization Strategies for Cost Reduction

Implement request batching. Processing 10 inference requests together often costs 30-50% less per request than individual inference. A batch processing pipeline for overnight analytics reduces costs substantially.

Cache inference results aggressively. If 30% of requests are repeats, caching eliminates 30% of compute. Caching overhead (storage, retrieval) is minimal compared to compute savings.

Use model quantization. Running int8 or 4-bit quantized models reduces memory requirements and latency, enabling cheaper hardware. Quality loss typically stays under 2% for well-quantized models.

Implement early stopping. If generating 500 tokens costs more than generating 250 tokens but users only need brief responses, truncating at 250 tokens saves 50% of output cost.

Distribute models across inference platforms based on cost. Use cheaper platforms (RunPod) for standard inference and expensive platforms (AWS) only when compliance or geographic requirements demand it.

Cost Accounting and Budget Management

Track actual costs against projections. Set up billing alerts at 50%, 75%, and 90% of monthly budgets to catch runaway costs.

Implement chargeback systems if multiple teams use shared infrastructure. Teams should understand their true cost impact to make economical design decisions.

Use tagging and cost allocation. Cloud providers support cost labeling by project, department, or user. Detailed tracking enables cost optimization at the granular level.

Review costs monthly. Hardware prices change. A GPU costing $1.99/hour last quarter might cost $1.79/hour this quarter as new capacity comes online. Renegotiating contracts quarterly can save 5-10%.

Advanced Infrastructure Planning

Capacity planning requires forecasting peak usage. A chatbot handling 10,000 daily queries with 3-second latency requires different infrastructure than one handling 1M daily queries with 10-second latency.

Peak vs average load matters significantly. If peak load is 10x average, over-provisioning for peak wastes money. Under-provisioning causes outages. Implement autoscaling to right-size for actual demand.

Geographic distribution adds complexity and cost. Running inference across multiple regions requires inter-region network transfer ($0.02-0.10 per GB depending on provider and direction), redundant infrastructure, and orchestration overhead.

Failover strategies impact costs. Active-active failover (multiple regions serving traffic simultaneously) costs 2x single-region. Active-passive failover (standby capacity) costs 1.5-1.8x. Cold standby (restore on failure) costs 1.2x but risks availability.

GPU Selection Deep Dive

RTX 4090 ($0.34/hr) handles small models efficiently. A quantized 7B model runs at 100+ tokens/second. Per-token cost is approximately $0.0000009/token at that throughput. This economics work for price-sensitive applications.

H100 ($1.99/hr PCIe) handles production inference. A 70B model generates 50 tokens/second. Per-token cost is approximately $0.000011/token. The higher hardware cost buys 2-3x throughput over RTX 4090.

A100 ($1.39/hr) serves intermediate workloads. Memory constraints (40GB or 80GB variants) limit model size. For models fitting A100 constraints, A100 cost-per-inference often beats H100.

B200 ($5.98/hr) targets throughput-extreme workloads. A 70B model achieves 100+ tokens/second. At extreme scale, B200 efficiency justifies premium pricing.

GPU selection directly impacts model selection. Deploying a 70B model on RTX 4090 requires quantization (quality loss). Deploying on H100 preserves quality but increases hardware cost. The tradeoff depends on revenue per inference.

Network and Data Transfer Optimization

Data transfer adds 15-25% overhead to infrastructure costs. Optimize by:

Keeping inference servers geographically close to users (reduce latency, bandwidth) Caching model weights locally (avoid repetitive transfers) Compressing inference outputs (reduce egress costs) Using regional endpoints (avoid cross-region data transfer)

A multi-region deployment can cut bandwidth costs 30-50% by keeping traffic regional and only transferring aggregate results.

VPC peering and private endpoints reduce data transfer costs on AWS and similar providers. Traffic within VPC is often free or heavily discounted compared to inter-region traffic.

Model weights caching is critical. A 70B model is 140GB. Transferring this over network incurs $10+ cost in egress. Caching model weights locally and loading from disk reduces this to startup cost only.

Monitoring and Cost Control

Implement monitoring dashboards showing:

  • GPU utilization percentage
  • Requests per second
  • Cost per request
  • Cost per unit of work (tokens, images, etc)

Use these metrics to identify inefficiency. A cluster with 20% GPU utilization is overprovisioned. A cluster with 95% utilization is undersized.

Cost anomalies (sudden doubling of cost with similar request volume) indicate infrastructure issues. Debug whether queries became more expensive (larger batch size, longer processing) or infrastructure scaled up unexpectedly.

Automated scaling policies should factor in cost. Scale up when utilization exceeds 80%. Scale down when utilization drops below 30%. Avoid oscillation by requiring scale events to persist for 5+ minutes.

Compliance and Production Requirements

Teams with HIPAA, SOC 2, or GDPR requirements have constrained provider choice. AWS and Azure offer these certifications. RunPod and Lambda don't.

This constraint impacts cost optimization. HIPAA requirements might force AWS usage ($40/hour for P4d) over cheaper RunPod ($1.99/hr for H100 PCIe). Compliance costs become the limiting factor, not optimization.

Data residency requirements limit geographic distribution. A workload requiring data in the EU must use European providers. This geographic limitation might mean higher costs and less flexibility.

FAQ

How much does it cost to run GPT-4 equivalent models self-hosted? A 70B parameter model equivalent to GPT-3.5 Turbo quality costs approximately $1,000-1,500 monthly on H100 hardware running 24/7. If targeting GPT-4 quality, a 175B model would require 4-6 H100 GPUs, costing $6,000-9,000 monthly. API-based GPT-4o becomes economical below 50B monthly tokens.

Can I use spot instances to reduce costs? RunPod offers spot pricing at 50-70% discounts. AWS and Lambda also support spot instances. Spot instances can terminate suddenly, requiring failover mechanisms. For non-critical workloads and batch processing, spot instances save money. For real-time APIs, reserved or on-demand instances are safer.

What's the cheapest way to run a language model in production? Quantized 7B parameter models on RTX 4090 GPUs cost approximately $300/month including infrastructure. This setup handles 5,000-10,000 daily queries with 2-3 second latency. API-based inference costs less than $100/month at these volumes but offers less control.

Do inference providers offer volume discounts? Most providers negotiate discounts for teams spending $5,000+/month. RunPod doesn't publish discounts but accepts custom pricing for large customers. Lambda Labs provides bulk discounts at $10,000+/month spend. AWS Reserved Instances provide 30-50% discounts for 1-3 year commitments.

How do I estimate data transfer costs accurately? Assume 20-30% overhead for output data transfer relative to compute costs. For a model generating 500 tokens (1,500 bytes), actual data transfer includes model parameters transferred, caching overhead, and multiple request-response cycles. Most infrastructure costs come from compute, not bandwidth.

What about latency and cost tradeoffs? Latency typically increases linearly with cost reduction. RTX 4090 GPUs show 3-4x higher latency than H100 but cost 6x less. For interactive applications, the cost difference often justifies H100. For batch processing, RTX 4090 is sufficient and economical.

Sources

  • RunPod pricing API documentation (accessed March 2026)
  • Lambda Labs pricing page (accessed March 2026)
  • AWS GPU instance pricing (accessed March 2026)
  • CoreWeave cluster pricing (accessed March 2026)
  • DeployBase.AI infrastructure cost benchmarks (March 2026)