Building the own data center costs serious money. Renting from RunPod, Lambda Labs, or CoreWeave almost always wins financially. But not always. Let's break the math.
Contents
- AI Data Center Cost: GPU Acquisition Costs and Hardware Economics
- Power Infrastructure Costs and Utility Expenses
- Cooling System Costs and Heat Removal
- Physical Infrastructure Costs and Real Estate
- Networking Infrastructure and Interconnect Costs
- Personnel and Operational Labor Costs
- Total Cost of Ownership Comparison
- Cloud GPU Rental Comparison Economics
- When Self-Hosted Makes Sense
- Regional Variations in Infrastructure Economics
- Risk Factors in Self-Hosted Economics
- Economics Sensitivity Analysis
- Capital Preservation and Financial Flexibility
- Economies of Scale at Large Deployments
- FAQ
- Related Resources
- Conclusion and Recommendation
- Sources
AI Data Center Cost: GPU Acquisition Costs and Hardware Economics
AI Data Center Cost is the focus of this guide. H100 costs $30-40K per GPU. Resellers knock off 10-15%. Volume buys get more.
A100: $10-15K. B200: $40-50K (scarcity premium). Expect normalization to $35-40K.
Eight H100s cost $240-320K. That's hardware only. Infrastructure, networking, mounting is extra. Larger clusters go into millions.
Bulk gets 5-10% discounts. Most teams don't buy in bulk.
3-4 year useful life. H100 cluster depreciates $60-80K annually. Monthly depreciation alone: $5-6.6K.
Power Infrastructure Costs and Utility Expenses
H100 draws 700W. Eight of them: 5.6 kW continuous.
Power costs vary by region. US average: $0.06-0.12 per kWh. Eight H100s at $0.09: $3,684/month, $44,208/year. That assumes 100% utilization. Reality is lower. But they run 24/7 whether developers use them or not.
California: $0.14-0.18/kWh → $6-8K/month. Texas: $0.05-0.07/kWh → $2.3-3.2K/month. Location matters. 35-40% difference.
Cooling System Costs and Heat Removal
Removing 5.6 kilowatts of heat requires substantial cooling infrastructure. Traditional data center cooling operates at 4:1 power ratio: one watt of cooling power required per four watts of heat.
This means 1.4 kilowatts of cooling power needed for the 8-GPU cluster. Liquid cooling systems, necessary for high-density GPU clusters, cost $15,000-40,000 installed for small clusters.
Annual cooling operational costs approach power costs. Cooling efficiency (power usage effectiveness) ratios of 1.5 mean cooling adds 50% to power costs. For the cluster:
Power: $44,208 annual Cooling: $22,104 additional Total power and cooling: $66,312 annually
These numbers assume dedicated facility cooling. Sharing infrastructure across multiple clusters improves efficiency per GPU significantly, reducing per-unit costs.
Physical Infrastructure Costs and Real Estate
Building a small data center or co-location facility requires physical space. Rack rental at co-location facilities costs $500-2,000 monthly depending on power density and location.
For owned infrastructure, space acquisition represents significant capital. A single 8-GPU cluster requires approximately 1 rack, occupying 42 RU. Building costs per square foot vary wildly: $100-500 per square foot for industrial space in tier-2 US cities, $300-1,000 per square foot in major metro areas.
A 1,000 square foot room suitable for 50-100 racks costs $100,000-500,000 in buildout alone. Amortized over 10 years, space represents $10,000-50,000 in annual carrying costs.
Rack and mounting hardware costs $2,000-5,000 per rack. 8-GPU cluster mounting costs $2,000-5,000 upfront.
For small deployments, co-location facilities avoid physical infrastructure investment entirely. Costs scale as clusters grow, making owned infrastructure competitive at 100+ GPU scale.
Networking Infrastructure and Interconnect Costs
High-bandwidth inter-GPU networking enables distributed training. InfiniBand network interfaces cost $1,000-3,000 per GPU. An 8-GPU cluster requires $8,000-24,000 in network hardware plus installation.
Networking switch infrastructure for connecting GPUs costs $5,000-15,000 depending on required bandwidth. Small clusters can share switch infrastructure; costs amortize across more GPUs.
Network cables, fiber optics, and installation add $2,000-5,000 for small clusters.
Annual networking maintenance and redundancy infrastructure costs $1,000-5,000 depending on cluster size and requirements.
Personnel and Operational Labor Costs
Deploying and maintaining infrastructure requires expertise. GPU cluster configuration, cooling management, and networking setup typically requires 2-4 weeks of senior engineer time.
At $150-250 per hour fully loaded cost, setup consumes $12,000-32,000 in personnel investment.
Ongoing operations require monitoring, updates, and troubleshooting. A dedicated infrastructure engineer earns $120,000-180,000 annually fully loaded. For teams with only one small cluster, this salary allocates entirely to GPU infrastructure.
Teams with 5+ clusters can share one engineer across clusters, reducing per-cluster allocation. Teams with one cluster allocate the full salary.
Total Cost of Ownership Comparison
An 8-GPU H100 cluster costs comprehensively:
Capital and Depreciation
- Initial GPU acquisition: $240,000-320,000
- Amortized over 3 years: $80,000-106,000 annually
- Supporting infrastructure (racks, cables, PSUs): $2,000-5,000 amortized
Operational Costs
- Power: $44,208 annually
- Cooling: $22,104 annually
- Space/real estate: $10,000-50,000 annually (varies dramatically by buildout)
- Networking: $5,000-10,000 annually
- Maintenance and repairs: $3,000-7,000 annually
- Monitoring and observability tools: $2,000-5,000 annually
Labor Costs
- Personnel: $60,000-180,000 annually
- Infrastructure management: $20,000-40,000 annually
- Emergency support: $10,000-20,000 annually
Estimated Annual Total: $256,312-423,312
At the midpoint ($340,000), per-GPU cost calculation:
- Monthly cost: $28,333
- Cost per GPU per hour: $28,333 / (8 GPUs × 730 hours) = $4.86 per GPU per hour
This assumes 100% utilization (unrealistic). At 50% utilization:
- Cost per GPU per hour: $9.72
- At 30% utilization (more realistic): $16.20
Real-world scenario: A team averaging 40% utilization pays $3.58/GPU/hour self-hosted versus $2.69/hour cloud. Cloud wins by 25% even in moderately favorable conditions.
Cloud GPU Rental Comparison Economics
RunPod H100 costs $2.69 per GPU per hour. This beats self-hosted infrastructure even at full utilization scenarios.
At 50% utilization, self-hosted costs double to $6.28 per hour, making cloud clearly superior by 2-3x. Even at perfect 100% utilization, self-hosted marginally beats cloud ($3.14 vs $2.69), but infrastructure risk eliminates the tiny 14% advantage.
CoreWeave's $49.24 per hour for 8xH100 cluster ($6.16 per GPU) beats self-hosted at any realistic utilization rate. The managed infrastructure and support justify the premium over bare-bones RunPod pricing.
Multi-Year Cost Analysis
Comparing three-year cumulative costs reveals hidden economics:
Self-hosted 8x H100 cluster over 3 years:
- Year 1: $220,000 (includes setup)
- Year 2: $180,000 (no setup)
- Year 3: $180,000
- Total 3-year: $580,000
RunPod equivalent ($2.69 × 8 GPUs × 8,760 hours × 3 years = $564,691):
- Slightly cheaper despite appearing more expensive hourly
- No capital risk if GPUs become obsolete
- No stranded infrastructure if workloads shift
The economics tighten significantly over multi-year horizons, with cloud becoming increasingly attractive once infrastructure risk is factored in.
When Self-Hosted Makes Sense
Multi-year sustained utilization above 80% favors self-hosted infrastructure. Long-term committed workloads with minimal variation justify capital investment. The key threshold is predictability. If the GPU demand varies 20% month-to-month, cloud wins. If demand stays within 5% range, self-hosted becomes viable.
Teams running continuous training pipelines benefit from owned infrastructure. Amortization across 3+ years creates economics favoring self-hosted. Consider a team training foundation models continuously. If they know they'll use 50+ GPUs for the next 3-5 years without interruption, self-hosting saves money.
Cost-plus pricing to internal teams drives adoption. If the organization charges cost-plus 30% to teams using GPU infrastructure, in-house clusters become attractive financially. This pricing model internally funds infrastructure improvements and expansion while making usage feel cheaper to engineering teams than public cloud.
Self-hosted infrastructure also makes sense when:
- Developers have dedicated site reliability engineers comfortable with GPU infrastructure
- Developers own appropriate real estate already
- Developers negotiate favorable power rates (under $0.07/kWh)
- The workloads require extreme customization (exotic CUDA kernels, specialized drivers)
- Developers have purchasing power with NVIDIA for volume discounts
- The utilization can be consolidated across multiple teams
Regional Variations in Infrastructure Economics
Building in low-cost power regions (Iceland, Pacific Northwest) reduces ongoing costs substantially. Power costs drop 30-40% versus US averages. However, facility costs and personnel remain similar.
International expansion worsens costs. Importing GPUs internationally adds tariffs and logistics. Personnel costs exceed US levels in many advanced economies.
Cooling-efficient climates (cool regions) improve economics. Air-side cooling versus water-side cooling saves significant operational cost in cool climates.
Risk Factors in Self-Hosted Economics
GPU depreciation risk: Models become obsolete faster than 3-year amortization timelines. B100 or future architectures may render H100s less desirable within 2 years.
Utilization risk: Actual utilization typically underperforms projections. Planning 80% utilization often yields 40-60% reality.
Staffing risk: Infrastructure expertise shortage makes scaling difficult. Adding engineers becomes expensive and introduces execution risk.
Opportunity cost: Capital deployed to GPUs cannot deploy to other business priorities. Renting GPU capacity preserves capital for other investments.
Economics Sensitivity Analysis
Power Cost Reduction (50% via relocation) Saves $22,104 annually, reducing total cost from $340,000 to $317,896. Self-hosted improves per-GPU cost from $4.86 to $4.48/hour at 100% utilization. Still doesn't beat cloud ($2.69).
Utilization Improvement (50% → 100%) Cuts per-GPU cost from $9.72 to $4.86/hour at 100% utilization, making self-hosted approach cloud competitiveness. However, achieving sustainable 100% utilization is extremely difficult. Most teams plateau around 50-70%.
Hardware Lifespan Extension (3 → 5 years) Reduces annual depreciation from $80,000 to $48,000. Lowers total cost by $32,000-40,000 annually. But hardware becomes obsolete technologically long before 5 years. Betting on 5-year hardware lifespans is risky.
Personnel Cost Reduction If developers can manage clusters with 0.5 FTE instead of 1 FTE, saves $60,000-90,000 annually. This requires exceptional automation. Most teams need dedicated ops engineers.
Volume GPU Purchase Discount (10% savings) Reduces GPU acquisition from $280,000 to $252,000. Saves $9,333 in annual amortization. Meaningful but not game-changing.
Combined Optimistic Scenario Even with power optimization, 3-year lifespan, 75% utilization, and 0.5 FTE personnel, self-hosted costs approximately $3.20/GPU/hour. Cloud at $2.69/hour remains cheaper, and cloud scales elastically without personnel bottlenecks.
Even with optimistic assumptions, cloud rental remains competitive. Most teams benefit from cloud infrastructure for flexibility and risk reduction.
Capital Preservation and Financial Flexibility
Cloud rental preserves capital. An organization with $250,000 of available capital can deploy 8 H100 GPUs for one year using cloud ($2.69 × 8 × 8,760 = $188,179), then pivot to different workloads.
Self-hosted deployment consumes the full capital, removing flexibility. If priorities change, the sunk infrastructure investment constrains options painfully.
This flexibility matters operationally. Market conditions, model developments, and application requirements shift unpredictably. Cloud infrastructure accommodates change; self-hosted requires sunken investment.
Economies of Scale at Large Deployments
Teams deploying 1,000+ GPUs approach self-hosted economics viability. Per-GPU costs approach cloud pricing when amortizing substantial infrastructure investments. At 1,000 GPUs, facility costs drop dramatically per unit. Networking becomes more efficient. Cooling systems approach theoretical efficiency limits.
Dedicated data centers, negotiated power rates ($0.04-0.06/kWh in optimal locations), and large-scale workforce optimization push self-hosted closer to parity with cloud providers.
However, cloud providers also scale; their economies of scale match or exceed large deployments. AWS, Google Cloud, and Meta operate data centers at 10,000+ GPU scale. Their purchasing power exceeds almost all independent operators. They negotiate better NVIDIA pricing, better power rates, and amortize infrastructure across more users.
The Trillion-Dollar Question: As GPUs become more commoditized and cloud providers build more capacity, does self-hosted ever make sense again? Currently (March 2026), the answer is "only at massive scale with very favorable circumstances." For teams that aren't hyperscalers, cloud remains the smart choice.
FAQ
Q: At what scale does self-hosted become economical? A: Self-hosted approaches parity at 100-200 continuous GPUs with 70%+ utilization. Most teams never reach this scale. Those that do benefit significantly from self-hosting, saving 20-30% annually.
Q: What if I own real estate already? A: Space cost becomes near-zero, improving self-hosted economics. A team with empty warehouse space might save $50,000+ annually in facility costs. But power, cooling, and networking remain substantial.
Q: How does GPU depreciation affect the equation? A: H100s may become obsolete in 2-3 years if H200 or B200 becomes standard. Buying today risks hardware becoming unmarketable in 18 months. Cloud rental transfers depreciation risk to providers.
Q: Can I mix cloud and self-hosted? A: Yes. Teams often run baseline workloads on owned infrastructure and burst to cloud for peak demand. This hybrid approach optimizes both cost and flexibility.
Q: What happens if my utilization drops from 80% to 40%? A: Self-hosted costs remain fixed (still paying for all infrastructure). Cloud costs drop 50%. This flexibility matters when workloads shift unexpectedly.
Related Resources
- RunPod GPU Pricing
- Lambda Labs Pricing
- CoreWeave Pricing
- NVIDIA H100 Pricing
- NVIDIA A100 Pricing
- GPU Provider Comparison
Conclusion and Recommendation
Self-hosted AI data center economics rarely beat cloud infrastructure for small and medium deployments. The $2.69-3.78 per GPU per hour cloud pricing defeats 3-5 year amortized self-hosted costs at realistic utilization rates. The gap widens when considering infrastructure risk, support costs, and operational complexity.
Large-scale operations (100+ continuous GPUs) start approaching self-hosted economics viability. Very large operations (1,000+ GPUs) justify dedicated infrastructure, particularly with favorable power and real estate access. But even at scale, cloud providers enjoy similar economies, keeping pricing competitive.
For most teams, cloud GPU rental eliminates infrastructure burden while preserving capital and flexibility. Only sustained, high-utilization, multi-year workloads justify self-hosted infrastructure investment. The decision requires analyzing specific cost structures, anticipated utilization, and risk tolerance.
Evaluate total cost of ownership carefully before building on-premise capacity. Cloud economics typically win decisively. As of March 2026, cloud infrastructure remains the default choice for teams lacking 100+ GPU scale and favorable power/space economics.
Sources
- NVIDIA GPU pricing and OEM costs (2024-2026)
- AWS, Google Cloud, and Azure power cost data
- Data center facility cost benchmarks
- RunPod, Lambda, CoreWeave pricing (March 2026)
- DeployBase GPU infrastructure analysis
- Industry GPU workload utilization studies