H100 Lambda Labs: Pricing, Reserved Capacity, and Multi-GPU Setups

Deploybase · February 5, 2025 · GPU Pricing

Contents

H100 Lambda Labs: Dedicated GPU Compute

H100 Lambda Labs runs $2.86/hr (PCIe) or $3.78/hr (SXM). Dedicated capacity, no spot terminations, NVLink multi-GPU setups. Good for production training that can't handle interruptions.

This covers pricing, reserved discounts, multi-GPU setup, and how it stacks against other providers.

Lambda Labs Pricing Overview

As of March 2026, Lambda Labs operates a straightforward pricing model without spot market variability:

On-Demand H100 Pricing

ConfigurationHourly RateMonthly Equivalent (730 hrs)AnnualPer-Token (50 tokens/sec)
H100 PCIe (1x)$2.86$2,088$25,056$0.0159
H100 SXM (1x)$3.78$2,759$33,113$0.0210
H100 PCIe (2x)$5.72$4,176$50,112$0.0318
H100 SXM (8x)$27.52$20,090$241,075$0.1529

Reserved Pricing Model and Discounts

Lambda offers prepaid contracts with significant savings across multiple term lengths:

TermSingle H100 PCIeSingle H100 SXM8x H100 SXMEffective Discount
On-Demand$2.86$3.78$27.52Baseline
3-Month Reserved$2.57$3.40$24.7710%
6-Month Reserved$2.43$3.21$23.3915%
12-Month Reserved$2.29$3.02$22.0220%

A single H100 PCIe reserved for 12 months costs approximately $20,044 annually (versus $25,056 on-demand), saving $5,012 per instance. For 8x SXM clusters, 12-month reservations save approximately $43,000 annually compared to on-demand.

Performance Benchmarks

Inference Throughput

Lambda's dedicated infrastructure delivers consistent numbers:

ModelParametersBatch SizePCIe ThroughputSXM ThroughputImprovement
Mistral7B162-72 tokens/sec68-78 tokens/sec+9%
Llama-213B142-52 tokens/sec48-58 tokens/sec+13%
Llama-270B132-42 tokens/sec40-50 tokens/sec+18%

NVLink provides proportionally higher gains on larger models due to reduced memory bandwidth bottlenecks.

Distributed Training Performance

Multi-GPU clusters show excellent scaling efficiency:

| Configuration | Throughput (13B Model) | All-Reduce Overhead | Scaling Efficiency | |---|---|---|---|---| | 1x H100 SXM | 550 tokens/sec | N/A | 100% | | 2x H100 SXM | 1,050 tokens/sec | 4.5% | 95.5% | | 4x H100 SXM | 2,080 tokens/sec | 5.2% | 94.8% | | 8x H100 SXM | 4,100 tokens/sec | 6.0% | 94.0% |

Strong scaling demonstrates NVLink efficiency for distributed training.

Instance Configuration and Setup

Launching H100 Instances: Complete Walkthrough

  1. Access Lambda cloud console at https://cloud.lambdalabs.com
  2. Click "Launch Instance" in the dashboard
  3. Select GPU configuration:
    • Single H100: Choose PCIe or SXM variant
    • Multi-GPU: Select 2x, 4x, or 8x clusters (pre-configured with NVLink)
  4. Choose region: US-West (lowest latency for US-based users), US-East, or Europe
  5. Select instance type from dropdown (instance names indicate GPU config)
  6. Choose filesystem: 20GB default, or custom SSD (costs ~$0.10/GB/day)
  7. Optionally attach persistent storage volume for model checkpoints
  8. Add SSH public key (create new pair if first time)
  9. Review instance specs and pricing
  10. Click "Launch" and wait 3-8 minutes for provisioning

Network Configuration and SSH Access

Lambda assigns static public IPs upon instance provisioning. Configure SSH access:

Host lambda-h100
    HostName <lambda-assigned-ip>
    User ubuntu
    IdentityFile ~/.ssh/lambda_key
    ServerAliveInterval 60

ssh lambda-h100

nvidia-smi

Default firewall rules permit SSH (port 22) and HTTPS (port 443). For custom application ports, configure firewall rules through Lambda dashboard under Instance Settings > Firewall.

Multi-GPU setups use NVLink for sub-microsecond GPU-to-GPU latency. No network stack involved.

Storage Integration and Dataset Management

Lambda offers persistent block storage separate from compute pricing:

aws s3 cp s3://my-bucket/dataset.tar.gz /tmp/
tar -xzf /tmp/dataset.tar.gz -C /root/data/
rm /tmp/dataset.tar.gz

mount_point="/mnt/persistent"
cp final_model.pt $mount_point/

A 500GB SSD persistent volume costs ~$3.65/day. If using volume for 10 days, pre-download at startup saves $30+ while requiring only 5 minutes additional setup time.

Multi-GPU Configurations and Distributed Training

Lambda pre-configures multi-GPU clusters with NVLink. 2x H100 SXM is $7.34/hr, 4x is $14.20/hr. No manual network setup.

Large-Scale Training

For training 70B-parameter models with full precision, 4x H100 clusters are standard. For 200B+ parameter models, 8x H100 clusters provide necessary bandwidth and memory parallelism.

import torch.distributed as dist
import torch.multiprocessing as mp

world_size = 8  # 8x H100 setup
rank = int(os.environ['RANK'])
dist.init_process_group('nccl')

Performance Characteristics

Bandwidth and Throughput

H100 SXM delivers 900 GB/s NVLink bandwidth between GPUs, approximately 14x higher than PCIe. For transformer models exceeding 30B parameters, SXM variants show 35-45% throughput improvement over PCIe-based setups due to reduced communication overhead.

Memory Utilization

Each H100 provides 80GB HBM3 memory. Across 8x H100 clusters, total GPU memory reaches 640GB, sufficient for most production models without memory optimization techniques.

Comparing Lambda to Other Providers

Lambda's H100 SXM at $3.78/hr is more expensive than RunPod H100 SXM at $2.69/hr, but below CoreWeave's cluster pricing ($6.16/GPU). The H100 PCIe at $2.86/hr fits standard server slots. The key trade-off: Lambda offers guaranteed dedicated capacity and simplified multi-GPU setups, whereas RunPod prioritizes lowest single-GPU cost on the spot market.

Vast.AI runs $2.50-4.00/hr but no multi-GPU cluster guarantees. Lambda wins if teams need coordinated distributed training with stable inter-GPU latency.

Cost Optimization Tactics

Right-Sizing the Cluster

Measure actual GPU utilization before committing to multi-GPU setups. Many training jobs achieve <70% GPU utilization, meaning a 2x cluster performs little better than a single H100 with gradient accumulation. Profiling methodology:

watch -n 1 'nvidia-smi --query-gpu=utilization.gpu,utilization.memory --format=csv,noheader'

Reservation Strategy and ROI

Lambda's reservations become cost-optimal when:

  • 3-month reservation: Break-even at 150 hours usage (saves $111 vs on-demand)
  • 6-month reservation: Break-even at 300 hours usage (saves $392 vs on-demand)
  • 12-month reservation: Break-even at 600 hours usage (saves $1,230 vs on-demand)

For production inference running 24/7 continuously, 12-month reservations save ~$4,708/year per H100 PCIe. Calculate ROI:

ROI breakeven = (Reserved Cost - On-Demand Cost) / (On-Demand Rate - Reserved Rate)
ROI for single H100 PCIe 12-month = $20,044 / ($2.86 - $2.29) = 351 hours
351 / 24 = 14.6 days continuous usage = ROI achieved in 2 weeks

For workloads running >2 weeks continuously, 12-month reservations are financially optimal.

Batch Inference Optimization

Consolidate inference requests into batches of 32-64 to maximize H100 throughput and minimize per-request cost:

Batch SizeThroughputCost per 1K Tokens
140 tokens/sec$0.0265
8200 tokens/sec$0.0053
32350 tokens/sec$0.0030
64380 tokens/sec$0.0028

A 70B-parameter model serving single requests costs 9x more per token than batch size 32. Implement request batching in inference server:

from fastapi import FastAPI
import asyncio

app = FastAPI()
request_queue = asyncio.Queue()

async def batch_processor():
    while True:
        batch = []
        # Collect up to 32 requests within 500ms timeout
        try:
            for _ in range(32):
                request = await asyncio.wait_for(
                    request_queue.get(), timeout=0.5
                )
                batch.append(request)
        except asyncio.TimeoutError:
            pass

        if batch:
            results = model.generate_batch([r.prompt for r in batch])
            for request, result in zip(batch, results):
                request.response_queue.put(result)

Persistent Volume Cost Management

Lambda's storage costs accumulate quickly. Optimization strategies:

  1. Download datasets at startup: Save $3/day per 500GB dataset by pre-downloading instead of persistent storage
  2. Compress checkpoints: 4-bit quantized models are 95% smaller than full precision, reducing persistent volume needs
  3. S3 integration: Keep checkpoints on S3 ($0.023/GB/month) instead of persistent volumes ($3/GB/month = 130x more expensive)

Example cost savings for 100GB checkpoint:

  • Persistent volume: $100 × $3/day = $300/day ($9,000/month)
  • S3 storage: $100 × $0.023/month = $2.30/month
  • Savings: $8,997.70/month by using S3

Multi-Instance Load Balancing

For production inference, distribute load across multiple Lambda H100 instances using round-robin load balancer:

import requests
from itertools import cycle

class LambdaLoadBalancer:
    def __init__(self, endpoints):
        self.endpoints = cycle(endpoints)

    def generate(self, prompt):
        endpoint = next(self.endpoints)
        response = requests.post(
            f"http://{endpoint}:8000/generate",
            json={"prompt": prompt}
        )
        return response.json()

balancer = LambdaLoadBalancer([
    "lambda-h100-1:8000",
    "lambda-h100-2:8000",
    "lambda-h100-3:8000"
])

This distributes load and provides automatic failover if one instance fails.

Troubleshooting Common Issues

NVLink Bandwidth Underutilization

Multi-GPU setups sometimes fail to saturate NVLink due to suboptimal ring-allreduce patterns or synchronous training. Implement async communication or use gradient accumulation to hide communication latency.

High Memory Fragmentation

Long-running training jobs can experience memory fragmentation causing OOM errors despite available memory. Enable PyTorch's garbage collection and periodically empty CUDA caches:

torch.cuda.empty_cache()
torch.cuda.synchronize()

SSH Connection Timeouts

Lambda instances behind firewalls may require explicit VPN or bastion host access. Configure SSH keepalive settings in ~/.ssh/config to maintain connections across network transitions. For full H100 specifications and cloud GPU pricing comparisons, check the dedicated reference pages.

FAQ

When should I use Lambda's reserved pricing versus on-demand?

Reserve capacity if committing 6+ continuous months. For 6 months of H100 PCIe usage, reserved saves approximately $2,354. If deployment duration is uncertain or varies seasonally, on-demand provides flexibility.

How does Lambda's multi-GPU setup compare to AWS p5 instances?

Lambda's 8x H100 SXM cluster costs $27.52/hr versus AWS p5.48xlarge at $55.04/hr on-demand (8xH100 SXM + CPU + storage). However, AWS instances include CPU, storage, and managed services. For pure GPU compute, Lambda is about 50% cheaper. See [AWS GPU pricing comparison](/gpus/models/nvidia-h100) for details.

Can I use Lambda instances for on-demand API serving with auto-scaling?

Lambda doesn't provide programmatic auto-scaling. Instances must be manually provisioned or managed through custom orchestration. For auto-scaling inference, consider RunPod's programmatic APIs or managed services like Replicate.

What's the optimal workflow for moving from single H100 to 4x cluster as model size grows?

Lambda supports incremental scaling without data loss. Workflow: (1) Train on single H100 PCIe ($2.86/hr), (2) periodically save full checkpoints to S3, (3) when model complexity requires more GPU, launch 4x H100 SXM cluster ($14.20/hr), (4) restore latest checkpoint and continue training with distributed training framework (PyTorch DDP). This approach minimizes upfront commitment while enabling smooth scaling as needs evolve.

How does Lambda's pricing compare to spot instances on other platforms?

Lambda has no spot pricing. All instances fixed-rate. But Lambda reserved at $3.02/hr beats AWS Spot at $3.68/GPU with guaranteed uptime.

Sources