H100 Paperspace: Pricing, Gradient Notebooks, and Limited Availability

Deploybase · February 8, 2025 · GPU Pricing

Contents

H100 Paperspace: Gradient Notebooks and Limited Access

H100 Paperspace (DigitalOcean subsidiary) offers H100 GPU access through Gradient notebooks and managed containers, though availability is significantly constrained compared to dedicated providers. Pricing varies regionally ($3.00-4.50/hr typical), and H100 capacity rarely exceeds 10-20 globally available instances at any time.

This guide covers Paperspace's H100 offerings, the Gradient notebook environment, availability management, and when to choose alternative providers.

Paperspace Pricing and Availability

Paperspace's pricing structure emphasizes simplicity but suffers from chronic H100 scarcity.

H100 Pricing Tiers and Monthly Analysis

PlanHourlyMonthly (730 hrs)StorageBest UseEffective Annual
H100 On-Demand$3.50-4.50$2,555-3,28520GBTemporary experiments$30,660-39,420
H100 Monthly$2,200-2,700Committed100GB30-day commitments$26,400-32,400
H100 AnnualUnavailableN/AN/ANot offeredN/A

Monthly plans provide approximately 20-30% discount versus hourly on-demand, but annual reservations are not available, limiting predictable long-term budgeting. Comparing effective costs: on-demand at $4.00/hr = $39,420/year, versus monthly commitment ~$2,450 = $29,400/year (25% savings).

Availability and Capacity Constraints

Paperspace maintains roughly 5-10 H100 instances across all regions, with significant availability variance:

Time WindowAvailable InstancesWait TimeAvailability Rate
Off-Peak (0-4 UTC)4-80-1 hour60-80%
Standard (8-20 UTC)2-42-4 hours30-50%
Peak (12-17 UTC)0-2Fully booked5-20%

For comparison, RunPod lists 200+ H100 instances with 95%+ availability, and Lambda maintains dedicated guaranteed capacity. Paperspace's scarcity makes it unsuitable for time-sensitive workloads.

Performance Benchmarks on Paperspace H100

Inference Throughput

Paperspace H100 achieves standard throughput for single H100 hardware:

ModelConfigurationThroughput
Llama-2 70BFull Precision40-50 tokens/sec
Mistral 7BFull Precision70-80 tokens/sec
Both4-bit Quantization2-3x improvement

Performance matches other providers (RunPod, Lambda) since underlying H100 hardware is identical.

Detailed Setup and Gradient Notebook Environment

IDE and Development Interface

Paperspace's Gradient provides a Jupyter-like environment accessible via browser. Key features:

  • Pre-installed PyTorch, TensorFlow, JAX with CUDA 12.2
  • File browser for dataset management
  • Terminal access for custom package installation
  • Git integration for version control
  • Collaborative notebook sharing with team members
import torch
from transformers import AutoModelForCausalLM

model = AutoModelForCausalLM.from_pretrained("gpt2")
model = model.to("cuda")

Launching Paperspace H100 Notebooks: Complete Walkthrough

  1. Access Paperspace console at https://www.paperspace.com/console
  2. Handle to "Notebooks" section
  3. Click "Create Notebook"
  4. Select Machine:
    • Filter by GPU: Select "H100"
    • Note: May show "Limited Availability" (0-10 instances global)
    • If unavailable, pre-register for waitlist notification
  5. Select Template:
    • PyTorch (recommended for training)
    • TensorFlow (if TensorFlow-specific)
    • Custom (for specialized packages)
  6. Configure:
    • Notebook name
    • Machine type: H100 (only option if available)
    • Billing: Hourly or Monthly (monthly provides 25% discount)
  7. Click "Start" and wait 2-5 minutes for provisioning
  8. Once running, access Jupyter IDE through browser

Storage and Persistence Strategy

Gradient provides 20GB (hourly) or 100GB (monthly) persistent storage. Optimal usage:

df -h /storage

import subprocess
subprocess.run(['aws', 's3', 'cp', 's3://bucket/data.tar.gz', '/tmp/'])
subprocess.run(['tar', '-xzf', '/tmp/data.tar.gz', '-C', '/storage/'])

For datasets exceeding 100GB, Paperspace's storage becomes impractical ($0.10/GB/day). Instead: save only final checkpoints to /storage, keep working data on instance ephemeral storage (/tmp), or use S3 integration for external storage.

import subprocess
subprocess.run(["aws", "s3", "cp",
                "s3://my-bucket/dataset.tar.gz", "."])
subprocess.run(["tar", "-xzf", "dataset.tar.gz"])

Cost Optimization Strategies for Paperspace H100

Pricing Tier Selection Guide

Calculate optimal pricing based on projected usage:

on_demand_hourly = 4.00
monthly_plan_price = 2,500
hours_needed_for_breakeven = monthly_plan_price / on_demand_hourly
if hours_needed <= 625:
    recommended = "On-Demand ($4/hr)"
    monthly_cost = hours_needed * on_demand_hourly
else:
    recommended = "Monthly Plan ($2,500)"
    monthly_cost = 2500

For projects exceeding 600 hours/month, monthly plans provide 25-30% savings.

Notebook-Based Workflows

Gradient suits interactive development:

  • Model fine-tuning with iterative experimentation
  • Dataset exploration and preprocessing
  • Inference testing with smaller batches
  • ML research prototyping

Avoid production deployment directly on Paperspace. Instead, develop locally, test on Gradient, then deploy to dedicated inference platforms.

Alternative Provider Fallback

Paperspace's availability constraints make fallback strategies essential:

providers = {
    "paperspace": {"cost": 4.00, "availability": 0.30},
    "runpod": {"cost": 2.69, "availability": 0.95},
    "lambda": {"cost": 3.78, "availability": 1.00}
}

selected = min(providers.items(),
    key=lambda x: x[1]["cost"] / max(x[1]["availability"], 0.1))

Session Management

Paperspace sessions terminate after 6-hour inactivity. Enable periodic activity to maintain longer sessions, or structure workflows as time-bound scripts.

Cost Optimization Strategies

Strategic Pricing Choice: On-Demand vs Monthly

Calculate optimal pricing tier based on expected usage:

  • Quick experiments (<4 hours): Use on-demand at $4.00/hr max
  • Week-long projects (40-100 hours): Use on-demand, avoid monthly commitment
  • Month-long projects (200+ hours): Monthly plan at ~$2,450/month provides 25-30% savings

Monthly commitment breakeven: ($2,450 / $4.00/hr) = 612.5 hours. For usage under 612 hours, on-demand is cheaper. For sustained 30-day usage (720 hours), monthly saves $950/month.

Availability Monitoring and Scheduling

Check availability before planning H100-dependent work. Paperspace's console shows real-time capacity per region. Availability patterns:

  • Morning UTC (8-17): Typically zero to low availability (demand peak in US time zones)
  • Late night UTC (18-6): 60-80% availability for European researchers
  • Weekends: Slightly improved availability (15-20% higher than weekdays)

Schedule compute-intensive tasks for off-peak windows to improve instance launch probability.

Fallback Strategy with Multi-Provider Deployment

Always have alternative providers identified. If Paperspace H100 unavailable, use:

Maintain flexible deployment processes supporting multiple backends:

import sys

providers = [
    ("paperspace", "https://paperspace.com", "limit_5_instances"),
    ("runpod", "https://runpod.io", "availability_200+"),
    ("lambda", "https://lambdalabs.com", "availability_guaranteed"),
]

def find_available_provider():
    for name, url, capacity in providers:
        try:
            response = requests.get(f"{url}/api/availability")
            if response.json()["available"]:
                return name
        except:
            continue
    # Default to RunPod if primary unavailable
    return "runpod"

Comparing Paperspace to Alternatives

Cost and Availability Trade-off

Paperspace's theoretical hourly rate ($3.50-4.50) appears reasonable, but scarcity makes effective cost higher. Calculate true cost including opportunity cost of waiting:

ScenarioPaperspace Effective CostAlternative Provider
Instance available immediately$4.00/hrRunPod: $2.69/hr
Wait 2 hours for availability$4.00 + (2 hrs × $0) + opportunity costRunPod: $2.69/hr (no wait)
Instance never availableInfinite (effectively $0 work)RunPod: $2.69/hr

For projects where time matters, RunPod's guaranteed availability at $1.99-2.69/hr provides more predictable compute planning despite higher nominal rate. Lambda's reserved multi-GPU clusters eliminate availability concerns entirely.

Detailed Provider Comparison

FeaturePaperspaceRunPodLambdaCoreWeave
H100 Availability5-20%95%+100%100%
Hourly Rate$3.50-4.50$2.69$3.78 SXM / $2.86 PCIe$6.16 (8xH100)
NotebooksExcellentCommunityN/AN/A
Setup Time<5 min<5 min<10 min<15 min
Multi-GPU SupportNoLimitedYes (2x-8x)Yes (8x native)
SupportEmailCommunityPhone/ChatEmail/Chat

Paperspace excels for interactive development when instances available. For production workloads, RunPod or Lambda are more practical despite higher costs on paper when accounting for availability and developer time.

Use Cases Favoring Paperspace

  • Interactive notebook-based research with immediate feedback
  • Brief experimentation (1-4 hour sessions) when instance available
  • Team collaboration through Gradient notebook sharing
  • Integration with DigitalOcean infrastructure (if already using platform)

Use Cases Requiring Alternative Providers

  • Sustained training jobs (>12 hours)
  • Scheduled batch processing
  • Production inference APIs
  • Multi-GPU distributed training

See the multi-GPU training guide for providers supporting coordinated clusters.

Technical Deep Dives

HBM3 Memory Management

H100's 80GB HBM3 provides exceptional bandwidth (3.3 TB/s) but limited to this single instance. Monitor memory carefully:

import torch
print(f"Allocated: {torch.cuda.memory_allocated() / 1e9:.2f}GB")
print(f"Reserved: {torch.cuda.memory_reserved() / 1e9:.2f}GB")
torch.cuda.reset_peak_memory_stats()

For models exceeding 50GB, quantization or sparse attention becomes necessary.

CUDA Architecture

Paperspace's H100 instances use NVIDIA's base CUDA 12.2 image. Confirm compatibility before deploying custom CUDA kernels:

nvcc --version  # Check CUDA version
nvidia-smi      # Verify GPU detection

FAQ

Should I commit to Paperspace's monthly H100 plan?

Only if requiring guaranteed H100 access within a one-month window. Monthly commitment ($2,200-2,700) provides modest discount versus hourly rates. Better to reserve with Lambda (15% discount for 6-month plans) or use RunPod spot pricing if availability uncertainty is acceptable.

How does Paperspace H100 compare to Gradient's other GPU options?

Paperspace offers A100 and A6000 alternatives with better availability. H100 provides ~2x A100's FP32 throughput, important for large model inference. For models under 30B parameters, A100 offers superior availability-to-cost ratio on Paperspace.

Can I export notebooks and deploy elsewhere?

Yes. Download Gradient notebooks as .ipynb files and deploy to other platforms with identical code. This portability is Paperspace's key advantage for development environments - experiment on Paperspace, deploy to production on RunPod or Lambda.

How should I handle Paperspace's 6-hour inactivity timeout for long-running jobs?

Paperspace terminates idle sessions after 6 hours. For longer training jobs: (1) disable sleep/screensaver on local machine and maintain SSH tunnel with periodic activity, (2) implement periodic checkpoint saves and prints to keep session active, or (3) migrate to always-on providers like RunPod/Lambda. Most practical: use Paperspace for development/testing only, then submit actual training jobs to RunPod or Lambda for 24/7 execution.

When should I avoid Paperspace H100 and choose alternatives instead?

Avoid Paperspace for: production inference (use RunPod), sustained training (use Lambda reserved), cost-sensitive batch work (use Vast.AI), and time-critical projects (availability too constrained). Paperspace excels for: interactive development, quick experiments, notebook-based research, and educational environments where availability constraints are acceptable.

Sources