Terrestrial GPU Annual Failure Rate

What is the annual rate at which datacenter GPUs (H100/A100-class) permanently fail — requiring physical replacement — in terrestrial operation?

What is the annual permanent failure rate of datacenter GPUs (H100/A100-class) in terrestrial operation?

Answer

The annual permanent GPU failure rate for H100-class accelerators in terrestrial datacenters ranges from 2.5% (optimistic) to 6% (conservative), with a central estimate of 4%. These rates cover only permanent failures requiring physical GPU replacement — the rate relevant for orbital deployment where failed GPUs cannot be hot-swapped.

Separately:

Analysis

Distinguishing Failure Categories

The most critical contribution of this page is separating failure types that have fundamentally different operational implications:

For the orbital model, only permanent failures matter. Transient failures cause downtime (reducing availability) but do not permanently reduce the GPU fleet.

Bounding the Permanent Failure Rate

Upper bound (~6%): all "Faulty GPU" events are permanent. Meta's 148 "Faulty GPU" interruptions over 54 days on 16,384 GPUs annualize to 6.1% meta-llama3-paper.2. This is an upper bound because Meta's automation handled nearly all failures with only 3 manual interventions meta-llama3-paper.4, implying many "Faulty GPU" events were resolved without physical replacement.

Middle estimate (~4%): overprovisioning convergence. The Delta study recommends 5% overprovisioning to maintain 99.9% job availability given a 2.2-hour recovery time cui-two-gpus-2025.6. If physical replacement takes 3–5 days nonuniform-tensor-parallelism.1, a 5% spare pool sustains a permanent replacement rate of roughly 3–5%/year. The NTP paper's finding that clusters spend 81% of time with >0.1% GPUs failed nonuniform-tensor-parallelism.2 corroborates this range.

Lower bound (~2.5%): lemon node analysis + H100 hardware improvements. Meta's lemon node analysis identifies 1.2% of the fleet as chronically defective, with GPUs as the root cause in 28.2% of cases revisiting-ml-cluster-reliability.3. The Delta study found dramatic H100 hardware improvements: zero PMU, GPU Fallen Off Bus, and NVLink errors across 146 days on 608 GPUs cui-two-gpus-2025.4. This suggests the permanent non-memory GPU failure rate is substantially lower than the "Faulty GPU" interrupt rate.

Cross-check: row remapping failures. The Delta study observed 8 row remapping failures on 608 H100 GPUs in 146 days cui-two-gpus-2025.7, annualizing to 3.3%/year for this specific permanent failure mode. This provides a floor on permanent failures from the HBM degradation pathway alone, consistent with the central estimate.

Cross-check: depreciation evidence. Hyperscalers depreciate GPUs over 5–6 years gpu-useful-life-2025.1, and Google's TPUs operate at 100% after 7–8 years gpu-useful-life-2025.2. A 4% permanent annual failure rate means ~81% of GPUs survive to year 5 without replacement — consistent with managed fleets where replacements maintain capacity.

The Google 1–3 year claim trendforce-gpu-lifespan-2024.1 likely refers to economic useful life (when GPUs become obsolescent) rather than physical failure. This is consistent with NVIDIA's 1-year release cadence making GPUs economically obsolete in 2–3 years even while physically functional.

HBM as a Distinct Failure Mode

HBM deserves separate treatment because it follows a distinct degradation pattern:

  1. High raw error rate: H100's per-GPU MTBE for uncorrectable ECC errors is 88,768 hours (~10%/year annualized), 3.2x worse than A100 cui-two-gpus-2025.2.
  2. Effective but finite mitigation: Row remapping succeeds 92% of the time cui-two-gpus-2025.3, reducing the unmitigated rate to ~0.8%/year. But spare rows (capped at 512) are consumed progressively.
  3. Eventual permanent failure: Once spare rows are exhausted, the next uncorrectable error becomes permanent. Eight RRFs in 146 days on 608 GPUs cui-two-gpus-2025.7 show this transition already occurring.
  4. Progressive degradation marker: Microsoft found that >10 correctable errors on a row increases regression probability by 77.8% microsoft-superbench.3, confirming that memory errors accelerate toward permanent failure.

For orbital deployment, HBM degradation is compounded by radiation effects (additional uncorrectable errors consuming spare rows faster), making the HBM failure trajectory a key driver of GPU attrition in orbit.

Overprovisioning vs. Capacity Decay

Terrestrial model: overprovisioning with active replacement. Operators maintain ~5% spare GPUs cui-two-gpus-2025.6, replace failed units within 3–5 days nonuniform-tensor-parallelism.1, and maintain dedicated spare buffers per customer nebius-fault-tolerant-2025.2. Automation detects and drains failed nodes with minimal human intervention meta-llama3-paper.4, and proactive lemon detection removes degrading hardware before it causes repeated failures revisiting-ml-cluster-reliability.3. Effective capacity stays at ~100%.

Orbital model: irreversible capacity decay. Failed GPUs in orbit cannot be hot-swapped. Cold spares add mass and cost, degrade from radiation and thermal cycling while idle, and once exhausted, every subsequent GPU failure permanently reduces fleet capacity. The permanent failure rate (~4% central) directly determines the capacity decay rate used in the effective lifetime integral.

The models are complementary, not competing. Terrestrial operators face the same underlying permanent failure rate but can compensate via replacement. The distinction matters because the orbital model should use the permanent failure rate (capacity that cannot be recovered), not the total interruption rate (which includes transient events that terrestrial operators recover from within hours).

Combination Method

The central estimate of 4% was derived by triangulating five independent data sources:

  1. Meta Llama 3 primary data meta-llama3-paper.2: 148 "Faulty GPU" events in 54 days on 16,384 H100s → 6.1% annualized upper bound on permanent GPU failures, discounted by the automation evidence meta-llama3-paper.4.
  2. NCSA Delta longitudinal study cui-two-gpus-2025.6: 5% overprovisioning recommendation, anchoring the steady-state failure-to-replacement ratio.
  3. Meta ML cluster reliability revisiting-ml-cluster-reliability.3: 1.2% fleet chronically defective, 28.2% GPU-caused, with explicit transient vs. permanent taxonomy revisiting-ml-cluster-reliability.2.
  4. NTP paper nonuniform-tensor-parallelism.1: 3–5 day physical replacement timeline constraining the replacement throughput of a 5% spare pool.
  5. Microsoft SuperBench microsoft-superbench.1: 10.36% node defect rate during A100 cluster build-out (includes performance degradation, not just hard failures).

The central value of 4% sits at the intersection of these constraints: above the 2.5% lower bound from lemon node analysis and H100 hardware improvements, below the 6% upper bound from treating all "Faulty GPU" events as permanent, and consistent with the 5% overprovisioning that the Delta study recommends for operational fleets.

Key Uncertainties

  1. No source directly reports the permanent GPU replacement rate. All estimates are derived indirectly from total interruption rates, overprovisioning recommendations, and lemon node analysis. The fraction of "Faulty GPU" interruptions that require physical replacement vs. automated recovery is the single largest uncertainty.

  2. H100 observation period is short. The Delta study's H100 data spans only 146 days cui-two-gpus-2025.1 — too short to observe wear-out failure modes or HBM spare row exhaustion at scale. The A100 data (895 days) is more reliable for long-term trends.

  3. Workload dependence. Meta's RSC-1 GPU swap rate is ~3x RSC-2 revisiting-ml-cluster-reliability.4, suggesting the failure rate depends heavily on workload intensity. The values here reflect training-class workloads at near-maximum utilization; inference workloads may be less severe.

  4. HBM degradation trajectory. The fixed spare row count (512) with increasing memory capacity means newer GPUs with more HBM (H100 96GB vs A100 40GB) may exhaust spares faster, causing an increasing permanent failure rate over time. This is not yet captured in the constant annual rate assumption.

Evidence

Long-Duration Field Studies

  1. The Delta HPC system at NCSA comprises 448 A100 GPUs (40 GB HBM2e) observed over 895 days (9.6 million GPU-hours) and 608 H100 GPUs (96 GB HBM3, GH200 Superchip) observed over 146 days (2.1 million GPU-hours), for a combined 11.7 million GPU-hours of operational data. — cui-two-gpus-2025

  2. H100 GPUs exhibit 3.2x lower per-GPU MTBE for uncorrectable ECC memory errors compared to A100: 88,768 hours per GPU for H100 vs 283,271 hours per GPU for A100. The per-GB MTBE is 8.5 million hours (H100 HBM3) vs 11.3 million hours (A100 HBM2e), a 24% reduction. — cui-two-gpus-2025

  3. H100 GPU memory error-recovery mechanisms (row remapping and error containment) successfully mitigate uncorrectable memory errors with a probability of 0.92 (92% success rate). Available spare rows for row remapping are capped at 512 rows, unchanged from A100 despite a 2.4x increase in memory capacity. — cui-two-gpus-2025

  4. H100 GPUs demonstrate significant improvements in GPU hardware resilience over A100: zero PMU SPI errors, zero GPU Fallen Off Bus errors, and zero NVLink errors were observed during the H100 operational period. GSP errors were reduced from 3,857 (A100, 895 days) to 3 (H100, 146 days), a >99.9% reduction. — cui-two-gpus-2025

  5. Overall node availability is approximately 99.4% for A100 GPUs and 99.3% for H100 GPUs, corresponding to 9–10 minutes of downtime per day. Mean recovery time after failure is 0.88 hours for A100 and 2.2 hours for H100. — cui-two-gpus-2025

  6. For a training job using 608 H100 GPUs with 2.2-hour recovery time, maintaining 99.9% job-level availability requires 5% overprovisioning (31 additional GPUs). If recovery time is reduced to 5 minutes, overprovisioning drops to 2%. — cui-two-gpus-2025

  7. Eight row remapping failures (RRFs) were observed on H100 GPUs during their 146-day operational period, indicating memory recovery failure due to exhaustion of spare memory rows. Zero RRFs were observed on A100 GPUs during their 895-day period. — cui-two-gpus-2025

Meta Llama 3 Primary Data

  1. During a 54-day snapshot of Llama 3 405B pre-training on 16,384 H100 GPUs (each at 700W TDP, 80GB HBM3), 466 total job interruptions occurred: 47 planned and 419 unexpected. — meta-llama3-paper

  2. Of the 419 unexpected interruptions, failure categories were: Faulty GPU 148 (30.1%), GPU HBM3 Memory 72 (17.2%), Software Bug 54 (12.9%), Network Switch/Cable 35 (8.4%), Unplanned Host Maintenance 32 (7.6%), GPU SRAM Memory 19 (4.5%), GPU System Processor 17 (4.1%), NIC 7 (1.7%), NCCL Watchdog Timeouts 7 (1.7%), Silent Data Corruption 6 (1.4%), GPU Thermal Interface/Sensor 6 (1.4%), SSD 3 (0.7%), Power Supply 3 (0.7%), Server Chassis 2 (0.5%). — meta-llama3-paper

  3. Approximately 78% of unexpected interruptions were attributed to confirmed or suspected hardware issues. GPU issues (all GPU-related categories combined) accounted for 58.7% of all unexpected issues. Despite 466 total interruptions, the system achieved >90% effective training time. — meta-llama3-paper

  4. Despite the large number of failures, significant manual intervention was required only three times during the 54-day period, with the rest handled by automation. — meta-llama3-paper

Meta Large-Cluster Reliability Study

  1. Meta's RSC-1 (16K GPUs) and RSC-2 (8K GPUs) A100 clusters, totaling 24K GPUs, experienced 4 million jobs over 11 months with 150+ million GPU-hours. Failure rates varied substantially over time, ranging from ~2.5 to ~17.5 failures per 1,000 node-days on RSC-1. — revisiting-ml-cluster-reliability

  2. Hardware errors are binned as transient (e.g., ECC error, link flap) or permanent (e.g., degraded hardware requiring repair or replacement by a vendor). — revisiting-ml-cluster-reliability

  3. "Lemon nodes" — servers causing repeating job failures not identifiable by standard health checks — represent 1.2% of RSC-1's footprint but are involved in 13% of daily jobs. Of lemon node root causes, 28.2% are GPUs, 20.5% DIMMs, 15.4% PCIe, 10.3% EUD, 7.7% BIOS, 7.7% NICs, 5.1% PSUs, 2.6% optics, 2.6% CPUs. — revisiting-ml-cluster-reliability

  4. GPUs are physically swapped in the cluster. RSC-1 GPUs are swapped at approximately 3x the rate of RSC-2, likely due to differing workloads. — revisiting-ml-cluster-reliability

Replacement Timelines and Overprovisioning

  1. Hardware failure recovery time is modeled as 3–5 days for physical GPU replacement, described as "perhaps on the low-side for replacing high-demand hardware." Software failure recovery takes approximately 3 hours. — nonuniform-tensor-parallelism

  2. Using Llama 3 failure rates, a GPU cluster spends 81% of its time with >0.1% of GPUs in a failed state. — nonuniform-tensor-parallelism

  3. The traditional DP-DROP fault tolerance method requires 90 spare racks to maintain full mini-batch size in a 32K GPU cluster, while Nonuniform Tensor Parallelism (NTP) reduces this to 2 spare DP replicas (16 racks). — nonuniform-tensor-parallelism

Industry Validation

  1. During a 90-day cluster build-out evaluation of 24k+ A100 GPUs (3k+ VMs) in Azure, SuperBench identified 10.36% of nodes as defective (exhibiting failure or performance regression). SuperBench has been deployed in Azure production for 2+ years, validating hundreds of thousands of GPUs. — microsoft-superbench

  2. Without proactive validation, baseline MTBI in Azure A100 clusters was 17.5 hours. In simulation, SuperBench's Selector increased MTBI to 22.61x baseline (262.05 hours). 38.1% of incidents previously required more than one day to resolve. — microsoft-superbench

  3. On A100 GPUs, row remapping with more than 10 correctable errors shows a 77.8% higher chance of end-to-end workload regression compared to 1-10 correctable errors, indicating correctable errors are a precursor to permanent degradation. — microsoft-superbench

  4. Nebius reports peak MTBF of 56.6 hours (169,800 GPU-hours) on a 3,000-GPU (375-node) production cluster and average MTBF of 33.0 hours. Average MTTR is 12 minutes across most installations. — nebius-fault-tolerant-2025

  5. Nebius maintains a "dedicated spare buffer of GPU capacity for each customer" to ensure quick provisioning of replacement nodes; faulty nodes are automatically drained and replaced with healthy spares. — nebius-fault-tolerant-2025

GPU Longevity

  1. Hyperscaler GPU depreciation schedules have been extended: Microsoft and Google from 4 to 6 years, Meta from 4.0 to 5.5 years, Oracle from 4 to 6 years. Amazon reversed from 6 back to 5 years in February 2025, incurring a $700 million operating income hit. — gpu-useful-life-2025

  2. Google's custom TPUs have been maintained at 100% utilization after 7–8 years of operation. Azure ran K80-class GPUs in production for 9 years (2014–2023) and P100s for 7 years (2016–2023). — gpu-useful-life-2025

  3. A "general architect at Alphabet" (unnamed) stated that datacenter GPUs last 1–3 years at 60–70% utilization, with 1–2 years typical under heavy AI workloads. TrendForce notes this "cannot be considered 100% accurate and requires further confirmation." — trendforce-gpu-lifespan-2024

Existing Sources (refined interpretation)

  1. Epoch AI calculates MTBF of ~50,000 GPU-hours for H100 GPUs based on Meta's Llama 3 data, treating all 419 unexpected interruptions (hardware + software + network) as "failures." This conflates permanent hardware failures with transient software/network issues. — epoch-gpu-failures

  2. Tom's Hardware reports 148 GPU hardware failures and 72 HBM3 failures across 16,384 H100 GPUs in 54 days. The 148 "GPU failures" annualize to ~6.1%. This is the journalism layer's interpretation of the primary data in the Meta Llama 3 paper [meta-llama3-paper.2]. — meta-llama3-failures