Orbital Satellite Operational Lifetime
What is the effective capacity-weighted lifetime (years) of an orbital compute satellite, accounting for GPU degradation, catastrophic failures, and regulatory constraints?
Answer
The effective capacity-weighted lifetime of an orbital compute satellite — the number of full-capacity-equivalent years of service delivered per satellite — ranges from 2.2 years (conservative) to 5.9 years (optimistic), with a central estimate of 3.8 years. This is shorter than the physical operational lifetime because five independent degradation mechanisms progressively reduce delivered capacity, and because a deployment delay (ground testing, launch scheduling, orbital commissioning) consumes part of the GPU's economic life before the satellite begins producing revenue.
The effective lifetime is derived from a structured reliability model that separates and independently quantifies each degradation mechanism:
| # | Mechanism | Optimistic | Central | Conservative | Source / Derivation |
|---|---|---|---|---|---|
| 1 | Bus loss (whole-satellite catastrophic) | 0.5%/yr | 1.0%/yr | 2.5%/yr | Starlink fleet data + high-power EPS premium |
| 2 | GPU accelerator attrition (individual GPU failures) | 3.2%/yr | 6.8%/yr | 16.1%/yr | terrestrial-gpu-failure-rate base + space-hardware-failure-rate space additions |
| 3 | SDC / error-rate growth (gradual radiation degradation) | 1.3% fixed | 1.3% fixed | 1.3% fixed | Negligible at LEO TID doses; constant mitigation overhead |
| 4 | Economic obsolescence (GPU generational improvement) | Caps physical life at 7 yr | 5 yr | 3.5 yr | gpu-useful-life + operator design targets |
| 5 | Spares / graceful degradation (capacity recovery) | ~10% recovery | ~5% recovery | ~0% | Small; not included in headline values |
The effective lifetime integrates these components:
Effective lifetime = (1 − SDC_overhead) × ∫₀ᵀ [(1 − λ_bus)(1 − λ_gpu)]ᵗ dt
where T = productive operating window (economic life minus deployment delay), λ_bus = annual bus loss rate, and λ_gpu = annual GPU attrition rate.
| Scenario | Economic Life | Deployment Delay | Productive Window (T) | Bus Loss (λ_bus) | GPU Attrition (λ_gpu) | Combined Decay | SDC | Effective Lifetime |
|---|---|---|---|---|---|---|---|---|
| Optimistic | 7 yr | 3 mo | 6.75 yr | 0.5%/yr | 3.2%/yr | 3.7%/yr | 1.3% | 5.9 years |
| Central | 5 yr | 5 mo | 4.58 yr | 1.0%/yr | 6.8%/yr | 7.7%/yr | 1.3% | 3.8 years |
| Conservative | 3.5 yr | 6 mo | 3.0 yr | 2.5%/yr | 16.1%/yr | 18.2%/yr | 1.3% | 2.2 years |
Deployment delay represents the time between GPU procurement and first revenue: ground testing, integration, launch scheduling, and orbital commissioning. The GPU's economic obsolescence clock starts at procurement, not deployment. Estimates: 3 months (optimistic, SpaceX vertical integration with rapid turnaround), 5 months (central), 6 months (conservative, per Patel's estimate patel-2024-ai-bottlenecks.2). This delay reduces the productive operating window and thus the effective lifetime — a 5-month delay on a 5-year economic life reduces T by 8%.
Combined annual decay = 1 − (1−λ_bus)(1−λ_gpu). The multiplicative combination is slightly lower than the arithmetic sum because a failed satellite does not also experience GPU attrition. GPU attrition rates reflect the terrestrial permanent failure rate (2.5%/4%/6% by scenario, from [terrestrial-gpu-failure-rate](terrestrial-gpu-failure-rate.md)) plus space-specific additions (0.7%/2.8%/10.1%, from [space-hardware-failure-rate](space-hardware-failure-rate.md)). The spares/graceful degradation recovery factor (mechanism 5) is documented below but not included in headline values, as the effect is small (<5% change in effective lifetime).
Confidence note: The effective lifetime is a model output, not a directly measured quantity — no orbital compute satellite has accumulated operational lifetime data. The model's sub-inputs have very different evidence quality: physical lifetime and SDC overhead are well-grounded in fleet data and radiation testing; bus loss rate is empirically anchored but requires an engineering judgment premium for high-power systems; GPU attrition — especially the SEL component — is the weakest link, with the destructive SEL rate for H100/B200 spanning an engineering judgment across a >6 order-of-magnitude uncertainty range in the underlying SEL literature. This single sub-input dominates the overall uncertainty. The 2.2–5.9 year scenario range captures this uncertainty, but the central value of 3.8 years should be understood as a weakly informed indicative estimate — likely somewhere in the range of 2–6 years — not a precisely calibrated figure. Because effective lifetime is the analysis's dominant sensitivity lever, this uncertainty propagates strongly into all downstream TCO and parity results. See the Evidence Quality Assessment section below for a component-by-component decomposition.
Why effective lifetime rather than separate replacement opex: In the TCO model, amortizing capex over the effective lifetime captures the cost of maintaining fleet capacity without a separate opex line for failure-driven replacement. A satellite that delivers 3.8 effective kW_IT-years per kW_IT of capex requires investing capex/3.8 per year to sustain capacity — this implicitly includes the cost of replacing failed satellites at the rate needed to keep the fleet at target capacity.
Physical Lifetime Bounds
The physical operational lifetime before deorbit is bounded by a combination of component degradation, economic obsolescence, and operator design targets. The subsystem-level evidence shows that no single component is the binding physical constraint for a well-designed satellite at 5-7 years: modern Li-ion batteries provide 10-12 year capability (2.4x margin for 5 years) saft-ves16-leo-battery.1, solar panels retain 85-97% power at 7 years mdpi-leo-degradation.1, and modern reaction wheels demonstrate very high reliability (3M+ failure-free hours across 800+ units) newspace-systems-rw.1. The binding constraints are, in practice, economic obsolescence (GPU generational improvement) and EPS reliability risk (especially for high-power compute satellites).
- Optimistic: 7 years — Physically achievable. First-generation Iridium achieved 20+ years at 780 km (2.5x its 8-year design life), limited by fuel depletion, not component failure iridium-lifetime-extension-spacenews.2. Iridium NEXT (launched 2017-2019, now 7-9 years old) has zero reported failures and expects to operate to 2035+ iridium-lifetime-extension-spacenews.1. Multiple Starlink V1.0 satellites from November 2019 are still operating at 6.3+ years [mcDowell-starlink-stats.1]. Solar panel degradation at ~0.2-1.5%/yr in LEO leaves ~85-96% power at year 7. GPU obsolescence becomes increasingly binding: a chip launched in year 0 will be 3-4 generations behind by year 7.
- Central: 5 years — The most robustly supported estimate, converging from multiple independent lines of evidence: SpaceX's own filing (5-year operational life spacex-fcc-million-satellite-filing.1), Starcloud's stated 5-year lifespan starcloud-first-ai-model-space.1, Starlink's ~5-year designed lifespan starlink-deorbit-stats.1, hyperscaler GPU depreciation cycles (5-6 years gpu-depreciation-schedules.1), and the satellite infant mortality pattern (Weibull beta=0.45 means failure rate decreases beyond the first 1-2 years castet-saleh-2009-satellite-reliability.1). Radiation TID is not the binding constraint: behind 10mm Al shielding, the 5-year dose (~0.7 krad) is far below FinFET tolerance thresholds google-suncatcher.1.
- Conservative: 3.5 years — Captures first-generation design risk. SpaceX proactively retired over 500 early V1 Starlink satellites at less than 5 years of age after identifying a component flaw in ferrite transformers [starlink-retirement-cgaa.1, starlink-mass-deorbit-pcmag.1] — a cautionary data point that novel bus designs can harbor hidden mid-life failure modes. Also captures elevated solar activity increasing drag/propellant consumption starlink-orbit-lowering.1, and rapid GPU obsolescence (1-year Nvidia cadence means 3+ newer generations available).
Analysis
The Structured Reliability Model
The effective lifetime is computed from five independent mechanisms, each quantified separately with distinct data sources and mitigation paths. This structured approach replaces the earlier method of applying a qualitative space multiplier to terrestrial GPU failure rates.
The mathematical formulation:
Effective lifetime = (1 − SDC_overhead) × ∫₀ᵀ S(t) dt
where S(t) = (1 − λ_bus)ᵗ × (1 − λ_gpu)ᵗ is the surviving capacity fraction at time t, and T is the physical/economic lifetime ceiling.
The integral evaluates to:
∫₀ᵀ S(t) dt = [αᵀ − 1] / ln(α), where α = (1 − λ_bus)(1 − λ_gpu)
For the central case: α = (1 − 0.01)(1 − 0.068) = 0.99 × 0.932 = 0.9227, T = 5 years → ∫ = [0.9227⁵ − 1]/ln(0.9227) = [0.667 − 1]/(−0.0805) = 4.13. Multiplied by (1 − 0.013) = 0.987 → 4.1 effective years.
Component 1: Catastrophic Satellite Loss (Bus Failure)
Bus loss represents total satellite failure from non-GPU causes — EPS failure, ADCS failure, propulsion loss, debris impact, or software/firmware critical failure. When a satellite is lost, all GPUs aboard are lost simultaneously.
Empirical anchors (updated March 2026):
| Fleet | Annualized Loss Rate | Notes |
|---|---|---|
| OneWeb | ~0.05-0.08%/yr | 2 failures / 656 units over 4-7 years oneweb-stats-mcdowell.1 |
| Starlink V2 Mini | ~0.9%/yr | 92 failures / 6,927 units over ~1.5 yr avg age (still maturing) [mcDowell-starlink-stats.3] |
| Starlink Gen1 (unplanned) | ~1.5%/yr | 348 unplanned failures / 4,714 units over ~5 yr avg age (includes immature early batches) [mcDowell-starlink-stats.2] |
| Iridium NEXT (7-9 yr) | ~0%/yr | Zero reported failures across 80 satellites, 7-9 years of operation iridium-lifetime-extension-spacenews.1 |
| Smallsat 220-500 kg | ~1.0%/yr | 96% mission success rate, multi-year missions smallsat-reliability-spacenews-2020.1 |
The satellite reliability literature shows an infant mortality pattern (Weibull beta=0.45), meaning failure rate decreases with age castet-saleh-2009-satellite-reliability.1. Satellites that survive the first 1-2 years become progressively more reliable. The strong performance of Iridium NEXT (zero failures at 7-9 years) is consistent with this: mature, well-designed bus hardware does not exhibit wear-out within a 5-7 year horizon.
However, Starlink's proactive retirement of 500+ early V1 satellites at <5 years [starlink-retirement-cgaa.1, starlink-mass-deorbit-pcmag.1] demonstrates that first-generation designs can harbor latent failure modes (in this case, a ferrite transformer defect) that would have caused mid-life failures if unaddressed. The satellites were still operational at retirement — SpaceX chose to replace rather than risk in-orbit failure.
Compute-satellite-specific risk premium: An orbital compute satellite operates at 5-130 kW — 2-50x higher power throughput than Starlink (~2-3 kW). The EPS is the dominant failure driver after infant mortality (27-44% of all spacecraft failures [tafazoli-2009-spacecraft-failures.1, kim-castet-saleh-2012-eps.1]), and EPS failures in LEO are more often fatal than in GEO kim-castet-saleh-2012-eps.1. Higher power means higher currents, more power conversion stages, and more thermal stress on power electronics. Additional risk comes from novel radiative thermal management at scale and larger deployable solar arrays. No LEO satellite has ever operated at 5-130 kW sustained power, so the bus failure rate at these power levels is genuinely unknown — the premium below is an engineering judgment, not an empirically derived value.
These factors justify a 2-3x premium over Starlink V2 Mini's ~0.9%/yr rate (or a larger premium over OneWeb's 0.05-0.08%/yr) for the central case:
- Optimistic (0.5%/yr): Mature compute satellite bus with Starlink heritage, mass-produced. Slightly above the Starlink/OneWeb mature rate due to higher-power EPS complexity.
- Central (1.0%/yr): Novel high-power bus design in early-mature production. Reflects ~3x Starlink V2 Mini rate for EPS stress, novel thermal management, and less design heritage. Consistent with 220-500 kg smallsat class success rates.
- Conservative (2.5%/yr): First-generation novel bus architecture. Comparable to Starlink V1.0 early-batch rates, before design iteration reduces failure rates. Higher-power EPS and novel subsystems amplify infant mortality risk.
Component 2: GPU Accelerator Attrition
GPU attrition is the rate at which individual GPUs fail permanently while the satellite bus continues to operate. The total orbital GPU attrition rate combines the terrestrial permanent failure rate with the space-specific additions:
| Mechanism | Optimistic | Central | Conservative | Derivation |
|---|---|---|---|---|
| Terrestrial permanent base rate | 2.5%/yr | 4.0%/yr | 6.0%/yr | terrestrial-gpu-failure-rate: 5 independent sources |
| Thermal cycling fatigue | 0.1%/yr | 0.6%/yr | 1.6%/yr | Norris-Landzberg model, SAC305 parameters pan-2005-norris-landzberg-sac.1 |
| Destructive SEL | 0.1%/yr | 1.0%/yr | 5.0%/yr | COTS catastrophic rate oliveira-2022-cubesat-radiation.1 + FinFET sensitivity ball-sheets-sel-7nm-finfet-2021.1 |
| Radiation soft errors | ~0%/yr | 0.2%/yr | 1.0%/yr | 14nm crash rate linux-see-cots-soc-2025.1 + mitigation spacecube-cots-iss.1 |
| TID degradation | ~0%/yr | ~0%/yr | 0.5%/yr | Negligible for 5yr at LEO doses tid-7nm-finfet-ro-2021.1 |
| Launch-induced damage | 0.5%/yr | 1.0%/yr | 2.0%/yr | Vibration/thermal interaction combined-vibration-thermal-bga.1 |
| Total orbital GPU attrition | ~3.2%/yr | ~6.8%/yr | ~16.1%/yr |
The terrestrial base rate now reflects only permanent failures requiring physical replacement — not transient faults recoverable by restart or automation. The previous analysis used a fixed 6% rate across all scenarios based on a journalism article's summary of Meta's Llama 3 data. The updated analysis draws on the primary Meta paper meta-llama3-paper.2, a peer-reviewed longitudinal study (SC '25, 11.7M GPU-hours) cui-two-gpus-2025.1, and Meta's explicit transient vs permanent failure taxonomy revisiting-ml-cluster-reliability.2 to derive a scenario-dependent range.
The optimistic scenario benefits most from this refinement: with a 2.5% permanent base rate and well-mitigated space additions (0.7%), the total orbital GPU attrition drops to ~3.2% — roughly half the previous estimate. The conservative scenario is unchanged because the 6% conservative base rate equals the previous fixed rate.
The implied multipliers (1.3x/1.7x/2.7x) remain consistent with MIL-HDBK-217's rating of LEO as equally benign to "ground benign" for non-radiation failure modes mil-hdbk-217-factors.1, with the additional space-specific attrition dominated by destructive SEL and launch-induced damage.
Component 3: SDC and Error-Rate Growth
A critical finding from the SDC research: radiation-induced gradual performance degradation is negligible for FinFET GPUs in LEO with adequate shielding. This mechanism contributes only a fixed ~1.3% capacity overhead, not a growing annual factor.
The reasoning:
TID parametric drift is irrelevant at LEO doses with adequate shielding. 7nm FinFET shows <1% circuit degradation at 380 krad tid-7nm-finfet-ro-2021.1. The expected 5-year dose depends strongly on shielding: ~15-17 krad behind 3mm Al researchgate-leo-radiation.1, ~1 krad behind 5.7mm Al nusat-tid-leo-2025.1, and ~0.7 krad behind ~10mm Al google-suncatcher.1. Even at minimal shielding (3mm Al), FinFET has a ~25x margin; at the 10mm Al assumed for compute satellites, the margin is ~500x. Threshold voltage shifts, leakage increases, and timing degradation are unmeasurable at mission-relevant doses.
Soft error rates do not increase with accumulated TID at relevant doses. In modern SOI SRAMs at 800 krad, the TID-SEU synergistic effect is only ~15% for 6T cells and actually decreases for 7T cells tid-seu-synergy-soi-sram-2022.1. At the shielded 5-year dose of <1-5 krad (depending on shielding depth), this effect is immeasurably small.
SDC is dominated by manufacturing defects, not radiation accumulation. Meta found ~3.6% of CPUs have SDC from manufacturing and aging defects meta-sdc-fleet-2022.1. Google observes "a few mercurial cores per several thousand machines" [google-cores-dont-count-2021]. These are intrinsic silicon defects, not radiation-induced trends.
Inference workloads are naturally resilient. Single bit-flips are masked ~70-85% of the time in LLM inference llm-soft-error-vulnerability-2025.1. SpaceCube demonstrated <1.3% radiation mitigation overhead over 4 years on ISS spacecube-cots-iss.1.
The 1.3% fixed overhead accounts for ECC, memory scrubbing, watchdog/restart systems, and application-level error detection. It does not grow over the mission because the mechanisms that would cause growth (TID-SEU synergy, parametric drift) are negligible at LEO doses for modern FinFET nodes.
Component 4: Economic Obsolescence
Economic obsolescence sets the ceiling on physical lifetime. With Nvidia releasing new architectures annually nvidia-one-year-cadence.1 and each generation delivering 2-4x inference performance, a GPU launched today is 2-3 generations behind within 3 years and 4-5 generations behind within 5 years. At some point, deorbiting and launching current-generation silicon becomes cheaper than operating obsolete hardware.
For inference workloads (less demanding of cutting-edge hardware than training), this crossover likely occurs at 4-5 years peraspera-realities.4. The "operate, deorbit, replace" model embraces this explicitly.
The physical lifetime bounds (7/5/3.5 years) already incorporate economic obsolescence as a co-equal constraint alongside physical degradation. The convergence of five independent indicators at 5 years — SpaceX filing spacex-fcc-million-satellite-filing.1, Starcloud starcloud-first-ai-model-space.1, Starlink precedent starlink-deorbit-stats.1, hyperscaler depreciation gpu-depreciation-schedules.1, and radiation tolerance limits google-suncatcher.1 — is remarkably consistent.
Obsolescence is asymmetrically worse for orbital: terrestrial GPUs follow a "value cascade" from training to inference to batch workloads and retain residual resale value. Orbital GPUs have zero salvage value and cannot be redeployed. If the value cascade extends terrestrial GPU economic life to 5.5-6 years, the orbital-terrestrial lifetime gap widens slightly.
Component 5: Spares and Graceful Degradation
Several mechanisms partially offset gross GPU failures:
Software routing (no amplification). For data-parallel inference, losing 1 of N GPUs reduces throughput by exactly 1/N (evidence item 26). There is no failure amplification — unlike training workloads where a single GPU failure can destabilize an entire NVLink domain. The per-GPU attrition model already correctly captures this 1:1 proportional degradation.
Soft error recovery. A small fraction (~5-10%) of gross GPU "failures" may be transient (SEU/SEFI recoverable by power cycling or software restart) rather than permanent. SpaceCube demonstrated >99.99% error-free operation with software mitigation spacecube-cots-iss.1, and the Blackwell RAS engine provides predictive failure detection (NVIDIA Blackwell RAS engine).
Cold spares. Unpowered spare GPU modules activated when primaries fail are theoretically viable but carry a mass penalty (~4.0-9 kg/kW_IT each). Assessment: likely suboptimal for mass-production orbital compute — the mass is better spent on active capacity, with fleet-level replacement providing the spare function. Sophia Space's SOOS aiaa-sophia-space.1 and Google's Suncatcher both emphasize software routing around failures rather than hardware spares.
Fleet-level replacement. At fleet scale (10,000+ satellites), individual failures average out to predictable aggregate loss rates. Multi-echelon inventory strategies (in-plane spares, parking orbit warehouses, ground stockpile) achieve 53% cost reduction vs direct replacement constellation-spare-strategy-2025.1. This is the primary recovery mechanism, and it is already captured by the effective-lifetime approach in the TCO model.
Net effect on headline values: The recovery factor (optimistic 10%, central 5%, conservative 0%) would reduce effective GPU attrition by ~0.3%/yr centrally. This changes central effective lifetime from 4.1 to ~4.2 years — a <3% improvement. We do not include this in the headline values because (a) the magnitude is within the uncertainty of other parameters, and (b) the 6.8%/yr central GPU attrition rate from the space-hardware-failure-rate analysis is derived from mechanisms that produce primarily permanent damage (destructive SEL, solder joint fatigue, permanent terrestrial failures). The recovery factor is noted here for completeness and as a potential modest improvement not captured in the conservative effective lifetime estimate.
Why 5 Years Is the Central Physical Design Point
Multiple independent lines of evidence converge on 5 years. We distinguish operator design targets (which reflect economic and engineering judgment) from empirical fleet data (which reflect observed hardware performance):
Operator design targets (Tier 4-5 evidence — statements of intent, not measurements):
- SpaceX filing: 5-year operational life stated spacex-fcc-million-satellite-filing.1
- Starcloud CEO: 5-year lifespan based on chip lifetime starcloud-first-ai-model-space.1
- Starlink precedent: 5-year designed lifespan starlink-deorbit-stats.1
Economic evidence (Tier 1-2 — observable market data):
- Hyperscaler GPU depreciation: 5-6 years gpu-depreciation-schedules.1
- GPU longevity: Google TPUs at 100% utilization after 7-8 years gpu-useful-life-2025.2
Empirical fleet survival data (Tier 1-2 — measured performance):
- Iridium first-gen: 20+ year operational life (2.5x design life), fuel-limited iridium-lifetime-extension-spacenews.2
- Iridium NEXT: zero failures at 7-9 years, life extended to 17.5+ years iridium-lifetime-extension-spacenews.1
- OneWeb: 0.3% cumulative failure over 4-7 years oneweb-stats-mcdowell.1
- Starlink V1.0: oldest satellites (Nov 2019) still operating at 6.3+ years [mcDowell-starlink-stats.1]
- Satellite infant mortality pattern (beta=0.45): reliability improves after first 1-2 years castet-saleh-2009-satellite-reliability.1
Component-level evidence (Tier 1-2 — measured degradation rates):
- Li-ion batteries: 65,000+ cycles, 2.4x margin for 5-year LEO mission saft-ves16-leo-battery.1
- Solar panels: 85-97% power retained at 7 years in LEO mdpi-leo-degradation.1
- Radiation TID: 5-year dose behind 10mm Al (~0.7 krad google-suncatcher.1) well within FinFET tolerance (380 krad tid-7nm-finfet-ro-2021.1)
Cautionary data:
- Starlink V1 proactive retirements: 500+ satellites retired at <5 years due to latent component defect starlink-retirement-cgaa.1 — first-generation designs can fail earlier than intended
The 5-year central estimate is the strongest part of the physical lifetime analysis. The fleet and component data show that no physical constraint binds at 5 years for a well-designed satellite. The 5-year design point is set by the economic obsolescence cycle (GPU generational improvement), not by hardware degradation.
Evidence Quality Assessment
The effective lifetime is a model output derived from five sub-inputs with very different evidence quality. The central value of 4.1 years is the result of a structured bottom-up calculation, not a directly measured quantity. No orbital compute satellite has accumulated operational lifetime data, so every sub-input involves extrapolation from non-identical contexts.
| Sub-input | Central Value | Evidence Basis | Evidence Quality | Type of Estimate |
|---|---|---|---|---|
| Physical lifetime (T) | 5 years | Operator design targets, fleet precedent, GPU depreciation, component degradation data | Medium-High — Multiple independent convergences; fleet data (Iridium, Starlink, OneWeb) validates 5+ years for bus hardware; no compute-satellite-specific data | Engineering consensus + empirical fleet data |
| Bus loss rate (λ_bus) | 1.0%/yr | Starlink/OneWeb fleet statistics + EPS risk premium for high-power systems | Medium — Fleet data is Tier 1 (directly observed); the 2-3x premium for high-power EPS is engineering judgment extrapolated from lower-power analogs | Empirical base rate + judgment premium |
| GPU attrition (λ_gpu) | 6.8%/yr | Terrestrial permanent failure rate (5 independent data sources) + space-specific additions (mechanism-level models) | Medium-Low — Terrestrial base rate (4%/yr) is well-sourced (SC '25, Meta, Microsoft); space additions (2.8%/yr) are bottom-up mechanism models with no orbital validation; SEL component (1.0%/yr central) is an engineering judgment across a >6 OOM uncertainty range | Empirical base + mechanism models + judgment |
| SDC overhead | 1.3% fixed | SpaceCube ISS data, Google Suncatcher, FinFET TID testing | Medium-High — Multiple empirical data points; negligible at LEO doses with adequate shielding | Empirical measurement |
| Economic obsolescence | Caps at 5 years | Hyperscaler depreciation schedules, Nvidia release cadence | High — Observable market data; depreciation schedules are public financial disclosures | Market observation |
The sub-inputs that drive the most scenario spread — GPU attrition (especially destructive SEL) and bus loss rate for high-power satellites — are also the ones with the weakest evidence. This is inherent to the problem: the quantities we most need to know (GPU failure rates in space, high-power bus reliability) have no empirical precedent to draw on. The model makes these uncertainties explicit and separable rather than hiding them in a single composite estimate.
Key Uncertainties
No public SEL characterization for NVIDIA H100/B200 — confirmed as genuinely unknown. Targeted research for SEL data on TSMC 4nm/5nm commercial devices found zero published results. NASA's statistical SEL database shows SEL rates vary across >6 orders of magnitude with no predictive trends [ladbury-2025-sel-statistics.2, ladbury-2025-sel-statistics.4]. The 0.1-5.0%/yr destructive SEL range used in this analysis is an engineering judgment spanning the plausible distribution, not an empirically bounded estimate. See space-hardware-failure-rate for the full assessment.
Bus failure rate for high-power compute satellites is an engineering judgment. Fleet data provides excellent empirical anchors for communication-satellite-class hardware (OneWeb 0.05-0.08%/yr oneweb-stats-mcdowell.1, Starlink mature batches ~0.3%/yr [mcDowell-starlink-stats.1], Iridium NEXT ~0%/yr at 7-9 years iridium-lifetime-extension-spacenews.1). The 2-3x premium applied for compute satellites at 5-130 kW reflects the higher EPS stress, novel thermal management, and absence of operational precedent — this is reasonable engineering judgment, but it is not an empirical measurement. The infant mortality pattern (Weibull beta=0.45 castet-saleh-2009-satellite-reliability.1) suggests the premium may be front-loaded: mature compute bus designs could approach communication-satellite reliability levels.
Design maturity trajectory. Bus failure rates will likely decrease through design iteration (as Starlink demonstrated: from 13% to 0.2% across generations starlink-failure-rates-wccftech.1), but the timeline depends on production volume and operational learning rates. Starlink's proactive retirement of 500+ early satellites at <5 years starlink-retirement-cgaa.1 shows that even experienced operators discover latent defects in first-generation designs.
Interaction between bus and GPU mechanisms. A high-power GPU failure causing EPS overload could trigger catastrophic satellite loss. This interaction between mechanisms 1 and 2 is not modeled and could increase effective whole-satellite loss rates beyond the sum of independent terms.
SEL Rate Sensitivity
The destructive SEL rate is the most uncertain sub-parameter in the effective lifetime model. NASA's statistical SEL database shows rates varying across >6 orders of magnitude with no predictive trends ladbury-2025-sel-statistics.2, and there are zero published SEL characterization results for TSMC 4nm/5nm commercial GPU dies (H100/B200). The table below isolates this uncertainty by varying only the destructive SEL rate while holding all other failure parameters at their central values: physical lifetime 5 years, bus loss 1.0%/yr, terrestrial permanent GPU failure 4.0%/yr, non-SEL space additions 1.8%/yr, and SDC overhead 1.0%.
| SEL Rate | Total GPU Attrition | Combined Annual Decay | Effective Lifetime (years) | vs Central |
|---|---|---|---|---|
| 0.1%/yr | 5.9%/yr | 6.8%/yr | 4.17 | +2% |
| 0.5%/yr | 6.3%/yr | 7.2%/yr | 4.13 | +1% |
| 1.0%/yr | 6.8%/yr | 7.7%/yr | 4.08 | central |
| 2.0%/yr | 7.8%/yr | 8.7%/yr | 3.97 | −2% |
| 5.0%/yr | 10.8%/yr | 11.7%/yr | 3.69 | −10% |
| 10.0%/yr | 15.8%/yr | 16.6%/yr | 3.25 | −20% |
Even at the extreme high end (10%/yr destructive SEL — near the conservative scenario's total space-specific GPU attrition), the effective lifetime declines by only 20% from the central value, from 4.1 to 3.3 years. This moderate sensitivity arises because the SEL rate compounds with other attrition mechanisms that already dominate: the 4.0%/yr terrestrial base rate and 1.8%/yr non-SEL space additions together account for 5.8%/yr before any SEL contribution. The effective lifetime is more sensitive to physical lifetime (economic obsolescence) and the terrestrial base rate than to the SEL rate alone. Nevertheless, a SEL rate at the high end of the plausible range (5–10%/yr) would meaningfully compress effective lifetime and shift the economics toward the conservative scenario.
Evidence
Regulatory Constraints
FCC 5-year deorbit rule (effective Sept 2024) The FCC adopted rules requiring LEO satellite operators to complete disposal within 5 years of mission end. This replaced the prior 25-year guideline. New licensees after Sept 29, 2024 must comply. The rule requires post-mission disposal within 5 years of mission end — it does not impose a limit on operational lifetime.
SpaceX FCC filing: 5-year operational life SpaceX's January 2026 filing for up to 1 million orbital data center satellites at 500-2,000 km explicitly states the data centers "are expected to have an operational life of five years." This is the operator's own design target, driven by engineering and economic considerations rather than the FCC disposal rule.
Starcloud: 5-year lifespan Starcloud CEO stated their "satellites should have a five-year lifespan given the expected lifetime of the Nvidia chips on its architecture."
Satellite Bus Reliability
Starlink V2 Mini fleet reliability [evidence:mcDowell-starlink-stats.1] As of March 2026, 11,641 Starlink satellites launched total. Gen2 V2 Mini: ~31 failures across 6,927 units (0.45% cumulative over ~1.5 years average age, implying ~0.3%/yr annualized). Gen1 (4,714 launched): 242 failures (5.1% cumulative over ~5 years). Total fleet: 327 failures (2.8% cumulative across all generations). — McDowell, planet4589.org
Starlink FCC semi-annual report (H2 2023) SpaceX took 14 satellites out of operation during the 6-month period, all retaining collision-avoidance capability. Only a single satellite (Starlink-32386) suffered a suspected power failure preventing standard disposal. 73 satellites reentered during the period. — Kratos Space, citing SpaceX FCC filings
OneWeb fleet reliability As of June 2023, 4 of OneWeb's 634 satellites had failed in orbit (0.63% cumulative over ~3 years, implying ~0.2%/yr annualized). By October 2024, Eutelsat CEO expressed confidence to extend satellite operational life beyond the initial design. — Space Intel Report
Spacecraft failure subsystem distribution Analysis of 156 on-orbit failures across 129 spacecraft (1980-2005): AOCS caused 32% of all failures (gyroscopes alone 17%); power subsystem caused 27%; together AOCS and power accounted for 59% of all failures. — Tafazoli, Acta Astronautica, 2009
EPS failure severity in LEO The electrical power subsystem (EPS) fails less frequently but more fatally in LEO than in GEO. After 10 years of operation, EPS accounts for 44.1% of all failures. 29% of LEO EPS failures are in the electrical distribution subsystem. — Kim, Castet, Saleh, Reliability Engineering and System Safety, 2012
Starlink failure rate improvement Starlink failure rates improved from ~83% (V0.9 prototype batch) to 3-5% (V1.0 early batches) to ~0.2% per batch (mature V1.0 batches, late 2020) to <0.5% cumulative (V2 Mini). This demonstrates ~10-50x improvement through design iteration within a single program. — Derived from McDowell tracking data
Smallsat mission success rates Smallsats (220-500 kg class) achieved 96% mission success rate (2009-2018). Most failures occur in the first 60 days: "If you can make it through your first two months, you'll likely make it through your entire design life." — SSC21-WKIII-02
CubeSat failure distribution CubeSat EPS causes more than 40% of all failures after 30 days of operation; communications accounts for ~26-30%. Most failures are immaturity failures, not environment-induced. Improved testing beats subsystem redundancy for improving reliability. — Bouwmeester et al., 2022
Radiation Degradation
LEO TID environment (light shielding) With 3 mm Al shielding, total ionizing dose in LEO is <10 krad(Si) over a 3-year mission — extrapolating linearly, ~15-17 krad(Si) over 5 years. This represents the dose behind minimal shielding; compute satellites with 10mm Al shielding experience ~5-10x lower doses (see items 13-14).
Google Suncatcher TPU radiation testing Google's Trillium TPUs (v6e Cloud TPU) survived proton beam testing to ~2 krad(Si) before HBM subsystems showed irregularities, with no hard failures up to ~15 krad. Google reports the expected shielded 5-year mission dose as ~700 rad(Si) (behind ~10mm Al equivalent), giving ~3x margin to first HBM irregularities and ~20x margin to hard failure threshold. Commercial AI silicon can survive a 5-year LEO mission with adequate shielding.
In-situ LEO TID measurements LabOSat-01 dosimeters on ÑuSat satellites in polar LEO (~490 km, 97.3° inclination) measured 0.5-1.9 krad over ~3 years depending on shielding depth. At ~2.9mm Al: ~1.9 krad/3yr (~3.2 krad/5yr); at ~5.7mm Al: ~0.6 krad/3yr (~1.0 krad/5yr). These in-situ measurements are consistent with the Google estimate of ~700 rad/5yr behind ~10mm Al — dose decreases roughly exponentially with shielding depth. — arXiv:2503.09520, 2025
7nm FinFET TID tolerance 7nm bulk FinFET ring oscillators show circuit-level degradation (logic gate delays, operating current) of less than 1% after TID exposure of 380 krad(SiO2) — roughly 500x the 5-year shielded LEO dose. — IEEE TNS, 2021
TID-SEU synergy in modern silicon In modern SOI SRAMs, TID irradiation to 800 krad increases 6T SRAM SEU cross-section by only ~15%, while 7T SRAM cross-section actually decreases by ~60%. This contrasts with micrometer-scale SRAMs where SEU cross-section increased by up to 1000x after TID exposure. At mission-relevant doses (<5 krad), the TID-SEU synergistic effect is immeasurably small. — MDPI Electronics, 2022
Derived Inputs (GPU Failure Rates)
Items 17-18 are derived from other research pages, not direct evidence. They summarize inputs consumed by this page's model.
Terrestrial permanent GPU failure rate [analysis] The terrestrial GPU failure rate analysis establishes the annual permanent failure rate for H100-class GPUs at 2.5%/4%/6% (optimistic/central/conservative). This reflects only failures requiring physical GPU replacement — not transient faults recoverable by restart. The range is derived from five independent data sources: Meta's Llama 3 primary paper [meta-llama3-paper.2], the NCSA Delta longitudinal study (2.5 years, 11.7M GPU-hours) [cui-two-gpus-2025.6], Meta's ML cluster reliability study [revisiting-ml-cluster-reliability.2], the NTP paper [nonuniform-tensor-parallelism.1], and Microsoft SuperBench [microsoft-superbench.1]. The Epoch AI MTBF of ~50,000 hours [epoch-gpu-failures.1] conflates all 419 job interruptions (hardware + software + network) as "failures" and should not be used for permanent failure rate estimates.
Space-specific GPU failure mechanisms [analysis] The space hardware failure rate analysis builds a bottom-up mechanism-level budget for space-specific GPU attrition: thermal cycling fatigue (0.1–1.6%/yr), destructive SEL (0.1–5.0%/yr), radiation soft errors (0–1.0%/yr), TID degradation (0–0.5%/yr), and launch-induced damage (0.5–2.0%/yr). These sum to 0.7–10.1%/yr space-specific additions on top of the scenario-dependent terrestrial permanent base rate, giving total orbital GPU attrition of 3.2–16.1%/yr.
SDC and Gradual Degradation
SpaceCube radiation mitigation overhead SpaceCube on ISS achieved >99.99% error-free operation of eight COTS PowerPC processors over four years using FPGA and software radiation mitigation (scrubbing, error detection, watchdog, checkpoint/restart). Radiation hardening by software overhead was <1.3%.
HBM radiation sensitivity HBM uncorrectable ECC error sensitivity: approximately one event per 50 rad of proton exposure. With 10mm Al shielding in dawn-dusk SSO (~150 rad/yr), approximately 1 uncorrectable error per 10 million inferences. Google assessed this as "likely acceptable for inference."
LLM inference soft error masking In LLM inference, most errors from a single bit-flip are masked (cause no output change). The abnormal outcome rate grows from ~15-30% for 1 injected fault. Larger models show a higher proportion of masked errors. Bit-flip vulnerability is strongly position-dependent: low-order bits cause <1% SDC rate while high-order bits (30-31) cause 23-24% SDC rate. — Chai et al., arXiv:2601.19912, 2025
GPU SDC terrestrial baseline GPU silent data corruption rate is 8.15×10⁻³ FIT per device at sea level (one error per 14,000 device-hours). Cosmic radiation causes 61.7% of faults; manufacturing variations contribute 4.3%. — Global Journal of Computer Science and Technology, 2025
Meta fleet SDC rates Meta found approximately 3.6% of CPUs in their fleet cause SDCs. Root causes are CPUs that are "born defective (escaped manufacturing testing), become defective (aging), or just differ from each other (timing variability)" — manufacturing defects, not radiation accumulation. — Meta Engineering Blog, 2022
Graceful Degradation
Sophia Space SOOS Sophia Space's SOOS monitors each tile's temperature and computational load, redistributing work when a quad or entire tile goes offline due to debris impact or hardware failure. The system gracefully degrades performance until a replacement module is launched. — Aerospace America/AIAA, 2026
Inference failure scaling: no amplification [analysis] Data parallelism for inference provides linear throughput scaling — losing one data-parallel replica reduces aggregate throughput by exactly 1/N. There is no failure amplification in data-parallel inference because each replica serves independent requests. This is a fundamental architectural property of data-parallel inference, in contrast to tensor-parallel training where a single GPU failure can halt an entire NVLink domain.
Epoch AI spare buffer analysis A 1M GPU cluster experiencing 20 node failures/hour with 1-day replacement time needs a buffer of only 480 nodes (0.3% of total capacity). Even with multi-day replacement delays, performance impact remains below 1%. — Epoch AI, 2024
Constellation spare strategy Multi-echelon inventory optimization for satellite constellations: in-plane spares for immediate replacement, parking orbit spares for hours-timescale resupply, ground stockpile with launch lead time. The indirect strategy (parking orbit spares with batch resupply) achieves 53% total cost reduction vs direct replacement. — arXiv:2509.09957, Georgia Tech, 2025
Solar Panel Degradation
GEO solar array measured degradation Telemetry from 11 GEO communications satellites showed GaAs cells degrade at 0.44-1.03%/yr and Si cells at 0.71-1.69%/yr. LEO radiation fluences are 5-10x lower than GEO.
ISS solar array degradation The ISS P6 silicon photovoltaic arrays showed measured short-circuit current degradation of 0.2-0.5%/yr, below the predicted rate of 0.8%/yr.
Solar panel lifetime implication At 0.2-1.5%/yr degradation in LEO, a 5-year satellite retains ~92-97% of initial power; a 7-year satellite retains ~90-96%. Solar panels are NOT the binding constraint on satellite lifetime.
GPU Obsolescence
Nvidia 1-year release cadence Nvidia has shifted from a 2-year to a 1-year release cadence for datacenter GPUs: Hopper (2022), Blackwell (2024-25), Rubin (2026), Feynman (2028). Each generation delivers roughly 2-4x performance improvement for AI inference.
GPU depreciation schedules AWS, Google, and Microsoft use 6-year depreciation schedules for servers/GPUs. Industry is converging toward 5-year useful life via a "value cascade" model. AI-native neoclouds use 4-5 year schedules.
Dylan Patel on orbital GPU economics Dylan Patel (SemiAnalysis) noted that the testing, assembly, and launch process for orbital GPUs could consume 6+ months, representing "10% of your cluster's useful life" if GPU useful life is 5 years.
"Operate, deorbit, replace" model Per Aspera analysis proposes that early orbital clouds will default to designing for a "5-7 year tour, then burn it up and launch Version N+1 with the latest silicon."
Starlink Operational Data
Starlink designed lifespan: 5 years Starlink satellites are designed for a 5-year operational lifespan. Of 10,801 launched, ~1,391 (~13%) have re-entered, with early batches showing 3-5% uncontrollable failure rates.
Starlink deorbit activity 2024-2025 SpaceX deorbited 472 Starlink satellites between December 2024 and May 2025, and another 218 between June and November 2025. Many were less than 5 years old, suggesting proactive fleet management and replacement with newer models.
Station-Keeping
- Orbital decay at LEO altitudes Atmospheric drag in LEO causes ongoing orbital decay, with rates increasing substantially during high solar activity. Station-keeping propulsion is required throughout the operational lifetime to maintain altitude. At ~575 km, uncontrolled satellites deorbit within a few years.
Fleet Operational Lifetime Data
Iridium NEXT: zero failures through 7-9 years, life extended to 17.5+ years Iridium NEXT satellites (launched 2017-2019, 12.5-year design life from Thales Alenia Space) were re-assessed in February 2024 to operate "well to at least 2035" (17.5+ years). CEO Matt Desch expects satellites will be in service "longer than 17.5 years." First-generation Iridium with similar design life lasted 20+ years before fuel depletion, not component failure. — SpaceNews, February 2024
First-generation Iridium: 20+ years in LEO Original Iridium satellites (8-year design life, launched 1997-2002) operated for 20+ years at 780 km altitude, completing ~100,000 orbits each. Decommissioning was driven by fuel depletion, not component failure. This is the strongest empirical data point for LEO satellite bus longevity: 2.5x design life achieved. — SpaceNews, February 2024; Iridium Communications
Starlink Gen1 unplanned failure rate: ~1.5%/yr [evidence:mcDowell-starlink-stats.2] Starlink Gen1 unplanned failures (excluding proactive retirements): early deorbits (114) + reentry after failure (127) + screened (100) + dead (7) = 348 out of 4,714 launched = 7.4% cumulative over ~5 years average age, implying ~1.5%/yr annualized. This includes immature V0.9 (100% loss) and early V1.0 batches with known design flaws. — Derived from McDowell tracking data, March 2026
Starlink proactive retirement of <5-year satellites SpaceX identified "a common issue in a small population" of early V1 Starlink satellites involving ferrite transformers. 100 satellites proactively retired February 2024; over 500 first-generation models retired total. Satellites were "currently maneuverable and serving users effectively" at time of retirement. — Multiple sources compiled by CGAA; SpaceX statements
Starlink mass deorbit: 472 satellites in 6 months SpaceX FCC filing shows 472 Starlink satellites deorbited December 2024-May 2025. 430 of 472 were first-generation. "Most of the satellites that reentered the atmosphere did so less than five years after beginning operations" despite being "designed to orbit the Earth for five years." — PC Magazine via Space4Peace, July 2025
OneWeb: 0.3% cumulative failure over 4-7 years As of March 2026, OneWeb has 656 satellites launched across 20 missions (2019-2024), with 654 in orbit and only 2 total down. 637 fully operational. Cumulative failure rate 0.3% over 4-7 years of operation (~0.05-0.08%/yr annualized). Design life 7+ years. — Jonathan McDowell, planet4589.org
Satellite infant mortality (Weibull beta=0.45) Analysis of 1,584 Earth-orbiting satellites (1990-2008) shows satellite failures follow an infant mortality pattern: Weibull shape beta=0.4521, scale theta=2,607 years. Beta < 1 indicates failure rate decreases with age — satellites that survive early life become progressively more reliable. This contradicts an earlier assumption of wear-out (beta=1.7). — Castet & Saleh, Reliability Engineering and System Safety, 2009
Li-ion batteries provide 10-12 year LEO capability Saft VES16 space-qualified Li-ion cells achieve 65,000+ cycles at 30-50% depth of discharge over 12 years. A 5-year LEO mission requires ~27,000 cycles; a 7-year mission ~37,800 cycles. Li-ion provides ~2.4x cycle margin for 5 years. Prior Ni-Cd/Ni-H2 batteries lasted only 5-7 years. — Saft manufacturer data
Modern reaction wheels: 3M+ failure-free hours NewSpace Systems T065 reaction wheel (first flown 2014) has accumulated over 3 million failure-free hours in orbit with no reported SEUs. Over 800 wheels sold, baselined on 4 constellation programs. — NewSpace Systems
Starlink orbit lowering increases drag SpaceX plans to lower ~4,400 Starlink satellite orbits from 550 km to 480 km throughout 2026, increasing atmospheric drag and propellant consumption for station-keeping. Reduces ballistic decay time by >80%. — Space.com, January 2026