UPDATED 16:01 EST / NOVEMBER 22 2025

AI

Resetting GPU depreciation: Why AI factories bend, but don’t break, useful life assumptions

Much attention has been focused in the news on the useful life of graphics processing units, the dominant chips for artificial intelligence. Though the pervasive narrative suggests GPUs have a short lifespan, and operators are “cooking the books,” our research suggests that GPUs, like central processing units before, have a significantly longer useful life than many claim.

In this Breaking Analysis, we use the infographics below to explain our thinking in more detail.

Premise

In January 2020, Amazon changed the depreciation schedule for its server assets, from three years to four years. This accounting move was implemented because Amazon found that it was able to extend the useful life of its servers beyond three years. Moore’s Law was waning and at Amazon.com Inc.’s scale, it was able to serve a diverse set of use cases, thereby generating revenue out of its EC2 assets for a longer period of time. Other hyperscalers followed suit and today, the big three all assume six-year depreciation schedules for server assets.

The question is, in the AI factory era, will this same dynamic hold or will the useful life of servers compress? It is our view that the dynamic will hold in that today’s most demanding training infrastructure will serve a variety of future use cases, thereby extending the useful life (revenue per token) of GPUs beyond their initial purpose. At the same time, we believe the more rapid innovation cycles being pushed by Nvidia Corp. will somewhat compress depreciation cycles from their current six years to a more conservative five-year timeframe.

Figure 3: The idea that GPUs will have a useful life of two to three years is incorrect in our view. Today’s training infra will support future workloads and applications such as inference.
Data sourced from public SEC filings and industry reports

The ‘six-year shift’

What the infographic shows:
Figure 1 shows the depreciation schedules for Amazon Web Services, Google Cloud and Microsoft Azure from 2017 to 2025. The chart highlights the coordinated progression from three- and four-year schedules to a uniform six-year useful life assumption beginning in 2023-2024.

Key takeaways:

  • AWS was first to extend asset life, triggering a fast-follow response.
  • By 2023, all three hyperscalers normalized on six years.
  • The expansion represents a noticeable change in operating income by spreading depreciation over a longer horizon.
  • The shift occurred before the massive wave of AI CapEx, meaning AI infrastructure inherits but may eventually challenge this standard.

Our research suggests that this shift reflects the hyperscalers’ confidence in workload diversification. Even as hardware aged, demand for general compute, analytics, web services and long-tail workloads sustained revenue generation across an asset’s lifecycle. The open question is whether AI GPU estates, which are more expensive, power constrained and evolving faster, behave the same way.

The new players emerge

What the infographic shows:
Figure 2 shows a horizontal bar comparison of depreciation schedules across hyperscalers and AI-native neoclouds (CoreWeave, Nebius, Lambda Labs). The big clouds have settled firmly on six-year depreciation, while neoclouds take more conservative positions (five years for Lambda Labs, four years for Nebius).

Key takeaways:

  • CoreWeave is the outlier, adopting an aggressive six-year posture despite an exclusively AI-intensive focus;
  • Nebius and Lambda Labs use shorter cycles, reflecting faster modernization cycles and possibly less heterogeneous workload mix;
  • Neocloud strategies give an indicator of where AI-pure-play economics may differ from general cloud economics.

In our view, Figure 2 is a harbinger for slight compression. AI-first clouds cannot afford stagnant infrastructure; performance/watt gains in successive GPU generations directly determine competitiveness. As neoclouds scale, their four- to five-year cycles will likely influence hyperscaler modeling —especially as accelerated computing consumes a larger share of CapEx.

Will six years hold for AI? The GPU value cascade

What the infographic shows:
Figure 3 depicts a three-stage lifecycle framework:

  • Years 1-2: Primary economic life to support foundational model training.
  • Years 3-4: Secondary life to support high-value real-time inference.
  • Years 5-6: Tertiary life to support batch inference ad analytics workloads.

Key takeaways:

  • The value cascade structurally extends GPU usefulness, even as generations turn over rapidly;
  • Training requires peak performance; inference tolerates lower latency constraints; batch/analytics operate at the long tail;
  • This is analogous to server repurposing that justified past depreciation extensions.

We believe this framework supports longer useful life assumptions than the two to three years many have stated. Silicon doesn’t die at end-of-training usefulness, rather it transitions to less demanding but still revenue-generating tasks. This is the essence of token revenue maximization in the AI factory era.

The financial impact

What the infographic shows:
Figure 4 shows three tiles summarizing impact:

  • Six years as emerging standard.
  • Two to three times extended life from cascading.
  • $7 billion annual impact to operating income (example using approximate figures from hyperscaler).

Key takeaways:

  • Extending useful life does lift GAAP profitability for hyperscalers;
  • GPUs, however, may actually have longer economic tails than CPUs because inference and internal workloads are insatiable. But perhaps not as long as pre-generative AI CPU cycles;
  • The financial leverage from longer depreciation schedules will become even more pronounced as GPU CAPEX scales into the trillions.

Our analysis suggests that AI factories magnify the issue but not to the extent many in the media have projected. A one-year change in useful life assumptions for trillion-dollar GPU estates can swing operating income by tens of billions; but in the grand scheme of operating profits for hyperscalers, it’s not game-changing. This dynamic will become a recurring narrative in earnings calls and investor guidance. As such, investors should look at operating and free cash flows to get a better sense of business performance.

The following section summarizes our thinking on this issue:

Premise 1: GPUs have resilient lifecycles

Key takeaways:

  • Training is a short window (one to two years) where only the newest silicon is competitive;
  • Inference demand is exploding and absorbs “last-gen” GPUs for years 3-4;
  • Utility workloads (internal processing, fine-tuning, retrieval augmentation) extend asset life to year 5 and potentially beyond.

We believe the proliferation of agentic applications, retrieval workflows, automation assistants and fine-tuning pipelines will expand the useful life of GPUs beyond the training phase. This reinforces the durability of extended economic cycles.

Premise 2: Neoclouds hint at slightly shorter lifespans for hyperscalers

Key takeaways:

  • AI-native clouds optimize for modernization speed over accounting benefits;
  • five-year midpoint appears to be the emerging equilibrium for AI-centric providers;
  • Accelerated computing is changing the shape of budgets – shorter depreciation aligns better with faster GPU innovation cycles.

Our research indicates that this five-year midpoint is where hyperscalers will ultimately converge as accelerated infrastructure dominates CapEx. Thoough general compute has historically survived at six years, AI factories operate under different competitive pressures. Moreover, the utility of legacy CPU technologies may compress as they become less useful. But on balance, we don’t see a dramatic alteration of the income statement as a result of AI.

Premise 3: Cash flow becomes the more meaningful KPI

Key takeaways:

  • High depreciation masks the true economic performance of AI factories;
  • OCF adds back depreciation and reveals that GPU estates generate strong cash even when earnings appear compressed;
  • As CapEx rises, earnings divergence from cash flow will widen.

In our opinion, investors must shift their valuation frameworks. Operating cash flow will become an increasingly important key performance indicator relative to non-GAAP and GAAP income. OCF will be a primary indicator of AI factory health, sustainability and ROI timelines. This becomes crucial as depreciation cycles compress.

Synthesis of the premise

Key takeaways:

  • GPUs are durable economic assets, not short-lived commodities;
  • A five-year cycle is emerging as the balance point between modernization and economic efficiency;
  • Investors must evaluate AI factories through a cash-flow lens rather than earnings alone.

The above graphic frames the strategic direction of the industry, which will likely extended economic utility with slightly shorter accounting life – reflecting both technological realities and revenue/token expectations.

Conclusion

Our research indicates that the useful life of GPUs will continue to benefit from the “value cascade,” allowing assets to generate revenue well beyond their initial training window. At the same time, the arrival of new architectures every 12 to 18 months will put pressure on formal depreciation schedules. We believe hyperscalers will ultimately converge on a five-year cycle – shorter than today’s six-year model but still supported by extended economic usefulness.

In the AI factory era, depreciation becomes a lever. The winners will be operators that maximize the utility curve of GPUs across training, inference and internal workloads while maintaining access to compute, land, water, power and the skills to build AI factories. As CapEx scales into the trillions, cash flow will increasingly be the metric investors will watch.

Disclaimer: All statements made regarding companies or securities are strictly beliefs, points of view and opinions held by SiliconANGLE Media, Enterprise Technology Research, other guests on theCUBE and guest writers. Such statements are not recommendations by these individuals to buy, sell or hold any security. The content presented does not constitute investment advice and should not be used as the basis for any investment decision. You and only you are responsible for your investment decisions.
Disclosure: Many of the companies cited in Breaking Analysis are sponsors of theCUBE and/or clients of Wikibon. None of these firms or other companies have any editorial control over or advanced viewing of what’s published in Breaking Analysis.
Images: theCUBE Research

A message from John Furrier, co-founder of SiliconANGLE:

Support our mission to keep content open and free by engaging with theCUBE community. Join theCUBE’s Alumni Trust Network, where technology leaders connect, share intelligence and create opportunities.

  • 15M+ viewers of theCUBE videos, powering conversations across AI, cloud, cybersecurity and more
  • 11.4k+ theCUBE alumni — Connect with more than 11,400 tech and business leaders shaping the future through a unique trusted-based network.
About SiliconANGLE Media
SiliconANGLE Media is a recognized leader in digital media innovation, uniting breakthrough technology, strategic insights and real-time audience engagement. As the parent company of SiliconANGLE, theCUBE Network, theCUBE Research, CUBE365, theCUBE AI and theCUBE SuperStudios — with flagship locations in Silicon Valley and the New York Stock Exchange — SiliconANGLE Media operates at the intersection of media, technology and AI.

Founded by tech visionaries John Furrier and Dave Vellante, SiliconANGLE Media has built a dynamic ecosystem of industry-leading digital media brands that reach 15+ million elite tech professionals. Our new proprietary theCUBE AI Video Cloud is breaking ground in audience interaction, leveraging theCUBEai.com neural network to help technology companies make data-driven decisions and stay at the forefront of industry conversations.