UPDATED 12:38 EDT / APRIL 06 2026

AI

The $100B question: AI’s appetite for compute is rewriting the rules of tech

We’re hitting a pivotal moment in artificial intelligence right now. The latest financial disclosures from the front lines of the AI arms race — OpenAI Group PBC and Anthropic PBC — don’t just give us a peek under the hood; they expose the core tension shaping the entire industry. The smartest machines ever built are also the most expensive to create and operate in the history of computing.

From software margins to infrastructure gravity

For two decades, the tech industry has been conditioned by the “magic” of software economics: 80%-plus gross margins, near zero marginal costs and infinite scalability.

That model is breaking. AI isn’t pure software; it’s infrastructure-as-a-service with a heavy dose of industrial-era physics. It is compute-bound, energy-intensive and capital-heavy. When OpenAI projects a staggering $121 billion in compute spend by 2028, it isn’t an outlier. It’s a signal that we have entered the era of “Infrastructure Gravity.”

The dirty secret: ‘Profitable’… if you ignore the biggest cost

Both OpenAI and Anthropic are now reporting profitability in two distinct ways: with training costs excluded, and with training costs included.

That alone should stop you in your tracks. It reveals that training costs are no longer a “R&D expense” — they are the cost of goods sold. Strip them out, and you have something that looks like software as a service. Put them back in, and you’re staring at one of the most capital-intensive industries ever created.

The 2025-2026 financial reality:

Annualized revenue: OpenAI ~$13 Billion | Anthropic ~$7 Billion

Valuation: OpenAI ~$300 Billion | Anthropic ~$183 Billion

Compute/training intensity: OpenAI – High (Consumer + Enterprise) | Anthropic – High (Enterprise + API)

Inference: The hidden tax on intelligence

While the training gets the headlines, inference is the silent killer. Serving AI responses currently consumes more than 50% of revenue for major providers. Every prompt has a cost. Every free user is a liability until they convert.

By 2026, inference is projected to account for 65% of all AI compute, and over a model’s lifetime, it can represent 80% to 90% of total costs. We’ve seen this with GPT-4: a ~$150 million training cost ballooned into an estimated $2.3 billion in inference costs within two years. This is a 15x multiplier that software-centric VCs aren’t used to seeing.

The strategic fork: Enterprise vs. consumer

We are seeing a clear divergence in how these giants chase the “means of production”:

  • Anthropic: Leaning heavily into the enterprise and API layers. Approximately 80% of its revenue now comes from business customers and API calls (notably powering tools like GitHub Copilot and Cursor). This brings predictable revenue and higher margins.
  • OpenAI: Straddling both worlds. It’s building a consumer platform (ChatGPT has more tan 800 million weekly active users) while maintaining a massive enterprise footprint. For them, the free tier isn’t a gift; it’s a data-collection infrastructure.

This isn’t SaaS. It’s the new hyperscale

If you zoom out, this starts to look familiar. We’ve seen this with Amazon Web Services Inc., Microsoft Azure and the telecom buildouts of the ’90s. Massive upfront capital, long payoff cycles and inevitable market consolidation.

The difference? Speed. AI is compressing decades of infrastructure buildout into a few years. This is why the “Big Four” hyperscalers are expected to spend more than $650 billion into capital expenditures by 2026 alone.

My angle: The beginning of the AI industrial era

What we’re witnessing is the birth of a new industrial layer: AI factories, compute supply chains and energy-to-intelligence conversion. This is why Nvidia Corp. sits at the center of the universe, and why “sovereign AI” is becoming a matter of national security.

The real battle: Bending the cost curve

The game ends in one of two ways. If hardware efficiency (such as Google LLC’s TPUs or Nvidia’s Blackwell) can drop the cost of intelligence faster than demand rises, the model becomes unstoppable. If not, we are looking at massive margin compression and a brutal era of consolidation.

The bottom line

Don’t get distracted by the demos. In this new era, the companies that win won’t just have the best models. They’ll have the best economics. Focus on the stack: compute, cost, distribution, monetization. That’s where the $100 billion question will be answered.

Image: SiliconANGLE

A message from John Furrier, co-founder of SiliconANGLE:

Support our mission to keep content open and free by engaging with theCUBE community. Join theCUBE’s Alumni Trust Network, where technology leaders connect, share intelligence and create opportunities.

  • 15M+ viewers of theCUBE videos, powering conversations across AI, cloud, cybersecurity and more
  • 11.4k+ theCUBE alumni — Connect with more than 11,400 tech and business leaders shaping the future through a unique trusted-based network.
About SiliconANGLE Media
SiliconANGLE Media is a recognized leader in digital media innovation, uniting breakthrough technology, strategic insights and real-time audience engagement. As the parent company of SiliconANGLE, theCUBE Network, theCUBE Research, CUBE365, theCUBE AI and theCUBE SuperStudios — with flagship locations in Silicon Valley and the New York Stock Exchange — SiliconANGLE Media operates at the intersection of media, technology and AI.

Founded by tech visionaries John Furrier and Dave Vellante, SiliconANGLE Media has built a dynamic ecosystem of industry-leading digital media brands that reach 15+ million elite tech professionals. Our new proprietary theCUBE AI Video Cloud is breaking ground in audience interaction, leveraging theCUBEai.com neural network to help technology companies make data-driven decisions and stay at the forefront of industry conversations.