TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem


Published: Updated: 
4 min read

AI Industry Pivots to Efficiency-First as Power Economics Force Reckoning

The AI infrastructure boom hits its physics constraint: massive computing flips from scale-at-all-costs to cost-per-token optimization. DeepSeek's $6M breakthrough and Kelly's 20-watt prediction signal 2026 efficiency inflection.

Article Image

The Meridiem TeamAt The Meridiem, we cover just about everything in the world of tech. Some of our favorite topics to follow include the ever-evolving streaming industry, the latest in artificial intelligence, and changes to the way our government interacts with Big Tech.

  • Chris Kelly, ex-Facebook CPO, tells CNBC: 'We run our brains on 20 watts. We don't need gigawatt power centers to reason'—signaling AI's infrastructure pivot from scale-first to efficiency-first

  • Data centers consumed $61 billion in deals during 2025, while OpenAI alone committed over $1.4 trillion to infrastructure partnerships, but power grid strain is now the hard constraint

  • DeepSeek's December 2024 breakthrough—an open-source LLM built for under $6 million—proved efficiency-first engineering beats brute-force spending, validating Kelly's thesis before major labs caught on

  • Watch for 2026 cost-per-token milestones: first lab to 10x efficiency gains without 10x power demand becomes the market leader

The AI industry's gold-rush phase just hit a wall—literally. After two years of hyperscalers burning billions to outbuild each other, the constraint isn't computing power anymore. It's electricity, cost, and the physics of efficiency. Chris Kelly, ex-Facebook Chief Privacy Officer, articulated what the numbers have been screaming: the next wave of AI competition won't be about who builds the biggest data center. It'll be about who figures out how to think smarter, not louder. That shift moves from theoretical to operational right now.

The numbers looked unstoppable for exactly one moment. OpenAI's $1.4 trillion in AI infrastructure commitments, $61 billion in global data center dealmaking during 2025, and Nvidia and OpenAI's 10-gigawatt project—equivalent to powering 8 million homes—painted a picture of unlimited scale. Then reality checked in.

Chris Kelly, who spent years as Facebook's Chief Privacy Officer before becoming general counsel, just named what everybody already knows: that model breaks. "We run our brains on 20 watts," Kelly told CNBC's Squawk Box on Tuesday. "We don't need gigawatt power centers to reason. Finding efficiency is going to be one of the key things that the big AI players look to."

It's not a prediction. It's a reset. And it's already happening.

The shift reveals itself in data that competitors have been trying to downplay for months. DeepSeek launched a competitive open-source model in December 2024 for under $6 million—a number that threw a spotlight on the gap between Chinese engineering pragmatism and American scaling ideology. When a startup in Beijing can match U.S. lab capability at 0.04% of the typical spend, the message stops being abstract. Cost architecture matters. Efficiency wins.

For context, this mirrors—but inverts—the cloud computing transition of the 2010s. Back then, the inflection point was moving from expensive on-premise servers to cheap cloud scale. The winner was whoever could offer the most compute for the least cost. Now the constraint has flipped. Compute is abundant. Power is the scarcity. And whoever can deliver reasoning without gigawatt-scale infrastructure becomes the next structural advantage.

What makes Kelly's timing commentary particularly sharp is that he's not just naming the shift—he's naming the mechanism. The companies that "reach a breakthrough in lowering data center costs," Kelly said, "will emerge as AI winners." Not the labs with the most GPUs. Not the ones with the flashiest benchmarks. The ones who crack efficiency. That's not a marginal optimization. That's the new competitive moat.

The pressure is already visible. The data center market has accumulated over $61 billion in infrastructure dealmaking in 2025 as hyperscalers rushed into what looks increasingly like an arms race with a false finish line. But the grid can't sustain the burn rate. Utilities have been flagging the looming crunch for months. The Northeast grid is strained. California's power demands from data centers are reshaping investment in nuclear. And here's the thing: throwing more money at bigger data centers doesn't fix the constraint. It just makes it worse. The inflection point is when labs realize that out.

Kelly expects this to accelerate the emergence of Chinese competitors, especially given Trump's recent approval of Nvidia H200 chip sales to the country. Open-source models from China, he noted, will provide "basic levels of compute" for generative and agentic AI—which means the efficiency-first model doesn't just reshape American labs. It reshapes competitive geography. Chinese players built on lean engineering principles. U.S. labs built on unlimited scaling. When efficiency becomes table stakes, the engineering culture that's been optimizing for power-per-output wins.

This is the moment the window for training-at-scale-cost closes. Not because companies stop building data centers. They won't. But because the ROI math inverts. Every company still trying to out-GPT their competitors is competing in the wrong dimension. The race now is inference efficiency. Model compression. Sparse compute. Attention mechanisms that don't require full-matrix operations. These are the engineering problems where innovation creates defensible advantage.

For OpenAI and Nvidia, the arithmetic gets uncomfortable. $1.4 trillion in commitments depends on continuous revenue growth to justify the infrastructure spend. But if efficiency gains compress the compute required per task by 5x or 10x, the utilization math changes. Suddenly you don't need that tenth gigawatt. That reshapes partnership economics, timeline expectations, and which infrastructure players survive 2026.

Kelly's broader point—the one that frames the entire transition—is almost deceptively simple: we already know how to build systems that reason efficiently. A human brain consumes 20 watts. Our current models consume millions of times that. The gap between current and possible is where competition moves next. Not bigger. Better.

The inflection point is real and immediate. For builders, this means infrastructure decisions made in January 2026 will either enable efficiency-first scaling or become stranded assets by Q3. Investors should watch cost-per-token metrics, not just revenue per data center—the first to achieve 5x efficiency gains at current power budgets wins structural advantage. Decision-makers at enterprises: your AI ROI now depends on model efficiency, not lab size. Professionals: the skills that matter in 2026 are inference optimization and sparse compute, not prompt engineering. The window for efficiency-first adoption is 6-8 months wide. Watch for the first major lab to announce a 50% power reduction while maintaining capability—that's the threshold crossing.

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiem

TheMeridiemLogo

Missed this week's big shifts?

Our newsletter breaks
them down in plain words.

Envelope
Envelope

Newsletter Subscription

Subscribe to our Newsletter

Feedback

Need support? Request a call from our team

Meridiem
Meridiem