🔮 Exponential View #545: Rethinking scale; GPUs are the new turbines; stuck graduates, robots & AI companions++
A weekly briefing on AI and exponential technologies
Good morning from London!
Software demands a new world
Silicon Valley spent decades abstracting away from physical reality. But AI is so computationally intensive that it’s dragging tech back into the world of concrete and copper wire. Software is no longer eating the world; it’s demanding a new one.
Banks and other forecasters think there’s $4-5 trillion in building and power needed to run AI. Meta, by its own admission, wants to build a data center the size of Manhattan. This is all demanding massive changes to our energy system.
Microsoft foresees a consistent data center crunch through 2026. As Kevin Scott said
It’s been almost impossible to build capacity fast enough since ChatGPT and GPT-4 launched… even our most ambitious forecasts are just turning out to be insufficient on a regular basis.
But as many as 95% of data centers built for cloud computing may struggle to meet the power demands of new generation AI chips, according to BofA. A single AI rack will draw up to 600kW by 2027 (up from ~50-120 kW/rack today).
This boom is creating new scarcities. Electricity now costs 267% more than it did five years ago in areas near a major data center in the US, fueling backlash as residents subsidize corporate power demand.
Another one is the ongoing demand for close-to-obsolete GPUs. Take Nvidia’s A100, five years after it was first launched, and even after its successor has begun shipping, is harder to get today than it was in 2020. Prices are rising and depreciation schedules are extending from three years to six years. New supply is being snapped up.
As the bottlenecks shift onto the physical infrastructure, it’s no wonder that companies building this new world are doing well. Stocks in nuclear power company Constellation Energy are up almost 5x since ChatGPT launched, outperforming even Microsoft.
Abundance needs to feed abundance though, and we’ll need new solutions. On the Open Circuit podcast this week, I argued that communities around data centers need a fair way to participate in the build‑out. Data centers need to be better grid citizens, perhaps by providing grid services – offering demand response and stability and paying when they strain the system. Others, like electrification nonprofit Rewiring America, suggest that hyperscalers co-fund a mass deployment of heat pumps, rooftop solar, batteries, weatherization and other technologies to cut residential peak load in data center alleys. Done at scale through 2029, these measures could offset the AI-driven load growth and lower household bills.
Recursive intelligence
After a spree of scaling, the frontier may be curving back toward elegance. Samsung’s 7-million-parameter tiny recursive model (TRM) scores well on two important benchmarks: 45% on ARC-AGI-1 and 8% on ARC-AGI-21. The TRM edges past much larger reasoning LLMs including DeepSeek R1, o3-mini and Gemini 2.5 Pro with less than one-ten-thousandth of their parameters. The trick is thinking in loops, as the model recursively critiques and revises its candidate answer until it improves.

This matters because ARC-AGI is designed to be easy for humans, hard for AIs. Most frontier models score in the low single digits on ARC-AGI-2 even with “thinking” modes enabled. Samsung’s tiny recursive network roughly doubles that mark. Still far from human level, but a striking efficiency frontier.
The broader pattern is now visible in that compute spent at inference (“test-time compute”) often beats raw parameter count. If these gains migrate from puzzles to real-world tasks, they’ll create a new family of small, recursive, edge-capable systems that run locally rather than in distant data centers. This could reshape deployment economics, especially where energy and memory are scarce.
Scale isn’t finished. Nor would this suppress compute demand or slow adoption. Data centers will remain supply-constrained, particularly around memory, while hundreds of millions of modern phones and edge devices can now host TRMs to open new deployment paths. For a certain class of tasks, inference could be distributed to billions of devices.
See also:
- ’s annual State of AI report is out and worth your time.
Tencent’s AI model beat Google’s “nano-banana” on text-to-image generation.
An update on graduate jobs and AI
Grads are having a terrible time getting jobs. The unemployment rate for US grads has averaged 6.6% over the past year (the national average is 4%). Many blame AI, but my working assumption is different.
Labor markets are complex and slow to change. They tightened after Covid, and now, amid genuine economic uncertainty, firms are focused on cutting costs. It’s simpler to pause hiring than to fire, and easiest of all to skip inexperienced candidates. If you’re running a company in this environment, you can spin a hiring slowdown as progress by pointing to AI wins. Two data points I came across this week back it up.
When the FT compared young graduates’ unemployment with other first-time entrants into the job market regardless of age, they found that “those without a degree are actually having a much harder time of it,” no matter their age. The slowdown is about (in)experience rather than age.
Yale’s Budget Lab’s latest report on AI and the job market further shows that there’s been no macro evidence that AI has disrupted employment levels, unemployment composition, or job transitions since 2022 when ChatGPT was released. Effects on the job market are coming from elsewhere, not AI, for now.
See also:
US economic growth has decoupled from employment.
Some are noticing AI may be creating a surplus in traditional roles and a shortage in AI skills. Hiring alone won’t fix this – we need reskilling.
Robots in the storm… and in the kitchen
Figure unveiled its third-generation humanoid this week and the first it claims can be manufactured at scale for real homes. Nearly every component was re-engineered for mass production and the company says they’ve rebuilt their supply chain with the goal to manufacture up to 100,000 humanoid units in the next four years. The focus is no longer on proving a robot works, but on proving it scales.
Meanwhile, China’s DEEP Robotics introduced an all-weather humanoid built for outdoor industrial use, rated to operate from –20°C to 55°C in real-world conditions. Last week, we wrote about the robotics flywheels and how China built a robotics ecosystem over the past decade that the West may have a hard time matching.
See also:
Humanoids as attack vectors. Security researchers have discovered that the Unitree G1 humanoid robot is hackable through Bluetooth flaws and can secretly streams its sensor data to external servers (including servers located in China).
Robots will come in all shapes and sizes:
Elsewhere
Across AI, technology, science and culture…
I talk the future of the web and AI agents with Cloudflare’s boss Matthew Prince:
Battery storage is now among the cheapest and fastest-deployable new energy capacity in the US.
Avoiding contrails on just a few percent of flights could halve aviation’s near-term warming at a cost of only pennies per passenger. (via
)People with stronger, more diverse social connections show slower biological aging.
America builds AI girlfriends and China builds AI boyfriends. This is a fascinating look at culture, companionship and geopolitics intertwining.
“In addition to customizing AI companions’ personalities, users can generate themes, plots, and side stories, deepening immersion for themselves and others.” Source: ChinaTalk Thanks for reading.
ARC-AGI-1 and ARC-AGI-2 are subsets of the Abstraction and Reasoning Corpus (ARC) benchmark designed to test whether models can infer abstract rules and generalize from very few examples – tasks that humans typically solve through pattern reasoning rather than memorization.
So true, but the theme is definitely, "extraordinary gifts (energy, market cap) will require extraordinary results (cure cancer, robots in the home supporting older people to live well) etc...
Keep the insights coming!
These are such pertinent societal points but do the big tech platforms have the embedded social conscience to factor these into their business models? These have evolved within the premise of light-handed regulations and largely eschewed “beyond compliance” participation in the broad, messy issues of human impacts.
Whether these issue arise from neurological plasticity and digital/social media consumption, crimes in the metaverse that harm real people in the physical world, or energy, utility and related built infrastructure consequences of the insatiable appetite of AI/quantum compute, it's a challenge to see social good as well as oligopolistic instincts of tech insiders driving next steps.
Casting the potential scenarios for non-tech publics to explore present costs and future benefits in social outcomes as well as privately held profits would help reduce the destructive friction of progress.
My non-tech 2 cents!