The AI industry runs on three scaling laws.
More data, more parameters, more compute makes a better model. Fine-tuning and distillation after training make it cheaper. More compute at inference makes it smarter. Every GPU cluster, every neocloud raise, every memory chip, every data centre permit fight in every single issue since the birth of this publication exists because of those three relationships.
All three assume the same thing: the model starts with human data.
This week, the researcher behind AlphaGo raised $1.1 billion to test whether that assumption is wrong.
I'm Ben Baldieri, and every week I break down the moves shaping GPU compute, AI infrastructure, and the data centres that power it all.
Here's what's inside this week:
Let's get into it.
The GPU Audio Companion Issue #104
Want the GPU breakdown without the reading? The Audio Companion does it for you, but only if you’re subscribed. If you can’t see it below, click here to fix that.
Ineffable Intelligence Raises $1.1b to Pursue Non-LLM AI
The post-LLM thesis now has $1.1 billion behind it.
Ineffable Intelligence raised $1.1 billion in a seed round led by Sequoia at a $5.1 billion valuation, making it Europe's largest seed round ever. UK Sovereign AI and the British Business Bank co-invested on top of the private round. The company is founded by David Silver, Professor at UCL and formerly Head of Reinforcement Learning at Google DeepMind, whose track record includes AlphaGo (beat the world Go champion, 2016), AlphaZero (mastered chess, Go, and shogi from self-play alone with zero human data, 2017), and AlphaProof (solved International Math Olympiad problems, 2024).
Why this matters:
AlphaZero mastered chess in four hours by playing itself. No human games, no grandmaster data, just the rules and compute. It surpassed all human knowledge of chess in an afternoon. Silver is betting that approach generalises into science, medicine, and engineering. If it does, the LLMs generating $30 billion a year in revenue are the equivalent of training a chess engine on grandmaster games when you could let it play itself. LLMs will still exist, but the question is whether they hit a ceiling that reinforcement learning does not.
The AI industry's three scaling laws all describe ways to make LLMs better: pretrain on more human data, refine with post-training techniques like distillation and RLHF, or apply more compute at inference for chain-of-thought reasoning. Every GPU cluster in this issue is optimised for one of those three relationships. Silver's work implies a fourth: experience scaling, where performance improves as the agent interacts with a simulated environment more, generating its own training signal through trial and error. The compute requirements are real (AlphaZero used thousands of TPUs for self-play), but the workload is fundamentally different. LLM training is one massive parallel job across the cluster. RL training is millions of small sequential loops, each one latency-sensitive, and bottlenecked by how fast you can simulate the environment rather than how many GPUs you can rack. If experience scaling works for real-world problems, the winning infrastructure shifts from "biggest cluster" to "fastest simulation," and the hundreds of billions committed to LLM-optimised clusters may be aimed at the wrong scaling relationship.
Sequoia leading this round is the signal that matters. The most successful venture firm in AI history is hedging against the very buildout it helped fund. That is not a research grant or a government initiative. That is $1.1 billion of private capital placed behind the one researcher most qualified to test whether LLMs are the destination or just the first stop.
China Orders Meta to Unwind a Completed $2 Billion AI Acquisition
The founders are barred from leaving China.
China's NDRC ordered Meta to unwind its $2 billion+ acquisition of Manus, the AI agent startup that built one of the most widely adopted agent systems in production. Manus raised $75 million from Benchmark in May 2025, then shut its China offices, laid off staff, and moved to Singapore in an attempt to bypass both US investment restrictions on Chinese AI firms and Chinese rules limiting IP transfer overseas. The move did not work. Co-founders CEO Xiao Hong and chief scientist Ji Yichao were summoned to Beijing in March and barred from leaving the country, even as Manus staff have already moved into Meta's Singapore offices and projects continue.
Why this matters:
This is the first time Beijing has ordered a completed US tech acquisition unwound. Last week the White House accused China of industrial-scale distillation of US frontier AI (Issue #103). This week Beijing ordered Meta to return a Chinese AI startup. Both governments are now blocking the other's access to AI intellectual property simultaneously, and the companies caught between them are the ones getting crushed.
"Singapore washing" just became a regulatory risk factor for every Chinese AI startup with Western investors. Manus moved to Singapore specifically to access foreign capital and sidestep both countries' restrictions. Beijing reached through the Singapore incorporation and ordered the deal unwound anyway, asserting jurisdiction based on the founders' nationality, the technology's origin, and the original R&D location rather than the place of incorporation.
Losing Manus does not kill Meta's agent strategy. Muse Spark, the Broadcom MTIA partnership, the CoreWeave $21B agreement, and data centre #32 in Tulsa are all intact. But it removes the team behind Manus's agent tooling, and it sends a warning to every hyperscaler considering acquisitions of Chinese-origin AI talent, regardless of where that talent has incorporated.
Oaktree Leads $350 Million for a Neocloud That Had 432 GPUs Six Months Ago
Sharon AI's capital structure is maturing faster than anyone expected.
Sharon AI (NASDAQ: SHAZ) entered into definitive agreements for $350 million in 6% Convertible Senior Notes due 2031, led by Oaktree Capital Management with its Value Opportunities strategy, alongside Two Seas Capital and other institutional investors. The notes carry a conversion price of approximately $48.24, representing a 20% premium to the market price, with founders locked up until March 2027. Latham & Watkins is advising Oaktree, a detail that signals this was structured as a proper institutional credit transaction rather than a growth-stage financing.
Why this matters:
We have tracked Sharon AI across the catalogue. In Issue #100, they had 432 GPUs online and a Canva contract. In Issue #102, we contrasted them with NewBird AI as proof that small starts are not disqualifying if the infrastructure DNA is right. Six months later, Oaktree is leading $350 million in convertible notes. The capital structure has outpaced what most observers thought possible for a company at this scale.
Oaktree manages $190 billion and is one of the largest credit investors in the world. Its involvement means the credit market is now underwriting GPU infrastructure at every scale tier, not just at CoreWeave ($8.5B A3-rated debt) or Firmus ($10B Blackstone facility) level. The asset class is institutionalising from the top down and the bottom up simultaneously.
Compare this to NewBird AI: $50 million, no named hardware, no named counsel. Both are NASDAQ-listed GPU plays. The substance between them is separated by two orders of magnitude, and the market is starting to tell the difference. Oaktree is the kind of investor that does not write cheques based on narratives. It writes them based on contracted revenue and asset value.
NVIDIA Gives Away Its Best Open Model Because Every Download Needs NVIDIA Hardware
The model is free because the infrastructure to run it is not.
NVIDIA launched Nemotron 3 Nano Omni, a 30B-A3B hybrid mixture-of-experts model that processes text, images, audio, video, documents, and graphical interfaces in a single pass. The weights, datasets, and training techniques are all open. It tops six leaderboards for document intelligence, video, and audio understanding, and delivers 9x higher throughput than comparable open omni models. Foxconn, Palantir, Docusign, Oracle, and Infosys are among the companies adopting or evaluating it, and the Nemotron family has passed 50 million downloads in the past year.
Why this matters:
The strategy is straightforward: NVIDIA does the expensive work of training frontier-class models on massive GPU clusters, then gives them away for free. Enterprises download the model, fine-tune it on their own proprietary data using NVIDIA's NeMo tools running on NVIDIA GPUs, and deploy on NVIDIA inference hardware. No API fees, no per-token charges, and the enterprise's data never leaves their own infrastructure. Every download is a customer who needs NVIDIA hardware to fine-tune and serve it.
This is the data sovereignty problem solved from the hardware side. Anthropic solves it from the cloud with Cowork on 3P (Issue #103). Google solves it with confidential Blackwell VMs. NVIDIA solves it by putting the model in your hands and letting you own everything. For defence, healthcare, and finance, where sending prompts to a third-party API is a compliance problem, owned hardware running open weights is the path of least regulatory resistance.
Nano Omni is not a chat model. It is a perception sub-agent designed for swarms, functioning as the "eyes and ears" alongside larger models that handle reasoning and execution. That means more data in, better decisions out. In theory, at least.
Qualcomm Lands Its First Data Centre Customer and Won't Say Who
A fifth silicon vendor just entered the AI data centre market.
Qualcomm beat Q1 earnings estimates on Wednesday and disclosed a data centre product win with an unnamed hyperscaler, with revenue expected by the end of 2026. The stock surged 15%. CFO Akash Palkhiwala called it "the first of many" and said Qualcomm would provide more detail at its analyst day next month. Qualcomm has been developing data centre silicon for several quarters but has not previously disclosed a customer of this scale.
Why this matters:
The AI data centre silicon market just added a fifth vendor. NVIDIA dominates. AMD has Meta's 6GW deal (Issue #93). Broadcom builds custom silicon for Meta (MTIA) and Google (TPUs). Amazon designs Trainium and Graviton in-house. Qualcomm entering with a hyperscaler win means there is now a mobile-first chip designer competing for inference workloads alongside the incumbents, and the unnamed customer is large enough to move Qualcomm's stock 15% on the announcement alone.
Qualcomm's Arm-based architecture is optimised for power efficiency, which is the binding constraint in every data centre story in this issue. The floating data centre in Singapore, the rejected facility in Minnesota, the $4 billion project that died over zoning are all fundamentally about power. A chip designed for mobile power envelopes entering data centres at hyperscaler scale suggests inference workloads are fragmenting across silicon architectures faster than the market expected.
"The first of many" from the CFO is the signal to watch. If Qualcomm lands a second hyperscaler before analyst day, the competitive dynamics in AI inference silicon shift significantly. NVIDIA's inference pricing power depends on being the only option at scale. Every new silicon vendor that wins a hyperscaler contract compresses that margin.
Hyperscalers Report Q1 Earnings Into a Memory Crisis They're Absorbing Without Raising Prices
H200 spot prices are up 68% since January. Memory prices are even higher. The hyperscalers are eating the cost.
Microsoft, Alphabet, and Meta all reported Q1 earnings this week against a backdrop of spiking energy costs from the Iran conflict, a worsening global memory shortage, and collective AI infrastructure capex now exceeding $500 billion for 2026. DRAM is forecast to cost $9.71 per gigabyte this year, up from $3.76 in 2025, a 158% increase that has already forced Microsoft to raise Surface PC prices by hundreds of dollars. Spot prices for NVIDIA H200 GPUs reached $3.82 per hour this month, up from $2.27 in January, according to data from Ornn, the startup building a GPU compute exchange. Amazon CEO Andy Jassy defended $200 billion in 2026 capex in his shareholder letter: "We're not going to be conservative in how we play this." AWS has no plans to raise prices despite increased costs.
Why this matters:
The hyperscalers are absorbing cost increases that would break smaller operators. DRAM up 158%. H200 spot up 68%. Oil prices spiking from the Iran conflict. Helium shortages hitting semiconductor manufacturing. And yet AWS is not raising prices, Microsoft is not raising cloud prices, and none of the three have revised capex guidance downward. The message: demand for AI compute is strong enough that the hyperscalers will eat hundreds of billions in cost increases rather than risk slowing adoption by passing them through.
Ornn's GPU pricing data is now being cited by CNBC as the reference for H200 spot markets. That price signal matters because it is the closest thing the market has to a real-time clearing price for GPU compute. The 68% increase from $2.27 to $3.82 per hour since January tells you that GPU scarcity is intensifying even as new capacity comes online. Supply is growing. Demand is growing faster.
$500 billion in collective hyperscaler AI capex for 2026 is the number that makes the Ineffable thesis so pointed. That is half a trillion dollars committed to LLM-era infrastructure in a single year. If experience scaling works for real-world problems, some portion of that spend is optimised for the wrong workload. If it does not, the hyperscalers have correctly priced the largest infrastructure buildout in history. Both outcomes are possible. Neither can be unwound.
Google Breaks Ground on a $15 Billion Gigawatt-Scale AI Hub in India, Utah Residents Learn About a 40,000-Acre Data Centre on Facebook
The biggest bottleneck in 2026 is not GPUs, power, or capital. It is permission.
Google held a groundbreaking ceremony on Monday for a $15 billion, gigawatt-scale "AI hub" in India comprising three campuses across Tarluvada, Adivivaram, and Rambilli-Achyutapuram. AdaniConnex and Nxtra by Airtel are leading construction. Google acquired the land in early April. Adani Group committed $5 billion to the project in December. In Virginia, Compass Datacenters cancelled the Prince William Digital Gateway, citing legal and regulatory barriers. And in Utah, residents of Box Elder County learned from Facebook that a 40,000-acre data centre backed by Kevin O'Leary and the Military Installation Development Authority was about to receive final approval from their county commission. The "Stratos Project" would consume more energy than the entire state. Hundreds showed up to protest. The vote was delayed to May 4.
Why this matters:
Google's India groundbreaking shows that capital is flowing to jurisdictions that say yes. India wants the investment, has the land, and has two construction partners ready to build. Meanwhile Virginia, the largest data centre market in the world, just lost another project to regulatory barriers. Utah residents found out about a facility that would use more power than their entire state through a Facebook post the morning of the vote. The industry's social licence is eroding in the jurisdictions where it needs it most.
The Utah story is the most extreme version of the permission problem we have covered. MIDA, a military infrastructure authority, used its special powers to bypass normal planning processes for a 40,000-acre project backed by a celebrity investor. Residents had no notice. The county commission was prepared to vote the same day. Community opposition stopped it, but the fact that a project of this scale nearly reached final approval without public awareness shows how badly the siting process has broken down on both sides.
Compass in Virginia and Google in India are the same story told in opposite directions. One jurisdiction's regulatory friction is another's competitive advantage. As US permitting timelines stretch and community resistance intensifies, the gigawatt-scale projects that define the next phase of AI infrastructure are increasingly being sited in countries where approval is faster and opposition is weaker. The capex follows the permit.
The Rundown
Read the stories in this issue back to back and count the assumptions.
China blocked a $2 billion deal to keep AI talent inside its borders. Oaktree underwrote $350 million in GPU procurement. NVIDIA gave away a model to sell hardware. Qualcomm landed its first data centre customer and the stock jumped 15%. The hyperscalers reported Q1 earnings into a memory crisis, with H200 spot prices up 68% since January and DRAM costs up 158%, and none of them raised prices or cut capex. Google broke ground on a $15 billion AI hub in India. Utah residents found out about a 40,000-acre data centre from Facebook. Virginia cancelled another project. Every one of those stories prices against the same three scaling relationships this issue opened on.
The GPU clusters, the memory chips, the new silicon vendors, the $500 billion in collective capex are all infrastructure for those laws.
David Silver raised $1.1 billion on the thesis that a fourth relationship matters more.
AlphaZero did not scale on human data. It scaled on self-play. The constraint was simulation speed, not dataset size. Sequoia backed that thesis with the largest seed round in European history, and the UK government added sovereign capital on top.
LLMs are not disappearing.
The revenue is real, the users are real, and the silicon sells faster than it can be manufactured.
But if experience scaling reaches problems that the other three cannot, then the $500 billion being committed this year is optimised for the wrong law. Not wasted. Just aimed at a ceiling that a different scaling relationship may not have.
See you next week.


