And just like that, half of xAI's founding team is gone.
Two co-founders left this week. That leaves six of the original twelve after just three years. Elon Musk’s response was to hold an all-hands about building an AI satellite factory on the moon.
With a giant catapult launch system.
Because what better way is there to deal with the loss of a potentially winning edge in a market where the difference between victory and defeat could literally come down to the individual brilliance of a single AI researcher, let alone three?
I'm Ben Baldieri, and every week I break down the moves shaping GPU compute, AI infrastructure, and the data centres that power it all.
Here's what else is inside this week:
Let's get into it.
The GPU Audio Companion Issue #89
Want the GPU breakdown without the reading? The Audio Companion does it for you, but only if you’re subscribed. If you can’t see it below, click here to fix that.
xAI Loses More Founding Team Members, Gains Space Catapults
Six of the original twelve xAI co-founders are gone.
Tony Wu and Jimmy Ba both announced their departures this week. Wu on Monday, Ba on Tuesday. The Financial Times reported Ba left amid internal tensions over efforts to improve xAI's models as Musk races to catch up with OpenAI and Anthropic. Musk then held an all-hands Tuesday night.
Per the New York Times, he told employees xAI needs a lunar manufacturing facility - a factory on the moon that will build AI satellites and launch them via catapult. While also claiming the departures may have been a “push, not a pull”. Not unexpected given the reorg, but given the number of departures and the timescale in which they happened, the lasting impact of xAI’s relative position remains to be seen.
Why this matters:
The departures come weeks after SpaceX completed its all-stock acquisition of xAI, creating a combined entity valued at $1.25 trillion (Issue #87). Everyone involved is in line for a significant windfall from a potential IPO targeting $1.5T later this year.
It's a natural exit point, for sure. But five departures in twelve months, plus a weekly wave of senior staff at the same time as a major direction change, likely isn't just founders cashing out.
xAI's Grok chatbot has faced regulatory probes in multiple jurisdictions after its image generator allowed mass creation of non-consensual explicit deepfake images. The kind of thing that creates friction on a technical team, emotional and otherwise. That means the product issues and the talent drain may not be unrelated.
Firmus Secures $10B to Build Australia's AI Factories
One of the largest private debt financings in Australian history for a company that most of the industry hasn't yet heard of.
Firmus locked in $10B in debt financing led by Blackstone Tactical Opportunities, Blackstone Credit & Insurance, and Coatue to fund Project Southgate, a national rollout of AI factories built on NVIDIA's DSX reference architecture. The facilities are already under construction across multiple Australian sites with thousands of GPUs planned, scaling to 1.6GW of infrastructure through 2028. Each factory is purpose-built around energy efficiency and token production. Blackstone's John Watson framed it as "picks and shovels": one of the firm's highest conviction themes.
Why this matters:
We've been tracking Firmus since the Singapore seawater cooling MoU (Issue #51) and the AU$500M Southgate expansion (Issue #74). This is a step change in scale. Now, with $10B in debt, not equity, this is a step change in scale for the Australian neocloud.
This deal allows Firmus to keep its cap table clean while scaling to 1.6GW, approaching the kind of capacity commitments we've seen from CoreWeave and Lambda in the US.
Australia is positioning as a serious AI compute destination. Competitive energy pricing, skilled workforce, and distance from the US-China export control flashpoints make it attractive for customers who need capacity outside traditional hubs. We've tracked the geographic diversification of AI infrastructure across the UK (Issue #86), the Middle East (Issue #74), and the Global South (Issue #44). Australia is the next node.
Submer Acquires Radian Arc For Full-Stack Edge Play
From liquid cooling to full-stack AI infrastructure in under two years. That's the Submer trajectory.
Barcelona-based Submer is acquiring Radian Arc Operations, a telco-focused GPU IaaS platform deployed across 70+ telecom and edge compute customers globally with thousands of GPUs in operation. The deal merges Radian Arc's carrier-embedded edge platform with InferX, Submer's NVIDIA Cloud Partner (NCP) neocloud launched earlier this year. The combined footprint now spans North America, Europe, the UK, India, the Middle East, and Asia-Pacific.
Why this matters:
Submer's evolution mirrors what we're seeing across the infrastructure stack. Companies that started with a single capability, liquid cooling in Submer's case, are acquiring their way into full-stack positions, with each acquisition layer driving down costs on the one hand, and adding recurring revenue and customer lock-in potential on the other.
Radian Arc’s position as a telco embeds GPU infrastructure inside carrier networks. That means data stays in-country, billing runs through telco systems, and latency potentially drops below what centralised data centres can match.
5GW of land and power pipeline is a serious claim. If even a fraction materialises, Submer is operating at a scale that puts it in conversation with the bigger neoclouds. The question, as always, is how much of that pipeline converts to operating capacity.
OpenAI Ships Cerebras-Powered, 1,000 tk/s Model
The OpenAI-Cerebras partnership just produced its first product.
OpenAI has released GPT-5.3-Codex-Spark, a smaller version of GPT-5.3-Codex and the company's first model built specifically for real-time coding. It runs on Cerebras' Wafer Scale Engine 3. And it delivers 1,000+ tokens per second - fast enough that interaction feels near-instant. The model is rolling out as a research preview for ChatGPT Pro users in the Codex app, CLI, and VS Code extension with a text-only, 128k context window with separate rate limits.
Why this matters:
This is the inference economics thesis playing out in real time. The same issue where Olix raises $220M for photonic inference chips, and Fractile invests £100M in in-memory compute, OpenAI ships a product that enables real-time collaboration - interrupting, redirecting, and iterating at conversational speed. That’s a different product, not just an iterative product improvement.
The $23B valuation in last week’s Series H was built on the thesis that wafer-scale silicon could carve out a meaningful role alongside GPUs. OpenAI's explicit validation of that premise, in production with paying users, is the strongest validation of that thesis yet.
This success feeds into the “two modes” roadmap. OpenAI described a future where Codex blends real-time interaction (Spark-class models on Cerebras) with longer-horizon reasoning (frontier models on GPUs), delegating to sub-agents in the background. If that architecture works, every coding tool will need both a speed tier and a depth tier. The infrastructure requirements and the hardware mix look very different from today's GPU-only stacks.
UK Chip Co’s Have a Moment: Olix Raises $220M, Fractile Commits £100M
Two UK chip startups, two different architectures, the same thesis: inference economics will decide who captures the next wave of AI infrastructure spend.
Olix has raised $220M led by Hummingbird Ventures, reaching a $1B+ valuation. The London-based startup is developing an Optical Tensor Processing Unit (OTPU): a photonic digital processor that performs bit-perfect computations (not analogue approximations) using a novel memory and interconnect architecture.
These funds will go to a new hardware engineering facility in Bristol for chip assembly, system testing, and a software lab. Backed by Oxford Science Enterprises, Kindred Capital, and the NATO Innovation Fund, Fractile is expanding from 70 to 110+ staff across London and Bristol.
Why this matters:
Both of these startups share a problem focus: memory. Fractile’s solution seeks to tackle the bandwidth issue with extreme compute proximity. Olix's decision to skip HBM entirely goes a step further, and is arguably the most aggressive supply chain bet in the alt-compute field.
With every other chipmaker fighting for the same constrained packaging and memory capacity that NVIDIA and AMD dominate, Olix is betting photonics lets them sidestep the queue altogether. If it works, the supply chain independence will be as valuable as the performance gains.
The inference silicon race keeps intensifying. We covered Mythic's $125M raise for analogue compute-in-memory chips, FuriosaAI landing LG with 2.25x better performance per watt than GPUs (Issue #69), and last week Positron and Cerebras raised $1.23B between them (Issue #87). Now add $320M more for UK-based alternatives. The capital flowing into non-NVIDIA inference hardware has never been higher.
$3T in AI Capex, and Nobody Can Find the Depreciation
The five biggest hyperscalers will spend $3T on property and equipment over the next four years. Wall Street can't see where the depreciation hits their income statements.
The WSJ's Jonathan Weil, the reporter who first challenged Enron's accounting in 2000, flagged the problem this week. Alphabet, Amazon, Meta, Microsoft, and Oracle don't report depreciation as a standalone line on their income statements. It's buried across cost of goods, R&D, and G&A with no consistency between companies. Investors can't model which expense lines carry the load. Morgan Stanley warned in January that consensus models don't adequately forecast the depreciation impact on earnings, and that margins may compress if revenue doesn't keep pace.
Why this matters:
Meta is already showing why this matters: its $60.5B in 2025 earnings included a $2.6B boost from extending estimated useful lives of servers and network assets. One accounting judgment. Material impact on reported profit. New US rules requiring disaggregated expense reporting won't take effect until 2028.
These companies built trillion-dollar businesses on software margins. Now they're spending like railroads but reporting like software companies. The WSJ was explicit in pointing this out: Union Pacific and Norfolk Southern report depreciation as standalone expense lines. The hyperscalers don't.
Useful life assumptions are the hidden lever. If those assumptions prove optimistic, the depreciation catch-up hits earnings all at once.
DataVita Wins Glasgow City Council's £45M Compute Contract
Scotland's largest data centre operator is on a roll.
Glasgow City Council awarded DataVita a £44.9M contract for core compute and storage services spanning five years and nine months. The deal includes options to extend beyond 10 years, with a projected total value between £ 80 M and £ 110 M. DataVita will deliver services from its Tier III-certified Scottish facilities, including DV1 at Chapelhall (40MW, 4,000 racks, PUE 1.15) and DV2 in the basement of 177 Bothwell Street, Glasgow (130 racks). The contract supports 400+ council applications and creates up to 25 new roles, including apprenticeships.
Why this matters:
DataVita is leading the Scottish AI infrastructure charge. On top of the £8.2B AI Growth Zone in North Lanarkshire, and landing CoreWeave as a tenant, DataVita is also winning the bread-and-butter enterprise compute contracts that keep the lights on.
Glasgow specifically cited DataVita's "commitment to renewable, energy-efficient operations" as aligned with the city's ambitions. DV1 runs on 100% green power with a PUE of 1.15. In an era where councils face political pressure over both digital resilience and climate commitments, that combination is a procurement advantage.
The contract is a signal about where Scottish public sector IT is heading. Away from in-house server rooms (Glasgow suffered a data centre outage over a decade ago when a fire suppression system damaged IT equipment) and toward multi-source delivery models with local providers.
The Rundown
This week's issue is really about one question: who's holding the bag when the music stops?
xAI is bleeding talent while Musk talks moon catapults. Firmus raises $10B in traditional project debt to build 1.6GW across Australia. Submer buys its way from cooling startup to full-stack neocloud via telco edge. OpenAI proves Cerebras can deliver a genuinely new product at 1,000 tokens per second. Olix and Fractile put UK inference silicon into the spotlight.
And then the accounting stories land.
There’s $3T in hyperscaler capex being depreciated in ways that nobody, including Wall Street's own analysts, can see clearly enough to model.
The through-line here is that the AI infrastructure buildout is now operating at a scale where the financial engineering is becoming as consequential as the actual engineering. Debt structures, depreciation assumptions, useful life estimates, off-balance-sheet vehicles are load-bearing walls.
And the people responsible for checking whether those walls can hold the weight with which they are being loaded are starting to have questions.
See you next week.


