Sharon AI just went public on 432 GPUs and three customers.

That's not a typo. The Australian neocloud priced its Nasdaq IPO at $30 per share this week, raising $125M in gross proceeds. The prospectus is a window into the economic guts of the smaller neoclouds.

And a stark reminder of just how early this market really is.

Here's what else is inside this week:

Let’s get into it.

The GPU Audio Companion Issue #91

Want the GPU breakdown without the reading? The Audio Companion does it for you, but only if you’re subscribed. If you can’t see it below, click here to fix that.

Sharon AI Lists on Nasdaq in $125M IPO with 432 GPUs

The first Australian neocloud has gone public in the US. The prospectus makes for interesting reading.

Sharon AI priced its Nasdaq IPO at $30 per share, raising $125M in gross proceeds led by Oaktree Capital Management and Two Seas Capital. The prospectus disclosed that as of September 2025, Sharon AI had 432 GPUs and 195 CPUs online across three colocation data centres in Australia, with 99% of 2024 revenue coming from just three customers. This, coupled with the recent funding rounds and Texas data centre divestiture, makes for interesting timing, to say the least.

Why this matters:

  • Two Australian neoclouds making major capital markets moves in back-to-back weeks. Firmus locked in $10B in project debt last issue (Issue #89). Sharon AI has taken the public equity route. With Firmus deploying thousands of GPUs across a 1.6GW national rollout, and Sharon AI going public on 432 GPUs with near-total customer concentration, the contrast is stark.

  • The prospectus risk factors read like a checklist of neocloud fragility: limited operating history, ongoing losses, capex intensity, supply chain exposure, and significant customer concentration. And while extreme customer concentration isn’t necessarily a deal-breaker in this industry (just look at NVIDIA), losing one would likely have a material impact. Sharon AI openly acknowledged this.

  • Public markets unlock deeper liquidity and lower capital costs on the one hand, but quarterly reporting forces transparency on metrics that private competitors never have to disclose on the other. That's the tension in taking a neocloud public. So, credit where credit is due to Sharon AI for choosing this path. If they succeed, we may yet see the floodgates open for small-neo IPOs.

Mistral Acquires Koyeb to Buildout Full-Stack AI Cloud

The French model builder just made its first acquisition, and it wasn't a model company.

Mistral AI has acquired Koyeb, a Paris-based serverless deployment platform founded by three former Scaleway engineers. Koyeb's 13-person team will join Mistral's engineering org under CTO Timothée Lacroix, with the platform transitioning into a core component of Mistral Compute. The deal came days after Mistral announced $1.4B in Swedish data centre investment and follows the company crossing $400M in ARR.

Why this matters:

  • Mistral's trajectory tells a story about where the value is moving. The company started as a pure model builder. Then it was a FluidStack customer renting GPU capacity (Issue #23). Then it became an NVIDIA Cloud Partner. Now it's acquiring infrastructure teams, investing in its own data centres, and working with HSBC on enterprise deployments (Issue #77).

  • Koyeb's serverless platform solves a specific problem Mistral couldn't build fast enough internally: on-premises deployment and GPU orchestration for enterprise clients. As Mistral pushes deeper into regulated industries, the ability to deploy models on a customer's own hardware becomes the dealbreaker.

  • The political tailwind behind non-US AI infrastructure is starting to generate real M&A activity, not just press releases.

AMD Backstops $300M Crusoe Loan for AMD GPUs

AMD is allegedly financing customers to win GPU deployments.

Goldman Sachs, per reporting from The Information and then Reuters, is providing Crusoe with a $300M loan backed by AMD chips and related equipment, with AMD guaranteeing the leaseback of its own GPUs if Crusoe can't find customers. The backstop reportedly allowed Crusoe to lock in a roughly 6% interest rate, likely significantly below what it would have secured without the guarantee.

Why this matters:

  • NVIDIA has long been investing in customers, backstopping loans, and creating circular financing structures to lock in GPU orders. AMD doing the same thing confirms that this route to market is becoming the go-to model for chip vendors competing for neocloud deployments. The chip sale is now a financial product, not just a hardware transaction.

  • The economics tell you how hard AMD is pushing. A 6% rate on GPU-backed debt is cheap. If the reporting is true, AMD effectively had to put its own balance sheet on the line, offering to lease back chips it already sold, to get Goldman comfortable at that price. That's a vendor saying, "We believe in this hardware enough to eat it if nobody else wants it."

  • Bold. Especially when it also means AMD is carrying contingent liability for Crusoe's commercial success.

Google, Anthropic Ship New “Most Capable Models Yet”

Two frontier releases in the same week, both rewriting the price-performance curve.

Google rolled out Gemini 3.1 Pro with a 77.1% score on ARC-AGI-2, more than double Gemini 3 Pro's reasoning performance on the benchmark designed to resist memorisation. Anthropic released Claude Sonnet 4.6 at unchanged Sonnet pricing ($3/$15 per million tokens), with internal testing showing developers preferred it over the previous Opus 4.5 flagship 59% of the time. Both models shipped across consumer, developer, and enterprise surfaces simultaneously.

Why this matters:

  • Two major model upgrades in one week, each deployed to hundreds of millions of users across multiple product surfaces. Every generation that ships drives another wave of inference demand, and both companies are compressing the gap between releases to months, not quarters.

  • Price-performance keeps collapsing. What required Anthropic's Opus-class pricing three months ago now ships at Sonnet rates. For API customers buying inference at scale, the cost per unit of useful intelligence is falling fast. That's deflationary pressure on inference revenue across every provider hosting these models.

  • OpenAI took a different approach entirely this week. Rather than shipping a new model, it hired OpenClaw creator Peter Steinberger to build consumer-facing agents. Because sometimes, hype and ready-made distribution are more important than rankings.

G42, Cerebras to Deploy 8 Exaflop Supercomputer in India

The UAE is building India's largest AI supercomputer, and it's not running on NVIDIA silicon.

G42 and Cerebras have announced an 8 exaflop AI supercomputer to be hosted in India, delivered in partnership with MBZUAI and India's Centre for Development of Advanced Computing (C-DAC). The system will operate under Indian governance frameworks with all data remaining within national jurisdiction, serving as a foundational asset under the India AI Mission. The announcement came on the sidelines of the AI Impact Summit 2026 in New Delhi, following the UAE President's January visit to India. Once operational, the supercomputer will be accessible to institutions, startups, SMEs, and government ministries.

Why this matters:

  • An 8 exaflop deployment in India is the second major Cerebras validation in two weeks, after OpenAI shipped its 1,000 tk/s Codex-Spark model on Cerebras hardware (Issue #89).

  • India is suddenly at the centre of multiple AI infrastructure plays in the same week. Submer's Radian Arc signed two Indian partnerships to embed GPU edge infrastructure in telecom and colocation networks (see below). Now G42 is deploying exaflop-scale national compute. The common thread is sovereignty: every deal is structured around Indian data residency, Indian governance, and Indian access.

  • The UAE-India corridor is becoming a significant AI infrastructure axis. Abu Dhabi is positioning itself as a sovereign AI infrastructure exporter, building compute capacity in partner nations rather than asking those nations to send their data to the Gulf. That's a different model from the US hyperscalers and, for now, it's winning deals.

Meta Goes Full-Stack NVIDIA

NVIDIA has announced a multiyear, multigenerational partnership spanning millions of Blackwell and Rubin GPUs, GB300-based rack-scale systems, and the first large-scale deployment of Grace-only CPUs in production. Meta has also adopted Spectrum-X Ethernet across its infrastructure footprint, a significant shift for a company that built its reputation on open networking through Facebook Open Switching System. The deal extends to Vera CPUs for large-scale deployment in 2027 and to NVIDIA Confidential Computing for WhatsApp's private processing.

Why this matters:

  • Meta has historically been the poster child for open infrastructure: Open Compute Project, custom switching, disaggregated everything.

  • Going full-stack NVIDIA, from silicon to network fabric, signals that the performance and integration requirements of frontier AI training have overtaken the open-source hardware philosophy.

  • NVIDIA's push into Ethernet networking has previously faced scepticism from competitors and operators loyal to them. Meta validating Spectrum-X at hyperscale changes that conversation overnight.

Submer's Radian Arc Signs Two Indian Partnerships Days After Acquisition

The ink on the Radian Arc deal barely dried before the India expansion started.

Radian Arc, now operating under Submer's InferX platform, has signed a binding Letter of Intent with GTPL Broadband to deploy GPUaaS across India's telecom network, and a separate MOU with Datasamudra to establish a GPU point-of-presence inside its Karnataka data centre at KIADB IT Park in North Bangalore. The GTPL deal is a three-year collaboration targeting universities, governments, and enterprises. The Datasamudra partnership focuses on sovereign compute and public-sector workloads with explicit compliance to India's data residency regulations. Both deals position Radian Arc's GPU edge infrastructure inside existing local networks rather than building standalone facilities.

Why this matters:

  • We covered the Radian Arc acquisition last week (Issue #89) and flagged that the real question was how much of Submer's 5GW pipeline converts to operating capacity. Two signed deals within days of closing suggest the acquisition came with a warm pipeline and provide some support to the 5GW claim.

  • The India strategy is embed-and-expand, not build-and-own. Both partnerships deploy GPU infrastructure inside someone else's network: GTPL's telecom footprint and Datasamudra's colocation facility.

  • That keeps capex light while giving Submer sovereign compute positioning in a market where data residency requirements are tightening. India's government has been increasingly explicit about wanting AI workloads processed domestically.

The Rundown:

Cast your mind back twelve months.

Beyond NVIDIA and the hyperscalers, the AI infrastructure conversation had about 3 names: CoreWeave, Lambda, and maybe Crusoe. They dominated the headlines, the funding rounds, and the discourse. Then came Nebius, Fluidstack, and Nscale, moving aggressively from tier-2 into the big leagues. If you were tracking independent AI infrastructure, that was the list.

Now look at the last couple of weeks.

We have new old names, and they’re making serious moves.

Sharon AI went public. Firmus locked in $10B last week. Submer closed an acquisition and signed two Indian deals within days. G42 and Cerebras are deploying an exaflop supercomputer without a hyperscaler in sight. Mistral bought its way from model builder to full-stack cloud. AMD backstopped debt just to get chips into racks.

New names, new geographies, new capital structures, and new routes to market.

Every week, new companies that didn't exist in the conversation a year ago are surfacing, raising serious money, braving the public markets, and moving beyond MOUs and LOIs into real contracts.

CoreWeave and Lambda aren't going anywhere, but the era where two or three US neoclouds defined the category is over.

The buildout has spread too far, into too many jurisdictions, with too many different financing models, for any small group to own the narrative.

That's what 2026 looks like so far. More names. More places. More ways to get it done. And to think, it’s only February.

See you next week.

ICYMI

I profiled Mirantis this week - check that out here:

Reply

Avatar

or to participate

Keep Reading