- The GPU
- Posts
- Issue #1: A $333M raise, 485,000 Hoppers, and a glimmer of AGI
Issue #1: A $333M raise, 485,000 Hoppers, and a glimmer of AGI
Feat. Vultr, OpenAI, Switch x Oklo, Ori, the GB200, and Microsoft
I’m Ben Baldieri and welcome to the first issue of The GPU.
Every week, I’m bringing you a sharp, no-fluff rundown of what’s shaping GPU compute, AI infrastructure, and the data centres powering it all.
From the power-hungry data centres keeping AI running, to NeoClouds outmaneuvering hyperscalers, and the software squeezing more from every GPU, this is your guide to understanding the moves that matter.
There’s a lot of noise out there. My goal? To cut through it.
Here’s what’s in today’s issue:
Vultr’s $333M raise to take on the big guys in AI cloud.
OpenAI’s o3 model, why it could redefine AI reasoning.
Switch betting on nuclear to feed AI’s insatiable energy appetite.
Ori making AI inference a breeze (and why hyperscalers should worry).
NVIDIA’s GB200 production delays: What’s the fallout?
Microsoft snapping up 485,000 Hopper GPUs, and outpacing everyone.
Let’s get into it.
The GPU Audio Companion Issue #1
Want the GPU breakdown without the reading? The Audio Companion does it for you—but only if you’re subscribed. Fix that here.
Vultr’s $333M AI Cloud Expansion
Vultr just closed a $333M funding round, valuing the company at $3.5B.
⛰️ Vultr has reached an exciting milestone: We’ve completed a growth financing round at a valuation of $3.5 billion! This achievement underscores our role as the largest self-funded cloud infrastructure company. Learn more here: blogs.vultr.com/financing2024
— Vultr (@Vultr)
6:10 PM • Dec 19, 2024
If you don’t know them, they’re one of the more established neoclouds. This is their first big raise, and it’s coming with some serious ambition. They’re scaling their GPU inventory, adding to their 32 global data centres, and making a direct play to steal workloads from AWS and Google.
Vultr isn’t alone.
NeoCloud providers like CoreWeave and Lambda Labs are all gunning for hyperscaler business, offering high-performance GPU compute at prices hyperscalers can’t always match.
Why this matters:
NeoClouds are fast, nimble, and laser-focused on AI. AMD Ventures’s involvement in this round signals it wants a bigger slice of the AI hardware market, especially as they’re stating they want to become Vultr’s “preferred” hardware partner.
Clearly this goes beyond just GPUs.
It’s about eating into NVIDIA’s dominance, one workload at a time.
OpenAI Unveils o3 Reasoning Model
OpenAI just ushered in a new frontier in reasoning.
Today OpenAI announced o3, its next-gen reasoning model. We've worked with OpenAI to test it on ARC-AGI, and we believe it represents a significant breakthrough in getting AI to adapt to novel tasks.
It scores 75.7% on the semi-private eval in low-compute mode (for $20 per task… x.com/i/web/status/1…
— François Chollet (@fchollet)
6:09 PM • Dec 20, 2024
The new o3 model scored 75.7% on ARC-AGI in low-compute mode ($20 per task) and 87.5% in high-compute mode (costing thousands).
OpenAI is saying these capabilities are new territory, built for adapting to novel tasks and solving problems AI hasn’t seen before.
Why this matters:
AI reasoning is the next big thing.
If o3 can handle novel tasks with fewer resources, it changes what’s possible for businesses deploying AI. But with costs scaling from affordable to astronomical, this could open up a whole new conversation about the economics of cutting-edge AI.
Good for hardware owners. Less good for the rest of us.
Switch Goes Nuclear
Oklo and @Switch are partnering to deploy 12 gigawatts of clean energy—one of the largest corporate power agreements ever signed.
Additional important information can be found at:
oklo.com/newsroom/news-…— Oklo (@oklo)
1:40 PM • Dec 18, 2024
These aren’t your grandfather’s reactors. They’re autonomous, small, and designed to run for 20 years without refuelling. And with AI’s power demands pushing everyone to rethink energy, Nuclear isn’t just a sustainability play.
It’s core to reliability when downtime isn’t an option.
Why this matters:
AI workloads need power, and lots of it. Switch’s 40 global data centres are already at the cutting edge, so the bet on nuclear shows just how critical stable, clean energy is becoming.
Expect more players to follow if this works.
Ori Redefines NeoCloud Flexibility
Ori just made deploying AI models as easy as a few clicks.
📢 Announcing Ori Inference Endpoints, an easy and scalable way to deploy state-of-the-art machine learning models as API endpoints.
Discover seamless AI inference with Ori Inference Endpoints ->
— Ori (@origlobalcloud)
1:39 PM • Dec 18, 2024
Their new Inference Endpoints let you spin up models like Llama 3 or Qwen, choose your GPU (H100 SXM, PCIe, or even L40S), and auto-scale as needed.
What makes this interesting? Their per-minute pricing model starts at $0.021 and scales to zero when you’re not using it.
Infrastructure without the overhead.
Why this matters:
NeoCloud providers like Ori are solving problems hyperscalers haven’t cracked:
Agility
Simplicity
Cost control
If you’re experimenting with generative AI, this is the kind of infrastructure that makes it painless.
NVIDIA’s GB200 Delays
NVIDIA’s GB200 system was supposed to be the next big thing for high-performance AI.
🔥The demand for AI servers continues to surge: #NVIDIA's #GB200 rack-mounted solutions gain attention, but the high-spec design brings new challenges. How will the supply chain respond? 💡More analysis from #TrendForce on key drivers: 🔗
— TrendForce (@trendforce)
4:00 PM • Dec 17, 2024
Now?
It’s delayed until Q2–Q3 of 2025.
The holdup?
Advanced manufacturing for its cutting-edge thermal and interconnect designs.
Why this matters:
Hyperscalers like AWS, Google, and Tier 1 neoclouds like Coreweave rely on NVIDIA’s hardware.
Delays could mean missed AI infrastructure deadlines, opening the door for AMD to grab some market share or tier 2 neoclouds room to compete.
Microsoft’s GPU Power Grab
Microsoft bought 485,000 NVIDIA Hopper GPUs in 2024.
With demand outstripping supply for Nvidia's most advanced chips, Microsoft’s hoard has given it an edge in the race to build the next generation of AI systems. on.ft.com/41yWBPc
— Financial Times (@FT)
7:30 AM • Dec 18, 2024
That’s twice as many as Meta (224,000) and way ahead of AWS (196,000) and Google (169,000).
Satya Nadella isn’t playing.
Why this matters:
The H100 is built for generative AI, with 80 billion transistors and 30 teraflops of FP64 performance. Even with NVIDIA’s Blackwell on the horizon, demand remains sky-high.
AI is the next gold rush, and Microsoft is ensuring they’ve got the infrastructure lead to dominate it.
The Rundown
It’s been a busy week.
Vultr’s $333M raise shows NeoClouds are winning on agility and cost. OpenAI’s o3 is breaking ground in reasoning. Microsoft’s 485,000 H100 GPUs cement their lead in AI infrastructure. Switch went nuclear, and NVIDIA’s GB200 delays show even giants have their limits.
The AI arms race is here. Reply to this email, and let me know where you think it’s headed.
ICYMI
I was invited to write a Christmas OpEd for The Tech Capital (thanks, Jack!).
Check it out here:
Keep The GPU Sharp and Independent
Good analysis isn’t free. And bad analysis? That’s usually paid for.
I want The GPU to stay sharp, independent, and free from corporate fluff. That means digging deeper, asking harder questions, and breaking down the world of GPU compute without a filter.
If you’ve found value in The GPU, consider supporting it.
☕ Buy Me A Coffee → https://buymeacoffee.com/bbaldieri
It helps keep this newsletter unfiltered and worth reading.
Reply