Inference Inflection: Cerebras, SpaceX, Leopold's $5.5B Bet
Cerebras's $60B IPO, Anthropic's SpaceX deal, and Leopold's $5.5B fund — three sides of one inference-supply story.
Three stories ran on parallel tracks this week. On Thursday, Cerebras priced its IPO at a $60 billion valuation after a year of withdrawn filings and national-security reviews, with shares closing at $280 and the company instantly worth more than half of Intel. The week before, Anthropic signed a deal with SpaceX to take over the entire 220,000-GPU Colossus 1 cluster in Memphis — and to begin scoping orbital data centers. And buried in a Fortune profile from earlier in the spring, a 23-year-old former OpenAI researcher named Leopold Aschenbrenner revealed his Situational Awareness Fund had grown from $225M to $5.5 billion in under two years, almost entirely by buying the unglamorous infrastructure underneath the AI boom.
Read on their own, each is a normal "AI is big" story. Read together — and read against Polymarket pricing Anthropic at 78–90% across nearly every category leadership market — they are the same story told from three angles: inference compute is being repriced as both the binding bottleneck of the agent era and a new investable asset class, in the same week. The capital stack is rewiring itself in real time, and a lot of public-equity investors are still pricing AI as a software story.
This piece pulls all three together.
1. The Cerebras print: what an inference-first IPO looks like
The Cerebras numbers are the first thing to anchor on. Per the S-1 and the post-IPO coverage:
- $60B valuation at pricing; revenue of $510M in 2025 (up 76% YoY).
- Hardware $358M, cloud services $152M — a meaningful shift toward selling tokens-per-second rather than just dinner-plate-sized chips.
- A $20B+ multi-year contract with OpenAI to deliver 750MW of low-latency inference compute through 2028, with an option to expand to 2GW through 2030.
- G42 and MBZUAI together drove a "large majority" of 2025 revenue. The OpenAI deal is the engine that re-rates 2026 and beyond.
The CFO comments in Latent Space's coverage are the tell. Asked about model size, the company said it currently serves trillion-parameter models — explicitly naming "OpenAI 5.4 and 5.5" — and that there is "no limit" to the model size it can serve. The pitch is no longer "we have a fast chip." It is "we are the production inference layer for frontier models that GPUs cannot serve at the latency users now demand."
The community context is worth flagging too. The same Hacker News audience that initially treated Cerebras as a curiosity has flipped completely. The thread on the original IPO filing news is now a useful time capsule of how the consensus changed.
The market response, per The Motley Fool, made it the biggest IPO of 2026 so far. Stock soared 68% on day one. The conventional read is "AI bubble froth." We think the better read is that retail and institutional capital have finally noticed that the binding constraint on every frontier-model product — ChatGPT Advanced Voice, Claude Code, the agent runtimes everyone is now shipping — is inference latency at production scale, not training FLOPs at the next milestone.
The HN discussion when Cerebras's investor list — Altman and Ilya among them — became public makes the point even more cleanly: this is not a niche bet anymore.
If you've been following our coverage of Anthropic's six-surface distribution push and the AWS $100B Claude dominance clock, this is the same story from the supply side: the same demand that makes Anthropic look like a category monopolist makes Cerebras look like the only US-listed pure-play on the supply.
2. The Anthropic-SpaceX deal: a hyperscaler is just a power-and-real-estate company
A week before the Cerebras print, Anthropic did something even stranger. It signed a deal with SpaceX — yes, the rocket company — to take over the entire compute capacity of xAI's Colossus 1 data center in Memphis. That is over 220,000 NVIDIA GPUs and more than 300 megawatts of power, per Bloomberg and Tom's Hardware. xAI built it; Anthropic rents it; both companies and SpaceX are exploring "multiple gigawatts of orbital AI compute capacity" together.
Two things to notice.
First, the demand context. Anthropic CEO Dario Amodei said Q1 2026 revenue and usage grew 80x against an internal plan of 10x. The New Stack frames the deal as "Anthropic recruited SpaceX's 220,000-GPU Colossus 1 to fix what Claude users kept complaining about" — the rate-limit complaints that filled r/ClaudeAI for most of April. Within hours of the deal, Claude Code's five-hour rate limits doubled for paid tiers, peak-hours throttling was removed for Pro and Max, and API rate limits for Opus models were "considerably" raised. The deal is, in operational terms, a 300MW patch to a customer-experience bug.
Second, Elon. Musk has spent two years calling Anthropic "woke," "misanthropic," and "evil". Then he handed them the keys to Colossus 1. His public quote: "Everyone I met was highly competent and cared a great deal about doing the right thing. No one set off my evil detector." The reason is not friendship. SpaceX has been the de-facto AI infrastructure financier for xAI for two years — pouring rocket revenue into GPUs — and the math now wants those GPUs leased, not held. Rocket cash flows fund the chips; Anthropic's token revenue services the chips; everyone takes a cut on the way through.
Semafor put it most cleanly: the Anthropic-SpaceX deal "shows how tokens are taking over the economy." A rocket company is now a hyperscaler because the unit economics of tokens-per-watt are now competitive with the unit economics of low-Earth-orbit launches. That is what an inflection looks like.
The Hacker News thread on the deal — which surfaced the same day the formal xAI announcement landed — surfaced two things worth highlighting. The technical analysis is that this is not a one-off rental; xAI's roadmap was to deprecate Colossus 1 in favor of the larger Colossus 2 cluster, so renting it to a competitor is more efficient than mothballing it. The cultural analysis is that the supposedly fragmented frontier-model market is, at the infra layer, a single shared pool. There is no "team Anthropic" and "team xAI" hardware stack. There is one pile of GPUs and a yield curve.
3. Leopold's $5.5B fund: the AGI thesis as a public-equity portfolio
The third leg is the one most people in tech are sleeping on. Leopold Aschenbrenner — the 23-year-old former OpenAI Superalignment researcher who wrote the Situational Awareness essay that has become the canonical AGI-investor primer — turned that thesis into a hedge fund called Situational Awareness LP. Per the February 2026 13F filing covered by Fortune, the fund went from ~$225M at launch in 2024 to $5.5 billion in U.S. equity exposure by Q1 2026.
What is in the book? Per The Motley Fool's breakdown of the top 7 holdings and Fortune's profile:
- Power companies and independent power producers.
- Bitcoin miners (cheap, transferable kilowatts).
- Chip-design companies and fab equipment makers (not just the headline names).
- Adjacent enablers — utility-scale storage, transmission, specialized real-estate.
What is not in the book? The headline AI names. No NVIDIA. No Broadcom. No Microsoft or Alphabet at material weights. The thesis is that those names are already priced for AGI, and the unpriced trade is one layer down — the megawatts and wafers that feed them.
Peter Diamandis spent EP #255 of Moonshots walking through the same thesis: the Anthropic compute shortage, SpaceX as a hyperscaler, Google's orbital data center patents, and Leopold's fund as a single connected story. The episode's most quoted line: "the singularity may become visible in space before it does on Earth." Whether or not you believe that, the capital flow implication is hard to argue with. The smart-money infrastructure trade is no longer in the SaaS names you already know.
4. Why the market still prices Anthropic at 78–90%
Here is the part the macro coverage usually misses. If inference compute is supply-constrained and Anthropic just publicly admitted to an 80x demand surprise, the textbook read is "the leader gets capped, the followers catch up." That is not what's happening on prediction markets.
Polymarket is pricing Anthropic across roughly every "best AI model" market this week at 78–90%. The May 16 markets show:
- "Best AI model overall" — Anthropic ~82%.
- "Best AI model, end of June 2026" — Anthropic ~69%.
- "Best coding model" — Anthropic ~90%.
- "Best AI model on May 16" —
claude-opus-4-6-thinkingat 99%.
These numbers are higher, not lower, than they were a month ago — after the compute-shortage story broke. The implied market view is not "Anthropic gets supply-constrained." It is "Anthropic will close the supply gap (via deals like SpaceX, AWS, Google, and presumably more), and once it does, demand will keep compounding from a leadership position."
That is consistent with what Cerebras's order book is saying and consistent with what Leopold's fund is buying. The market does not believe the bottleneck is permanent; it believes the bottleneck is priced into the wrong layer. Capital is racing to fund the layer that unlocks the supply.
If you want our full take on Anthropic's pricing-power story, the $1 trillion valuation monopoly framing piece lays out the demand side. This week's three stories are the supply side of the same thesis.
The HN thread when Cerebras filed to come back — after the previous withdrawn S-1 — caught the moment the market started taking the supply story seriously again:
5. The "follow the money" picture
Stand back and the capital stack from this one week looks like this:
| Layer | Story this week | What it tells you |
|---|---|---|
| Tokens | Anthropic 80x demand surprise; Claude rate limits doubled overnight | Demand outran every plan |
| GPUs | 220,000 NVIDIA GPUs at Colossus 1 transferred from xAI to Anthropic | Physical pool, not team pool |
| Wafers | Cerebras $60B IPO; 750MW OpenAI deal; supply gated by TSMC through 2028 | Inference-first chips win an asset class |
| Power | "300 MW" headlined in every story; Leopold's fund overweights IPPs and BTC miners | Megawatts are the real bottleneck |
| Capital | Situational Awareness LP +$5.3B in 18 months on this exact thesis | Public equity is catching up to the physical layer |
| Orbit | Anthropic + SpaceX scoping "multiple gigawatts" of orbital compute | The exotic optionality nobody is priced for |
Almost every one of these layers used to be priced as a feature of "AI software." This week, each one became its own market. That is what a supply-side inflection looks like.
6. What this means if you build with AI
A few operational takeaways for builders.
Latency, not capability, is now the customer-facing variable. Cerebras's pitch — "we serve trillion-parameter models at speeds GPUs can't match" — only makes sense in a world where users notice the difference. If your product depends on real-time agent loops (voice, code completion, coding agents, browser-using agents), the binding constraint on your UX in 2026 is what fraction of inference the underlying lab routes to specialized wafer-scale silicon vs. shared GPU pools. That is now a procurement decision your model provider is making for you. Ask them.
Rate-limit policy is supply-driven, and supply is now political. When Anthropic doubled rate limits the same week as the SpaceX deal, that wasn't a strategy decision — it was a capacity decision. As more inference moves to deals like Cerebras-OpenAI and SpaceX-Anthropic, expect the rate-limit relief curve to track those announcements directly. If you can read a press release, you can predict your API ceiling six months out.
The "circular deal" critique has run its course. The reflex skepticism — "OpenAI invests in NVIDIA which invests in CoreWeave which sells to OpenAI" — assumes the money is making round trips through a fixed pool. That was a reasonable read a year ago. With Cerebras going public, with SpaceX renting Colossus to Anthropic, and with Leopold's fund flowing into power and miners, the pool is being widened by genuinely outside capital. See our Google-Anthropic $40B circular deal piece for the prior frame; this week's stories meaningfully break it.
Watch the orbital line item. It sounds like science fiction. So did "rocket company becomes hyperscaler" before this month. Per the CNBC writeup, Anthropic and SpaceX explicitly committed to scoping "multiple gigawatts" of orbital compute. The cost of getting megawatts to low Earth orbit, divided by the cost of getting megawatts to Memphis, has been closing for two years. If it closes by 2028, the entire physical-layer thesis re-rates again — and Leopold's fund is one of the few public vehicles structured to benefit.
The bottom line
Cerebras's IPO, the Anthropic-SpaceX deal, and Leopold's fund are not three AI stories. They are one story about a market that has finally figured out that inference is the asset. Not the model weights, not the chat interface, not even the chips on their own — the entire stack of wafers, power, latency, and rent that turns weights into tokens at the speed users have now learned to demand.
The Polymarket pricing — Anthropic at 78–90% despite an admitted compute shortage — is the cleanest signal that capital is no longer treating the supply problem as a ceiling on the leader. It is treating it as an investable bottleneck. That is what an inflection looks like.
We're going to watch two things over the next six weeks. First, whether Cerebras's print pulls more inference-specialist silicon into public markets — Groq, SambaNova, and the AI-ASIC arms at Broadcom and Marvell are obvious candidates. Second, whether the orbital-compute line in the Anthropic-SpaceX deal turns into an actual capex commitment. If both happen, the $1T Anthropic monopoly thesis and the Leopold thesis end up describing the same trade from opposite ends.
For builders, the practical move is to start treating model-provider supply policy as a first-class input to your roadmap — the same way you already treat cloud-provider region availability and GPU prices. Inference is the inflection. The capital is just catching up.
About ComputeLeap Team
The ComputeLeap editorial team covers AI tools, agents, and products — helping readers discover and use artificial intelligence to work smarter.
💬 Join the Discussion
Have thoughts on this article? Discuss it on your favorite platform:
Related Articles
How to Use AI for Stock Market Research (Beginner's Guide)
A beginner's guide to using AI for stock market research. Learn about sentiment analysis tools, AI screeners, and portfolio analyzers — with honest expectations about what AI can and can't do.
AI vs Human Financial Advisors: The Complete Comparison
Should you trust an AI with your finances or stick with a human advisor? We break down the costs, capabilities, and best use cases for each.
The ComputeLeap Weekly
Get a weekly digest of the best AI infra writing — Claude Code, agent frameworks, deployment patterns. No fluff.
Weekly. Unsubscribe anytime.