Cerebras IPO Targets $26.6 Billion — Backed by the Same Company That Funds Its Biggest Competitor
- Aisha Washington
- 5 days ago
- 7 min read
Cerebras Systems, the AI chipmaker positioning itself as the fastest inference alternative to NVIDIA, announced its IPO on May 4 at a target valuation of $26.6 billion. Its most important backer — the company that loaned Cerebras $1 billion, signed a $20 billion computing contract with it, and whose CEO is a personal angel investor — is OpenAI, the same company that accepted $30 billion from NVIDIA in its $122 billion funding round six weeks earlier. The Cerebras IPO is the most structurally peculiar public offering in recent AI history: a company going public on the strength of a relationship that could also be its undoing.
The basic facts are straightforward. Cerebras plans to sell 28 million shares at $115 to $125 each, raising up to $3.5 billion. Banks are already fielding roughly $10 billion in orders for those shares — nearly three times the target — suggesting the price will close above the stated range. The stock will trade on the Nasdaq under the ticker CBRS, with pricing expected around May 13. What the numbers don't show is the web of dependencies underneath.
What Happened
Cerebras has been building toward this Cerebras IPO for years, but the shape of it changed dramatically over the past six months. In December 2025, OpenAI loaned Cerebras $1 billion in working capital. In exchange, OpenAI received warrants for 33.5 million Cerebras shares at an exercise price described in the filing as "a fraction of a penny." Shortly after, in early 2026, Cerebras signed a multi-year agreement worth more than $20 billion with OpenAI to deliver 750 megawatts of computing capacity through 2028.
Then in February, Cerebras closed a venture round at a $23 billion valuation, with AMD among its investors — a notable detail, since AMD is simultaneously one of NVIDIA's main GPU competitors and now a Cerebras backer. By the time the IPO launched in May, Sam Altman, Greg Brockman, Ilya Sutskever, and Adam D'Angelo were all listed as angel investors.
The angel investor list reads like an OpenAI alumni directory. That is not a coincidence. OpenAI needs what Cerebras offers: an inference chip fast enough to power real-time applications at costs that don't require NVIDIA to approve every contract. Building Cerebras into a viable supplier — by lending it money, anchoring its revenue, and backing its founders personally — is OpenAI's hedge against the GPU supply chain it depends on.
The filing acknowledges the risk directly. If Cerebras fails to deliver computing capacity on the agreed timelines, OpenAI can terminate part or all of the contract. The $1 billion loan could become repayable under certain circumstances. The IPO is, in part, a way for Cerebras to reduce that dependency before it becomes a liability in the public markets.
Why the Cerebras IPO Matters Beyond the Numbers
The chip market that Cerebras is entering in 2026 looks different from the one where NVIDIA became dominant. Training large models — which dominated AI infrastructure spending between 2020 and 2025 — favors massive GPU clusters, parallelism, and NVIDIA's CUDA software ecosystem. Inference, which is what happens every time someone uses ChatGPT, Gemini, or a code assistant, favors something different: low latency, high throughput, and single-chip efficiency.
Cerebras' Wafer-Scale Engine 3 (WSE-3) was designed specifically for inference. The chip is 46,225 square millimeters — compared to 814 square millimeters for NVIDIA's H100 — and contains 900,000 cores with memory bandwidth seven thousand times greater than the H100. In practical terms, Cerebras claims the WSE-3 delivers 450 tokens per second running Llama 3.1 70B, roughly 20 times faster than NVIDIA GPU-based cloud solutions running the same model.
The use case where this matters most is not consumer chatbots. It is latency-sensitive enterprise applications: financial risk models that must respond in under 100 milliseconds, real-time API calls in agentic workflows, medical diagnostics requiring immediate output. You cannot achieve sub-100ms inference by stacking more H100s together. The architecture does not work that way. Cerebras' single-chip approach is, for these specific workloads, not a cheaper alternative — it is the only viable option at that performance level.
This is why OpenAI signed a $20 billion contract rather than simply licensing NVIDIA capacity for inference. OpenAI's own product roadmap depends on inference speed. If GPT-5.5 Instant delivers 52 percent fewer hallucinations and responds faster than GPT-5.3, part of that improvement comes from where and how inference is run.
The cerebras ipo is, at its core, a bet that inference speed becomes a competitive moat rather than a commodity.
The Part Nobody Is Talking About
The $10 billion in IPO orders and the 3× oversubscription are the numbers getting the most coverage. The number worth examining is the exercise price on OpenAI's 33.5 million warrants: a fraction of a penny per share.
At Cerebras' IPO price of $125 per share, those warrants are worth approximately $4.2 billion. OpenAI will have lent $1 billion and received the right to purchase stock worth $4.2 billion for essentially nothing. That is not unusual in venture lending — warrants are the lender's upside — but it becomes structurally significant at IPO scale. Once Cerebras is public and the lockup expires, OpenAI could become one of its largest shareholders without having paid for the position.
The customer concentration problem compounds this. According to the IPO prospectus, Mohamed bin Zayed University of Artificial Intelligence (MBZUAI), a UAE state-backed institution, accounted for 62 percent of Cerebras' 2025 revenue. G42, the Abu Dhabi technology group, contributed another 24 percent. Together, two UAE-linked entities generated 86 percent of total sales.
This is not a new risk. Cerebras filed for its first IPO in 2024, but that attempt was delayed after national security reviewers scrutinized G42's ties to Chinese technology companies. The clearance eventually came in 2025. The second filing does not reduce the UAE exposure — it shifts the concentration from G42 (down from 85 percent of 2024 revenue to 24 percent) to MBZUAI, which was not the subject of the original review.
Independent analyst William Keating noted the core paradox: Cerebras has solved its G42 problem by becoming more dependent on a different UAE customer, while simultaneously taking on OpenAI as a customer whose contract contains termination provisions that could unwind the entire business model. "The real concentration risk is not NVIDIA," he wrote. "It is that Cerebras has two customers who can each inflict existential damage."
What the $100 billion in orders reflects is institutional appetite for AI chip supply chain diversification. Investors want NVIDIA alternatives to exist. Whether Cerebras is the right vehicle is a separate question from whether the demand for that vehicle is real.
Cerebras vs NVIDIA: The Actual Comparison
Cerebras and NVIDIA are not competing for the same customers in the same way AMD and NVIDIA do. The comparison that matters is narrower.
NVIDIA's H100 and Blackwell families win on training workloads, software ecosystem depth, and multi-modal flexibility. CUDA has sixteen years of tooling, libraries, and developer familiarity behind it. For any organization building or fine-tuning large models, NVIDIA remains the default and will likely remain so through the next hardware generation.
Cerebras wins on inference latency for specific architectures. The WSE-3's single-chip design eliminates the inter-chip communication overhead that slows GPU clusters at low-batch inference. For a financial institution running real-time compliance checks, or an agentic workflow requiring synchronous model calls, the performance gap is meaningful and cannot be closed by adding more H100s.
NVIDIA's Blackwell architecture is expected to narrow the inference speed difference on standard benchmarks. It will not replicate the single-chip latency profile. The cerebras stock thesis is essentially a bet that a class of latency-sensitive enterprise applications becomes large enough to sustain a $26 billion company. Groq, the nearest comparable inference chip startup, is valued at roughly $3 billion. The gap between Groq's valuation and Cerebras' reflects, in large part, the OpenAI relationship.
AMD's presence as both a Cerebras investor and a NVIDIA competitor adds a layer of strategic complexity. AMD benefits from Cerebras validating alternative AI chip architectures, but its MI300X is a direct inference competitor in the mid-latency tier where Cerebras also plays.
What Happens After the Cerebras IPO
The immediate milestone is the May 13 pricing. If the book closes above $125 — which the $10 billion order flow suggests is likely — it will set the tone for how the market values AI infrastructure pure-plays that are not NVIDIA.
The 90-day lockup expiration is the next structural event. OpenAI's warrants are tied to equipment delivery milestones rather than the standard lockup, which means the timeline for OpenAI becoming a major Cerebras stockholder depends on how quickly Cerebras can fulfill its compute commitments under the $20 billion contract. Every data center rack Cerebras deploys for OpenAI brings the warrant vesting trigger closer.
The strategic variable that most affects the Cerebras IPO thesis long-term is customer diversification. If 24 months after IPO, MBZUAI and G42 still account for more than 70 percent of revenue, the valuation will have difficulty sustaining itself regardless of how many OpenAI racks are deployed. The AWS partnership — Cerebras chips deployed in Amazon cloud data centers — is the most visible path to non-UAE, non-OpenAI revenue, but it carries no public commitment on volume.
The question Cerebras has not yet answered publicly is what happens if OpenAI decides the inference market is better served by its own chip design. Several of the largest AI companies — Google (TPU), Amazon (Trainium), Microsoft (Maia) — have moved toward custom silicon for their most critical workloads. OpenAI has not announced an equivalent program, but the warrants give it both the incentive and the equity upside to accelerate one.
The Cerebras IPO opens a window into something larger than one company's public debut. AI infrastructure is fracturing. The NVIDIA monoculture that defined the first wave of the AI build-out is giving way to a more fragmented supply chain — custom chips, inference specialists, wafer-scale designs — where the relationships between companies are as important as the silicon itself.
Whether Cerebras is the right long-term bet depends less on the WSE-3's benchmark numbers than on whether OpenAI stays a customer, whether UAE revenue converts into global diversification, and whether the latency-sensitive inference market grows fast enough to justify a $26 billion entry price. The oversubscription says the market thinks it will. The prospectus says it might not.
For anyone building AI-powered workflows that depend on real-time model output, the infrastructure questions Cerebras is raising — how fast inference needs to be, and who controls the hardware that delivers it — are not abstractions. Speed at the chip level shapes what AI knowledge retrieval can do in real applications. The Cerebras IPO is a bet on where that limit is moving.