For years, every serious conversation about AI compute infrastructure ended in the same place: Nvidia. The H100, the B200, the GB200 — the GPU cluster has become the default unit of AI scale. That consensus is about to face its first real public market test.
Cerebras Systems, the Santa Clara-based chip startup behind the world’s largest semiconductor die, is targeting a Nasdaq IPO in April or May 2026, raising approximately $2 billion at a valuation between $22 billion and $25 billion. Morgan Stanley is leading the offering. The company cleared its last major regulatory hurdle in February when CFIUS approved the deal after Cerebras restructured Gulf-based investor G42’s stake to non-voting shares.
The $10 Billion Vote of Confidence
The most significant data point in Cerebras’s IPO story is not its chip architecture — it’s a contract. OpenAI signed a multi-year compute agreement worth $10 billion with Cerebras, making it the largest AI infrastructure deal ever signed outside of Nvidia’s own customer base.
That anchor contract changes the risk calculus for public market investors significantly. Cerebras is not asking the market to bet on a technology thesis; it is asking the market to value a business with a named hyperscaler customer, committed revenue, and demonstrated production deployment.
The deal also carries a message about OpenAI’s own supply chain strategy. Nvidia’s production constraints — particularly around the CoWoS advanced packaging used in its H100 and B200 lines — have pushed hyperscalers to diversify. Google has its TPUs, Amazon has Trainium, and Microsoft has the Maia chip program. OpenAI, previously the most Nvidia-dependent of the major AI labs, appears to be making a parallel move.
What the WSE-3 Actually Is
Cerebras’s Wafer Scale Engine 3 is genuinely different from anything else in commercial production. Where a Nvidia B200 GPU contains approximately 208 billion transistors on a roughly 800mm² die, the WSE-3 is fabricated across an entire 300mm wafer — 4 trillion transistors, 900,000 cores, and 44GB of on-chip SRAM. There is no inter-die communication overhead because there is only one die.
The performance claims are aggressive: Cerebras says the WSE-3 delivers 21 times the performance of a Nvidia DGX B200 system at one-third the cost per token and one-third the power draw. These figures are model-dependent and task-specific, and independent benchmarking remains limited. But the economics are compelling enough that one of the most sophisticated AI buyers in the world — OpenAI — signed a $10 billion contract on the basis of them.
The architecture has a structural limitation: it excels at inference and large-scale training on sparse models but is less flexible for the iterative, heterogeneous workloads that GPU clusters handle naturally. Cerebras’s go-to-market has increasingly focused on inference at scale, where its memory bandwidth and latency characteristics are most advantageous.
The IPO Landscape and What It Signals
Cerebras closed a $1 billion Series H in February 2026 led by Tiger Global, with AMD, Fidelity, Benchmark, Coatue, and Altimeter participating. The round valued the company at approximately $16 billion pre-money — the IPO target represents a 40–55% step-up in roughly two months, reflecting both the OpenAI contract announcement and improved public market conditions for AI infrastructure names.
The timing is deliberate. After a difficult 2025 for AI-adjacent IPOs — where multiple unicorns saw post-listing declines — the window has reopened in Q1 2026 as the venture funding environment recovered and AI compute demand metrics improved across hyperscaler earnings calls.
If Cerebras prices at the top of its range, the implied valuation will be a reference point for every subsequent AI chip company considering a public offering. More importantly, it will represent the first time public market investors have explicitly priced the possibility that GPU clusters are not the permanent default for AI infrastructure.
That vote, whatever the outcome, is worth watching.