
Dear WIMM readers,
Today, I am inaugurating a new category: Watchlist. And to begin, a fitting candidate: the Cerebras IPO. Below you can also find the schedule for the rest of the week:
13 May: [Watchlist] The expected IPO of Cerebras
14 May: Portfolio update
15 May: Macroeconomics (May 2026)
16 May: [Market update] META and Palantir
17 May: [Deep dive] JPMorgan
Onto the update:
From Bloomberg:
Cerebras Systems Inc. increased the size of its initial public offering, now seeking to raise as much as $4.8 billion, as demand for the artificial intelligence chipmaker and data center operator’s shares continues to build.
The company is offering 30 million shares for $150 to $160 each, according to a filing with the US Securities and Exchange Commission on Monday. Cerebras had previously marketed 28 million shares for $115 to $125 each.
At the top of the price range, Sunnyvale, California-based Cerebras would have a market value of $34.4 billion, based on the shares outstanding in its filing. The IPO has drawn orders for more than 20 times the number of shares available, people familiar with the matter have said.
The IPO is expected to price on May 13, according to terms of the deal seen by Bloomberg News. Cerebras filed confidentially for a listing earlier this year, months after withdrawing a previous registration.
Unfortunately, we can’t buy at IPO prices in Europe. Probably, some regulatory EU bullshit. So, we haven’t seen the S-1 form yet, but I can speculate on the public information available.
Is this stock worth it at this price? Here are some points:
Cerebras is not “Nvidia, but smaller.”
It is a very different bet. Nvidia won the first AI compute era because GPUs were flexible (= good for training, good for inference, supported by CUDA, and able to scale across massive clusters). Cerebras is not trying to beat Nvidia at Nvidia’s own game. It is trying to win a narrower game, which is, very fast inference. (That’s why NVIDIA acquired Groq)The core idea is simple: make the chip huge.
Most chips are limited by how much silicon can be exposed at once during manufacturing. Cerebras breaks that convention by turning an entire wafer into one giant chip. Instead of connecting many chips together, it builds one massive piece of silicon with compute and memory sitting very close to each other.The advantage is speed, especially memory speed.
In AI inference, the bottleneck is often not only “how much compute do you have?” but “how fast can you move data?” Cerebras has a lot of fast on-chip memory, which means it can generate tokens very quickly when the model and context fit inside the chip. That makes the experience feel much more responsive.This matters because inference is becoming the main event.
The first AI story was training, for which you had to build bigger models, buy more GPUs, and connect more data centers. The next story is inference, with billions of users, apps, copilots, agents, devices, and workflows constantly asking models to produce answers. Training is the factory, but the inference is the product.Cerebras is strongest when a human is waiting.
If you are using an AI coding assistant, a voice agent, a wearable device, or a real-time business tool, latency matters. The difference between waiting two seconds and twenty seconds is not technical; it is emotional. Fast inference makes AI feel intelligent. Slow inference makes AI feel like enterprise software.But the moat has a boundary.
Cerebras works best when everything fits inside its very fast memory. If the model is too large, or the context becomes too long, or the agent needs a huge memory state, the advantage becomes less obvious. The company’s strength is speed, not infinite capacity.The big strategic distinction is answer inference versus agentic inference.
Answer inference is when a human asks a question and waits. Agentic inference is when software agents do tasks in the background: search, verify, code, test, transact, document, repeat. For answer inference, speed is king. For agentic inference, memory, cost, state, and reliability may matter more.Cerebras may own the premium front-end of AI.
Think of it as the Ferrari of inference. It may not carry all the freight. It may not run every warehouse. But when the user experience depends on speed, Cerebras has a compelling reason to exist. Voice AI, real-time assistants, coding, reasoning-heavy interfaces, and premium AI products could all benefit.The investment question is whether speed is a niche or a platform.
If the future of AI is mostly humans waiting for better answers, Cerebras becomes strategically important. If the future is mostly agents working quietly in the background, then cheaper memory-rich systems may win. Cerebras has a real technological advantage. The question is whether that advantage sits at the center of the AI economy, or at its high-end edge.
From Yahoo Finance:
Nvidia CEO heralds ‘inference inflection’ as next phase of AI boom, backed by $1 trillion in orders.
So, I will buy Cerebras and add it to my AI portfolio (ie. my 3rd portfolio), unless I will see something fishy in the IPO forms.

“Where is my moat?” is designed for individual readers, though the occasional forward is absolutely fine. If you’d like to set up multiple subscriptions for your team with a group discount (minimum 3 seats), reach out to me directly.
Thanks for your support & have a wonderful day!

