I’m assuming most readers are at least somewhat aware of DeepSeek over the past week, which culminated in a large sell-off in AI stocks on Monday.
I thought it worth putting down some thoughts about where this may go.
Quick background
A small Chinese company, DeepSeek, released a large language model that matches or betters many of the leading Western models. Reported costs are around 10% (i.e. 90% lower) vs similar models. Some are suggesting the end of the AI boom.
Interesting features
DeepSeek is not from one of the big Chinese IT companies, such as Baidu, Tencent, or Alibaba, which are spending a lot on AI.
DeepSeek was trained on NVIDIA chips, but far fewer than (say) OpenAI is using. DeepSeek acknowledged that it is limited by chip availability and is actively seeking more computing power.
The 90% lower figure might be overstated, maybe because of black market NVIDIA chips, or maybe hyperbole. Competitors are suggesting 30-50%. Which is still great! But 2-3x better rather than 10x better.
The resulting model, which answers the questions, is small enough to run on many home PCs.
Most of the benefits seem to be from a change of learning methodology. But not a revolutionary change – it has effectively taken and improved the OpenAI model. OpenAI basically did the same thing to Google. All the models are running the same (broad) underlying calculation.
It is open source. So, everyone else is scrambling to copy the insights into their (mostly) private models.
Given that it is open source, it is not that difficult to turn off the Chinese censorship of results.
What does this mean for AI spending?
Advertisement
It is hard to see how this is going to decrease AI spending. No CEOs are waking up today and deciding to apply the improvements and then use that to cut costs in their AI department. But spending will be more productive.
DeepSeek has been running largely for free. It probably can’t do that at scale unless the Chinese goverment wants to fund AI hardware. Presumably in order capture the data. Given the US was looking to ban TikTok over user data, it would seem likely that Chinese AI might face similar issues.
Net effect: more productive spending is likely to mean more spending.
Advertisement
Trillion-dollar question for NVIDIA
Will you need the highest-end chips?
Most of the appeal for NVIDIA is that:
a) it has a big lead at the top end of the market
b) in a world of constrained computing power, the top end was so important
c) thus, it seemed NVIDIA’s economic moat would last for a while.
DeepSeek clearly shows that you don’t need the biggest budget to achieve a breakthrough.
Advertisement
The question now is whether data centres will install lower-end chips for much less or stick with the higher-end chips.
My initial guess? Currently, the world can’t build data centres fast enough. Both power and (experienced) labour are the bottlenecks. Given these constraints, it is unlikely that someone would choose to build (say) three data centres with less powerful chips rather than two data centres with the most powerful chips.
Importantly, the new data centres have a different design. It is about networking vast arrays of GPUs. There is nothing in DeepSeek suggesting we go back to old data centre design.
Switching costs are low
Advertisement
You are only as good as your latest model. Many of the service models being built now allow you to change the AI engine quickly and easily. That is going to keep profits low for many of the AI providers (Google, OpenAI, Anthropic).
It is all about services.
Training / Inference mix
First, remember that we are taking really small trends and trying to extend them out years.