How the U.S. chip bans led to a monster called DeepSeek

The Chinese AI company DeepSeek has put the AI industry in an uproar. Denied the most powerful chips thought needed to create state-of-the-art AI models, DeepSeek pulled off some engineering master strokes that allowed the researchers to do more with less. The DeepSeek-V3 and DeepSeek-R1 models the company recently released achieved state-of-the-art performance in benchmark tests and cost much less time and money to train and operate than comparable models.

And the cherry on top: The company’s researchers showed their work—they explained the breakthroughs in research papers and open-sourced the models so others can use them to make their own models and agents.

The main reason DeepSeek had to do more with less is that the Biden administration put out a series of restrictions on chip exports saying that U.S. chipmakers such as Nvidia couldn’t ship the most powerful GPUs (graphics processing units, the go-to chip for training AIs) to countries outside the U.S.

This effort started in October 2022, and has been updated and fine-tuned several times to close loopholes. Biden released an executive order shortly before leaving office further tightening restrictions. DeepSeek apparently played by the rules. It made do with H800 chips the U.S. allowed Nvidia to sell in China, instead of the more powerful H100 that U.S. tech and AI companies use. 

With less powerful chips, the researchers were forced to find ways of training and operating AI models using less memory and computing power. 

The DeepSeek models use a “mixture of experts” approach, which allows them to activate only a subset of the model’s parameters that specialize in a certain type of query. This economizes on computing power and increases speed. DeepSeek didn’t invent this approach (OpenAI’s GPT-4 and Databricks’s DBRX model use it), but the company found new ways of using the architecture to reduce the computer processing time necessary during pretraining (the process in which the model processes huge amounts of data in order to optimize its parameters to correctly respond to user queries).

In DeepSeek-R1, a reasoning model comparable to OpenAI’s most recent o1 series of models (announced in September), DeepSeek found ways of economizing during inference time, when the model is “thinking” through various routes to a good answer. During this process of trial and error, the system must collect and store more and more information about the problem and its possible solutions in its “context window” (its memory) as it works.

As the context window adds more information, the memory and processing power required leaps up quickly. Perhaps DeepSeek’s biggest innovation is dramatically reducing the amount of memory allocated to storing all that data. In general terms, the R1 system stores the context data in a compressed form, which results in memory savings and better speed without affecting the quality of the answer the user sees. 

DeepSeek said in a research paper that its V3 model cost a mere $5.576 million to train. By comparison, OpenAI CEO Sam Altman said that the cost to train its GPT-4 model was more than $100 million.

Since the release of DeepSeek’s V3, developers have been raving about the model’s performance and utility. Consumers are now embracing a new DeepSeek chatbot (powered by the V3 and R1 models), which is now number one on the Apple ranking for free apps. (However, that success has attracted cyberattacks against DeepSeek and caused the company to temporarily limit new user registrations.) 

For the past two years, the narrative in the industry has been that creating state-of-the-art frontier models requires billions of dollars, lots of the fastest Nvidia chips, and large numbers of top researchers. Across the industry and in investment circles that assumption has been challenged. As a result, Nvidia stock fell nearly 17% Monday as investors question their assumptions about the demand for the expensive GPUs. And it’s all happening because a small shop of Chinese researchers knew they’d need some big engineering breakthroughs in order to create state-of-the-art models using less than state-of-the-art chips. 

https://www.fastcompany.com/91267968/how-the-biden-chip-bans-created-a-monster-called-deepseek?partner=rss&utm_source=rss&utm_medium=feed&utm_campaign=rss+fastcompany&utm_content=rss

Vytvořeno 5mo | 28. 1. 2025 0:40:02


Chcete-li přidat komentář, přihlaste se

Ostatní příspěvky v této skupině

Chime’s cofounder on the company’s IPO: ‘We’re just getting started’

A dozen years after its launch, fintech company Chime rang the bell this morning at the Nasdaq MarketSite in Times Square to ce

12. 6. 2025 20:20:06 | Fast company - tech
What is a fridge cigarette? The viral Diet Coke trend explained

It hits at a certain time in the afternoon, when a familiar craving strikes. You walk to the kitchen. The satisfying sound of a can cracking, the hiss of bubbles. It’s time for a “fridge cigarette

12. 6. 2025 20:20:06 | Fast company - tech
This startup wants AI to help manage software infrastructure, not just write code

Many developers find that AI programming assistants have made writing code easier than ever. But maintaining the infrastructure that actually runs that code remains a challenge, requiring engineer

12. 6. 2025 18:10:21 | Fast company - tech
Apple fumbled its personal AI debut, but the alternative was far worse

Welcome to AI DecodedFast Company’s weekly newsletter that breaks down the most important news in the world of AI. You can sign up to receive this newsletter every week 

12. 6. 2025 18:10:18 | Fast company - tech
Greenhouse and Clear team up to fight fake job applications flooding tech hiring

Fraudulent job applications have become a serious issue in the era of

12. 6. 2025 13:30:02 | Fast company - tech
‘We’re on the cusp of more widespread adoption’: Laura Shin on Trump, stablecoins, and the global rise of cryptocurrency

With the first family actively engaged in memecoin ventures, speculation about the future of cryptocurrency has never been hotter. Laura Shin, crypto expert and host of the podcast Unchained

12. 6. 2025 11:10:06 | Fast company - tech
Thanks to AI, the one-person unicorn is closer than you think

When Mike Krieger helped launch Instagram in 2010 as a cofounder, building something as simple as a photo filter took his team wee

12. 6. 2025 11:10:04 | Fast company - tech