Chinese AI startup DeepSeek has disrupted the global AI conversation by revealing that its R1 reasoning model was trained for just $294,000. The figure, disclosed in a Nature paper, stands in sharp contrast to the hundreds of millions reportedly spent by U.S. rivals like OpenAI and Anthropic.
A Rare Look Inside DeepSeek’s Operations
DeepSeek made headlines in January 2025 with low-cost models that rattled global markets. Since then, the company has kept quiet. This new disclosure offers a rare glimpse into its methods.
The paper revealed that the R1 model was trained on 512 Nvidia H800 GPUs, chips tailored for China after U.S. export restrictions. Training lasted just 80 hours, a fraction of the usual time for large-scale AI systems.
Training Costs: DeepSeek vs. U.S. Giants
For comparison, OpenAI has admitted its foundation models cost well over $100 million to train. Against that backdrop, DeepSeek’s reported efficiency is eye-opening.
Skeptics, however, remain cautious. Some U.S. officials suspect the company accessed restricted hardware like Nvidia’s H100s. DeepSeek clarified that it owns A100 GPUs but used them only for early tests before training R1 with H800s.
Model Distillation Debate
One factor behind the low DeepSeek AI model cost may be model distillation—a method where one AI learns from another. While this boosts performance and reduces costs, critics argue it skirts intellectual property rights, especially if proprietary models are indirectly used.
DeepSeek has admitted to using Meta’s open-source Llama for earlier versions and acknowledged its V3 training data included AI-generated answers, possibly from OpenAI systems. The company maintains this was incidental.
Why It Matters
This disclosure sparks big questions for the global AI race:
- Can China produce competitive AI at a fraction of U.S. costs?
- Will concerns over transparency and IP slow global adoption?
If scalable, DeepSeek’s approach could shift the balance of power in AI, challenging U.S. dominance and accelerating China’s influence.
FAQs
1. What is DeepSeek’s R1 model?
A reasoning-focused AI model designed for advanced problem-solving.
2. How much did it cost to train?
Just $294,000, per the company’s Nature paper.
3. What hardware was used?
512 Nvidia H800 GPUs over 80 hours, with A100s in early tests.
4. Why is model distillation controversial?
It lets AI learn from other models, raising IP concerns.
5. Why does this matter globally?
It could make AI development cheaper, boosting China’s position in the tech race.
Related: Microsoft $30B UK Investment in Artificial Intelligence Growth
More from AI
How to Translate Kindle Books The Right Way, Even With DRM Restrictions
The use of digital books has turned reading into a silent activity among millions of individuals who love having shelves …
How Simple Online Writing Can Turn Into Real Monthly Income
There are moments in a person’s life when something ordinary reveals a quiet possibility. It might happen while reading a …
Canva AI Design Model: New Tools, Editable Layers, and Smarter Creative Power
The creative world has acquired a different aspect and Canva is leading in the transformation. As the Canva AI design …










