$5.5 million. That's what DeepSeek spent to train a model that outperforms GPT-4 on several benchmarks. OpenAI reportedly invested around $100 million to train GPT-4, approximately 18 times more. When DeepSeek released its app in January 2025, it became the most-downloaded app in the world within hours. NVIDIA's stock lost around $600 billion in market capitalization in a single day.
Think AI is an American game? Think again.
This article covers all current numbers, data, and facts about DeepSeek: from user counts and downloads to training costs and benchmarks, API pricing, and company metrics.
- DeepSeek shook the AI market in January 2025: 173M downloads, 96.9M monthly active users, and the #1 app worldwide in the App Store
- DeepSeek-V3 was trained for only $5.5M, roughly 1/18 of estimated GPT-4 training costs, yet achieves comparable benchmark results
- Only 150-200 employees, no external VC funding until 2026: DeepSeek is the most capital-efficient frontier AI lab in the world
1. What is DeepSeek?
DeepSeek is a Chinese AI research company founded in July 2023 in Hangzhou. Behind it is Liang Wenfeng, who previously co-founded the quantitative hedge fund High-Flyer Capital. High-Flyer invested heavily in building a large GPU cluster before the US introduced export restrictions on high-performance chips to China.
The stated goal:
Frontier AI research with a fraction of the resources used by Western labs. DeepSeek publishes its models as open source on HuggingFace and releases technical reports transparently, a practice that is rare among commercial AI providers.
What's special about the team: Instead of experienced industry AI researchers, DeepSeek deliberately hires fresh graduates and doctoral students. Only around 20-25% of employees have more than three years of professional experience.
2. Downloads and User Numbers
January 2025 was the month DeepSeek turned the AI world upside down. The numbers speak for themselves:
- 173M downloads total (by May 2025)
- 96.9M Monthly Active Users (April 2025)
- 22.15M Daily Active Users at peak (January 2025)
- 57.2M app downloads (34.6M Google Play + 22.6M App Store by May 2025)
For comparison: ChatGPT took two months after its November 2022 launch to reach 100 million users. That was a record at the time. DeepSeek came close to that in just three months, without the global marketing budget of a Silicon Valley giant.
2.1 Geographic Distribution
DeepSeek is globally distributed, with a clear home base:
The relatively low US share of 5% is notable. It's partly due to political concerns and the fact that American companies and agencies often avoid DeepSeek for security reasons.
3. Models and Technical Specifications
DeepSeek has released several powerful models in a short time. Key point: all major models are open source and can be run on your own hardware.
3.1 DeepSeek-V3
DeepSeek-V3 is the flagship model for general tasks. The technical specs are impressive:
- 671 billion parameters total, with 37 billion activated per token (Mixture-of-Experts architecture)
- Training cost: $5.5M (2.788M H800 GPU hours)
- Available as API and open source on HuggingFace
Benchmark | DeepSeek-V3 | GPT-4o | Claude 3.5 Sonnet |
|---|---|---|---|
| MMLU | 88.5% | 87.2% | 88.3% |
| HumanEval | 65.2% | 90.2% | 92.0% |
| MATH | 61.6% | 74.6% | 71.1% |
| BBH | 87.5% | 83.1% | 88.0% |
| GPQA | 59.1% | 53.6% | 65.0% |
The fact is: DeepSeek-V3 is on par with GPT-4o on the MMLU benchmark. On some coding benchmarks, Western models lead; on general knowledge tests, results are nearly identical. All of this for $5.5 million in training costs.
3.2 DeepSeek-R1
DeepSeek-R1 is the reasoning model, comparable to OpenAI's o1. It's optimized for logical reasoning, mathematics, and complex coding tasks:
- MMLU: 90.8% (higher than V3 at 88.5%)
- AIME 2025: 87.5% (mathematical olympiad problems); AIME 2024 per original paper: 79.8%
- Open source, weights available on HuggingFace
3.3 DeepSeek-V3.2 and V4 (2026)
The latest models push even further:
- V3.2-Exp: $0.028 per million input tokens, AIME 2025: 96.0% (surpasses GPT-5 High at 94.6%)
- V4: $0.30/$0.50 per million tokens (input/output)
4. API Pricing
The lowest price is DeepSeek's strongest argument in the developer community. Here's a direct comparison with the leading competitors:
Model | Input ($/1M Tokens) | Output ($/1M Tokens) |
|---|---|---|
| DeepSeek V3.2-Exp | $0.028 | $0.028 |
| DeepSeek V3.2 | $0.26 | $0.38 |
| DeepSeek R1 | $0.55 | $2.19 |
| DeepSeek V4 | $0.30 | $0.50 |
| GPT-4o | $2.50 | $10.00 |
| Claude 3.5 Sonnet | $3.00 | $15.00 |
| Gemini 2.0 Flash | $0.075 | $0.30 |
DeepSeek V3.2-Exp costs only 1.1% of what Claude 3.5 Sonnet charges. For many use cases, quality is comparable. This has put significant pressure on the entire industry's API pricing.
5. Benchmark Comparison with Competitors
Benchmarks aren't a perfect measure. However, they provide a good overview of a model's relative performance on standardized tasks.
6. Revenue and Company Data
DeepSeek is an unusual company: small, efficient, yet on par with billion-dollar labs.
- Annualized revenue: not officially confirmed (external estimates: $100-300M, 2025)
- API calls per month: 5.7B (2025)
- Funding: Internally backed by High-Flyer Capital, no external VC round until April 2026
- First external round (Apr. 2026): Targeting $10B+ valuation
- Employees: only 150-200 (vs. OpenAI's 3,500+)
Revenue per employee (estimated, based on unconfirmed revenue figures): approximately $1.1-1.5M. This is very high for an AI startup, clear evidence that DeepSeek's approach of working with a small, highly qualified team works not just scientifically, but economically.
7. The DeepSeek Shock: Market Impact
January 27, 2025 will go down in technology market history. That day, NVIDIA lost approximately $600 billion in market capitalization in a single trading session, a record in stock market history.
The trigger was the announcement that DeepSeek-V3 achieves results comparable to Western labs at a fraction of their resource costs. Investors asked: if AI training becomes this cheap, do we still need thousands of NVIDIA GPUs?
- NVIDIA stock decline: approximately -17% in one day (January 27, 2025)
- NVIDIA market cap loss: approximately -$600B
- Political response: Intensification of US chip export controls for China
- Regulatory response: Privacy authorities in Italy, India, and Australia investigated DeepSeek
Sam Altman, CEO of OpenAI, commented on X: "impressive model, particularly around what they're able to deliver for the price." From a direct competitor, that's almost a compliment.
8. Founders and History
Liang Wenfeng is not a typical AI founder. He studied mathematics and engineering before entering quantitative trading. With his co-founder Xu Jin, he founded High-Flyer Capital, one of China's most successful quantitative hedge funds.
High-Flyer recognized early that AI would be the next major lever in financial markets. The company began investing heavily in NVIDIA GPUs before the US introduced export restrictions. DeepSeek emerged from this GPU infrastructure as an internal research project, spun off as an independent company in 2023.
Date | Milestone |
|---|---|
| Jul. 2023 | DeepSeek founded |
| Nov. 2023 | DeepSeek Coder V1 (coding specialist) |
| Jan. 2024 | DeepSeek LLM 67B (general model) |
| May 2024 | DeepSeek-V2 (MoE architecture, 236B parameters) |
| Dec. 2024 | DeepSeek-V3 (671B, training cost: $5.5M) |
| Jan. 2025 | App launch + #1 globally in App Store |
| Jan. 2025 | NVIDIA stock drops -17% in one day |
| Feb. 2025 | DeepSeek-R1 (reasoning model, open source) |
| Apr. 2025 | 96.9M Monthly Active Users |
9. Availability and Open Source
DeepSeek is accessible through multiple channels:
- App: iOS and Android (free)
- Web chat: chat.deepseek.com
- API: platform.deepseek.com
- Open source (HuggingFace): V3 and R1 as open-source models
The open-source release of model weights has sparked a wide range of community projects. Developers run DeepSeek models locally via tools like Ollama. With local deployment, privacy concerns disappear entirely.
10. Interesting Facts and Records
- #1 app worldwide in the App Store in January 2025, just days after launch.
- Largest single-day market cap loss triggered by an external announcement: NVIDIA lost approximately $600B through DeepSeek's announcement.
- Cheapest frontier model in the world: DeepSeek V3.2-Exp at $0.028 per million input tokens.
- Complete V3 training for $5.5M: Less than the budget of an average Hollywood film production.
- 150-200 employees competing with Western AI labs each employing thousands.
- Open source despite commercial success: DeepSeek releases frontier models freely, even though they would be extremely valuable commercially.
- Team of recent graduates: Most developers are fresh university graduates without years of industry experience.
- "Sputnik moment" of AI: That's how many analysts describe the DeepSeek moment, it made the Western world reconsider its assumptions about AI dominance.
For more comparisons, see our articles on Claude Statistics and ChatGPT Statistics.






