About DeepSeek
Where your cheap-tier AI bill really comes from.
DeepSeek was founded on 17 July 2023 in Hangzhou by Liang Wenfeng, who also founded the High-Flyer hedge fund that backs the lab. The company released DeepSeek-V2 in May 2024 with 236 billion parameters of which 21 billion are active per token, then DeepSeek-V3 in December 2024 with 671 billion parameters and 37 billion active, and DeepSeek-R1 on 20 January 2025. R1 is the reasoning-focused model that triggered a major US tech selloff the week after its release, on the basis that a Chinese lab had trained a frontier-class reasoner for a reported six million dollars in compute against the hundred-million-dollar figure attached to GPT-4. The model weights are published under the MIT License; the training data is not.
For a warehouse the API is what matters. The line-up exposed at api.deepseek.com is deepseek-chat for the general workhorse and deepseek-reasoner for the chain-of-thought tier, both routed under the v4-flash family with v4-pro as the higher-performance variant. Pricing is roughly an order of magnitude below the comparable OpenAI or Anthropic tiers, with cache-hit input tokens charged at a small fraction of the cache-miss rate, an off-peak discount window, and a context-caching feature that matters for any workload that reuses the same system prompt. Reasoner output includes the chain-of-thought tokens in the bill, so a single hard question on deepseek-reasoner can spend more output tokens than ten chat-mode answers. The endpoints to pull are the chat-completions usage records, the model list, the balance and billing endpoints and the cache-hit accounting, so finance and product can split spend per API key, per model, per cache state and per peak window.