DeepSeek has significantly reduced the prices of its artificial intelligence (AI) models, including its latest V4 offering, which now costs 97% less than comparable OpenAI products. This strategic move aims to attract more enterprise clients, developers, and agent-based users.

The company announced permanent price cuts for "input cache hits" for application programming interface (API) users, reducing costs to one-tenth of original levels. The minimum input cost is now approximately USD 0.14 per million tokens, effective immediately.
To promote its new V4-Pro model, DeepSeek also offered an additional 75% discount through May 5. This makes DeepSeek-V4-Pro as inexpensive as USD 0.0036 per million input tokens, significantly lower than its American rivals.
In comparison, OpenAI’s GPT-5.5 charges USD 0.5 per million cached input tokens. A conversation on GPT-5.5 can be 32 times more expensive than on DeepSeek-V4, considering input typically exceeds output in length.
DeepSeek's aggressive pricing reflects heightened competition within China's foundational model market. Other high-profile start-ups, such as Kimi K2.6 and Zhipu GLM-5.1, recently increased prices on their latest flagship versions.
DeepSeek V4 is one of the few models to break this trend, prompting speculation of further price competition. Following DeepSeek-V4's release, OpenRouter, a US-based model aggregation platform, reported a notable surge in usage.
DeepSeek-V4-Pro recorded 13.6 billion tokens recently, nearly four times its previous day’s volume on OpenRouter. Hu Yanping, a distinguished professor at Shanghai University of Finance and Economics, noted DeepSeek's goal of attracting more users and lowering industry pricing expectations.
Professor Hu also suggested that DeepSeek's price cuts might have a relatively limited impact on top-tier models like GPT-5.5 and Claude 4.7 Opus. DeepSeek stated its V4 models are optimised for mainstream agent tools.
Users must configure Claude Code by setting the model to "deepseek-V4-pro" for 1-million-token context support, while OpenCode and OpenClaw require an upgrade. This pricing overhaul accompanies technical upgrades and deeper integration with Huawei Technologies’ Ascend ecosystem.
This integration is expected to further enhance cost efficiency. DeepSeek-V4 is viewed as enabling more complex agent-based applications at a lower cost, expanding the scope for scalable AI deployment.
A report from Goldman Sachs highlighted that integration with Ascend supernodes could strengthen DeepSeek's cost competitiveness and support wider adoption. Arena.ai evaluated DeepSeek-V4-Pro's performance as comparable to GPT-5.4-high and Gemini-3.1-Pro in agent-based web development tasks.
Artificial Analysis found DeepSeek-V4 to be far more cost-efficient than leading closed-source models such as Claude, despite performance lags. On the third-party benchmark firm’s flagship Intelligence Index tests, DeepSeek’s Flash variant cost approximately USD 113 in inference terms.
In contrast, Anthropic’s latest Claude Opus 4.7 model incurred costs around USD 4,811 for the same tests.
DeepSeek has reduced prices for its V4 AI models by 97% compared to OpenAI's GPT-5.5.
The V4-Pro model is now available for as low as USD 0.0036 per million input tokens, with a 75% promotional discount through May 5.
DeepSeek's pricing strategy aims to attract enterprise clients, developers, and agent-based users amidst increasing competition in China's AI market.
Source: SCMP

