Chinese AI company DeepSeek has unveiled its long-awaited V4 model.
On Friday, the Hangzhou-based startup released its newest large language model in a preview capacity. The release comes over a year after it shook markets and reset the entire conversation around AI with its V3 and R1 models, which showed that frontier-level performance could be achieved through an open-source model trained on a relatively tiny budget.
V4 may not have as large of an effect on markets as its predecessor did. So far, the biggest impacts have been on the shares of other Chinese companies. Shares in Semiconductor Manufacturing International Corporation jumped 10% in Hong Kong trading. That Chinese chipmaker makes Huawei’s Ascend AI processors, which DeepSeek said it used to train its new model. (Huawei has been trying to catch up to the performance of Nvidia’s GPUs, which most companies use for training AI models.) Meanwhile, shares in Minimax and Knowledge Atlas, two of DeepSeek’s competitors, sank by more than 9%.
Yet the increasingly narrow performance gap between DeepSeek and leading U.S. models, as well as its rock-bottom prices, will raise questions about the competitive moat that surrounds leading U.S. labs like OpenAI and Anthropic—and the constraints that still hold back China’s AI development.
So, how good is DeepSeek’s new model?
DeepSeek released two versions on Friday: “DeepSeek-V4-Pro,” which it said had “performance rivaling the world’s top closed-source models,” and “DeepSeek-V4-Flash,” a smaller and cheaper variant. The Pro version has 1.6 trillion parameters, the tunable knobs within a neural network that serve as a measure of a model’s size, and a context length of 1 million tokens. (A token is the fundamental building block used by an LLM to understand and generate text; a million tokens is roughly equivalent to 750,000 words)
The startup claims its model beats all other open-source models in agentic coding and reasoning. DeepSeek also pointed to benchmarks that showed V4 performing favorably against Anthropic’s Claude Opus 4.6, OpenAI’s GPT-5.4, and Google’s Gemini 3.1-Pro. The startup’s tech report says that V4 “falls marginally short of GPT-5.4 and Gemin i3.1-Pro, suggesting a developmental trajectory that trails state-of-the-art frontier models by approximately 3 to 6 months.”
Just as important is how much DeepSeek is charging for access to its model. DeepSeek’s V4-Pro model will cost $3.48 for one million tokens of output; by comparison, OpenAI and Anthropic charge $30 and $25 respectively for the same amount of work. Even Kimi, from fellow Chinese AI startup Moonshot AI, costs $4 for a million tokens of output. DeepSeek’s V4-Flash costs even less, at just $0.28 for a million tokens.
That pricing puts DeepSeek at odds with a trend across the wider AI sector. Both OpenAI and Anthropic have hiked prices and imposed rate limits to manage surging demand. Chinese developers have followed suit, also increasing prices and removing subscriptions that offered unlimited usage.
DeepSeek’s prices could get even cheaper: It expects to lower V4-Pro prices later in the year as Huawei scales up production of its new Ascend 950 AI processors
Why DeepSeek mattered—and whether it still matters
DeepSeek first captured global attention in December 2024, when it released its V3 large language model. The startup claimed it trained V3 on just $5.6 million worth of processors, which AI researcher Andrej Karpathy called a “joke of a budget.” DeepSeek later released R1, a reasoning model that matched the equivalent offering from OpenAI.
That sparked a massive sell-off in U.S. tech stocks, as investors repriced how much money was needed to train and run AI models. At one point, tech stocks lost $1 trillion in value.
While markets eventually recovered, DeepSeek’s decision to release its model on an open-source basis ended up being more significant. The startup’s release built on momentum started by Alibaba’s Qwen to inspire several other Chinese labs to release their own open-source models. Even OpenAI was pushed to release its own open-source model, gpt-oss.
DeepSeek’s open-source approach allowed developers to download its model for free, tweak it for their own purposes, and run it on local hardware. That helped win DeepSeek, and later labs that also went open-source, support from resource-strapped developers, particularly those outside of the U.S. and Europe.
V3 and R1 also sparked a revival of investor interest in China’s beaten-down tech sector. In recent months, several AI startups have surged in public markets as investors try to build exposure to China’s AI ecosystem. In January, MiniMax and Knowledge Atlas listed in Hong Kong, as some of the first AI labs to tap public markets: The two startups are up 370% and 700% respectively from their offer price.
This also means that DeepSeek is releasing V4 into a new competitive landscape. Alibaba, Moonshot AI, MiniMax and Knowledge Atlas have all released high-performing open-source models this year.
Both The Information and the Financial Times report that DeepSeek is trying to raise money from Tencent and Alibaba in a funding round that would value the lab at $20 billion. DeepSeek, which is owned by High-Flyer, a Chinese hedge fund, isn’t short of cash. The reason, according to the FT, is talent: DeepSeek needs to raise money to convince its AI researchers to stick around, and not get poached by other labs with larger valuations.
The AI frontier is also about to be pushed further out. This week, OpenAI released GPT 5.5, the latest version of its flagship LLM, as well as an updated version of its image generation model. Anthropic is also testing its latest Mythos model with a select group of companies; the Claude developer claims these restrictions are necessary to ensure the model can’t be misused upon a more public release.
‘Full support’
DeepSeek worked closely with Huawei to ensure its new model runs on the Chinese tech giant’s Ascend AI processors. On Friday, Huawei announced its Ascend processors would offer “full support” for DeepSeek’s models.
Processing power is in short supply, as tech companies snap up GPUs, CPUs and memory chips in order to build more data centers and AI processors. Chinese developers have an additional problem: U.S. export controls bar them from buying the most advanced processors made by companies like Nvidia.
Ironically, these export controls may have helped Chinese AI startups, including DeepSeek, learn how to operate in a world of scarce processing power. Because Chinese developers have been forced to train and run models with limited resources, they had to come up with ways to make their models much more efficient. This power-efficiency has, in turn, allowed the Chinese developers to reduce prices substantially. U.S. pressure also forced China’s semiconductor sector to accelerate plans to manufacture chips domestically.
DeepSeek’s ability to train and run its model on Huawei chips could represent another step away from U.S. chipmakers like Nvidia and AMD.
“The best AI researchers in the world, because they are limited in compute, also come up with extremely smart algorithms,” Nvidia CEO Jensen Huang said on the Dwarkesh Podcast last week, adding that “the day that DeepSeek comes out on Huawei first, that is a horrible outcome for [the U.S.]”
U.S. officials and developers have long claimed that DeepSeek’s success is based on Nvidia processors smuggled into China, or which the Chinese company has managed to access illegally in other countries in Asia, such as Malaysia through front companies, as well as illicit distillation of leading U.S. models. Distillation is the process of training a smaller, more efficient model on the outputs of a larger one.
On Thursday, Michael Kratsios, science advisor to U.S. President Donald Trump, alleged that Chinese AI developers were engaged in “industrial-scale campaigns” to copy U.S. technology. “There is nothing innovative about systematically extracting and copying the innovations of American industry,” Kratsios wrote in a memo, and promised to work closely with the private sector to protect them against these attacks.
OpenAI and Anthropic have also accused Chinese AI developers, including DeepSeek, of conducting “illicit” distillation attacks.
On Friday, China’s foreign ministry called Kratsios’s claims “groundless” and “a smear against the achievements of China’s AI industry.”












