Chinese artificial intelligence firm DeepSeek rocked markets this week with claims its new AI model outperforms OpenAI’s and cost a fraction of the price to build.
The assertions — specifically that DeepSeek’s large language model cost just $5.6 million to train — have sparked concerns over the eyewatering sums that tech giants are currently spending on computing infrastructure required to train and run advanced AI workloads.
But not everyone is convinced by DeepSeek’s claims.
CNBC asked industry experts for their views on DeepSeek, and how it actually compares to OpenAI, creator of viral chatbot ChatGPT which sparked the AI revolution.
What is DeepSeek?
Last week, DeepSeek released R1, its new reasoning model that rivals OpenAI’s o1. A reasoning model is a large language model that breaks prompts down into smaller pieces and considers multiple approaches before generating a response. It is designed to process complex problems in a similar way to humans.
DeepSeek was founded in 2023 by Liang Wenfeng, co-founder of AI-focused quantitative hedge fund High-Flyer, to focus on large language models and reaching artificial general intelligence, or AGI.
AGI as a concept loosely refers to the idea of an AI that equals or surpasses human intellect on a wide range of tasks.
Much of the technology behind R1 isn’t new. What is notable, however, is that DeepSeek is the first to deploy it in a high-performing AI model with — according to the company — considerable reductions in power requirements.
“The takeaway is that there are many possibilities to develop this industry. The high-end chip/capital intensive way is one technological approach,” said Xiaomeng Lu, director of Eurasia Group’s geo-technology practice.
“But DeepSeek proves we are still in the nascent stage of AI development and the path established by OpenAI may not be the only route to highly capable AI.”
How is it different from OpenAI?
DeepSeek has two main systems that have garnered buzz from the AI community: V3, the large language model that unpins its products, and R1, its reasoning model.
Both models are open-source, meaning their underlying code is free and publicly available for other developers to customize and redistribute.
DeepSeek’s models are much smaller than many other large language models. V3 has a total of 671 billion parameters, or variables that the model learns during training. And while OpenAI doesn’t disclose parameters, experts estimate its latest model to have at least a trillion.
In terms of performance, DeepSeek says its R1 model achieves performance comparable to OpenAI’s o1 on reasoning tasks, citing benchmarks including AIME 2024, Codeforces, GPQA Diamond, MATH-500, MMLU and SWE-bench Verified.
Read more DeepSeek coverage
In a technical report, the company said its V3 model had a training cost of only $5.6 million — a fraction of the billions of dollars that notable Western AI labs such as OpenAI and Anthropic have spent to train and run their foundational AI models. It isn’t yet clear how much DeepSeek costs to run, however.
If the training costs are accurate, though, it means the model was developed at a fraction of the cost of rival models by OpenAI, Anthropic, Google and others.
Daniel Newman, CEO of tech insight firm The Futurum Group, said these developments suggest “a massive breakthrough,” although he shed some doubt on the exact figures.
“I believe the breakthroughs of DeepSeek indicate a meaningful inflection for scaling laws and are a real necessity,” he said. “Having said that, there are still a lot of questions and uncertainties around the full picture of costs as it pertains to the development of DeepSeek.”
Meanwhile, Paul Triolio, senior VP for China and technology policy lead at advisory firm DGA Group, noted it was difficult to draw a direct comparison between DeepSeek’s model cost and that of major U.S. developers.
“The 5.6 million figure for DeepSeek V3 was just for one training run, and the company stressed that this did not represent the overall cost of R&D to develop the model,” he said. “The overall cost then was likely significantly higher, but still lower than the amount spent by major US AI companies.”
DeepSeek wasn’t immediately available for comment when contacted by CNBC.
Comparing DeepSeek, OpenAI on price
DeepSeek and OpenAI both disclose pricing for their models’ computations on their websites.
DeepSeek says R1 costs 55 cents per 1 million tokens of inputs — “tokens” referring to each individual unit of text processed by the model — and $2.19 per 1 million tokens of output.
In comparison, OpenAI’s pricing page for o1 shows the firm charges $15 per 1 million input tokens and $60 per 1 million output tokens. For GPT-4o mini, OpenAI’s smaller, low-cost language model, the firm charges 15 cents per 1 million input tokens.
Skepticism over chips
DeepSeek’s reveal of R1 has already led to heated public debate over the veracity of its claim — not least because its models were built despite export controls from the U.S. restricting the use of advanced AI chips to China.
DeepSeek claims it had its breakthrough using mature Nvidia clips, including H800 and A100 chips, which are less advanced than the chipmaker’s cutting-edge H100s, which can’t be exported to China.
However, in comments to CNBC last week, Scale AI CEO Alexandr Wang, said he believed DeepSeek used the banned chips — a claim that DeepSeek denies.
Nvidia has since come out and said that the GPUs that DeepSeek used were fully export-compliant.
The real deal or not?
Industry experts seem to broadly agree that what DeepSeek has achieved is impressive, although some have urged skepticism over some of the Chinese company’s claims.
“DeepSeek is legitimately impressive, but the level of hysteria is an indictment of so many,” U.S. entrepreneur Palmer Luckey, who founded Oculus and Anduril wrote on X.
“The $5M number is bogus. It is pushed by a Chinese hedge fund to slow investment in American AI startups, service their own shorts against American titans like Nvidia, and hide sanction evasion.”
Seena Rejal, chief commercial officer of NetMind, a London-headquartered startup that offers access to DeepSeek’s AI models via a distributed GPU network, said he saw no reason not to believe DeepSeek.
“Even if it’s off by a certain factor, it still is coming in as greatly efficient,” Rejal told CNBC in a phone interview earlier this week. “The logic of what they’ve explained is very sensible.”
However, some have claimed DeepSeek’s technology might not have been built from scratch.
“DeepSeek makes the same mistakes O1 makes, a strong indication the technology was ripped off,” billionaire investor Vinod Khosla said on X, without giving more details.
It’s a claim that OpenAI itself has alluded to, telling CNBC in a statement Wednesday that it is reviewing reports DeepSeek may have “inappropriately” used output data from its models to develop their AI model, a method referred to as “distillation.”
“We take aggressive, proactive countermeasures to protect our technology and will continue working closely with the U.S. government to protect the most capable models being built here,” an OpenAI spokesperson told CNBC.
Commoditization of AI
However the scrutiny surrounding DeepSeek shakes out, AI scientists broadly agree it marks a positive step for the industry.
Yann LeCun, chief AI scientist at Meta, said that DeepSeek’s success represented a victory for open-source AI models, not necessarily a win for China over the U.S. Meta is behind a popular open-source AI model called Llama.
“To people who see the performance of DeepSeek and think: ‘China is surpassing the US in AI.’ You are reading this wrong. The correct reading is: ‘Open source models are surpassing proprietary ones’,” he said in a post on LinkedIn.
“DeepSeek has profited from open research and open source (e.g. PyTorch and Llama from Meta). They came up with new ideas and built them on top of other people’s work. Because their work is published and open source, everyone can profit from it. That is the power of open research and open source.”
BARCELONA — China’s Huawei isn’t the only smartphone maker adding a third display to its devices.
At the Mobile World Congress (MWC) trade show in Barcelona, a number of firms were showing off their display technology innovations.
The South Korean tech giant Samsung revealed its new “trifold” concept devices at the event: the Flex G and Flex S.
The Flex G has three screens and folds flat inwards and outwards, a bit like a book. The Flex S, on the other hand, has a more zigzag-like shape. It’s meant to resemble an “S” — hence the name.
The Flex S is another concept device Samsung showed off at MWC. It folds in a more zigzag-like way to make an “S” shape.
Samsung stressed that its Flex G and S models were only concept devices — so don’t expect to find them on shelves anytime soon.
Still, it’s a sign of where smartphone makers are seeing the next wave of innovation.
‘Sea of sameness’
The smartphone market has hit something of a plateau over recent years, with many models not straying far from the standard form factor of a bar-shaped device.
Apple set the tone for what the devices in our pockets would look like when it launched the first iPhone in 2008. But smartphone makers are now trying to pull the market out of this so-called “sea of sameness.”
On Tuesday, British consumer tech startup Nothing launched its new Phone (3a), a 329-euro ($356.28) budget model with a quirky design and LED light system that lights up when you get calls or notifications.
Nothing co-founder Akis Evangelidis — who is planning a move to India as the startup plans an aggressive expansion push in the country — told CNBC the company is trying to shake up the smartphone market with something more fun and unique.
Using the Indian market as an example, Evangelidis said: “People are walking away from pure functional needs when it comes to product. They aspire to brands that have more of an emotional benefit, and I think that’s where the opportunity is.”
Innovating on display
However, although smartphone makers have been aggressively working to release new folding devices, the category remains a relatively niche area of the market.
Plus, folding phones can represent a big jump for the average consumer.
For one, they tend to be bulkier than non-folding phones because of the additional screen. And they’re not cheap, either. According to data from market research firm IDC, the average selling price of folding phones is nearly three times higher than that of normal smartphones — roughly $1,218 vs. $421 for non-folding phones.
While the foldable phone market grew 6.4% year-over-year to 19.3 million units, the category “represents only 1.6% of total global shipments,” according to Francisco Jeronimo, vice president EMEA for devices at IDC.
Nevertheless, this year at MWC, phone companies showed they’re getting better at developing folding phones that can better cater to everyday users.
For example, Oppo showed off its new Find N5 device this week. It only has two screens, but it’s a lot thinner than competing folding phones, such as Samsung’s Galaxy Fold 6.
Samsung currently holds the leading position in the global foldables segment. In 2024, it commanded a 32.9% share of the market. Huawei was close behind, with 23.1%, while Motorola was the third-biggest folding phone manufacturer with 17% market share.
And despite the punchy prices, these companies are betting consumers will be willing to pay for a more premium-grade experience.
MongoDB shares cratered more than 20% after the database software maker shared weak guidance that signaled a slowdown in growth.
For the fiscal 2026 year, the company said it expects adjusted earnings to range between $2.44 to $2.62 per share and revenue of $2.24 billion to $2.28. Analysts were expecting EPS of $3.34 and $2.32 billion in revenue.
The weak guidance stems from slower growth in the company’s Atlas cloud-based database service. The revenue projection would imply 12.7% growth, the slowest for the company going back to its 2017 stock market debut.
Finance chief Srdjan Tanjga said during an earnings call that the company is seeing slower-than-expected growth in new applications harnessing its Atlas cloud-based database service. However, MongoDB is beefing up hiring and going after deals with larger companies.
Read more CNBC tech news
For the fiscal first quarter, MongoDB forecasted 63 cents to 67 cents in adjusted earnings per share on $524 million to $529 million in revenue. Analysts polled by LSEG had expected EPS of 62 cents and revenue of $526.8 million.
Citing MongoDB’s weak outlook and slowdown in growth, Wells Fargo analyst Andrew Nowinski downgraded shares to equal weight and lowered his price target.
“With a smaller pool of multi-year deals, we believe it will be difficult to significantly outperform expectations in FY26 and therefore expect shares to remain range-bound,” he wrote.
MongoDB’s outlook offset stronger-than-expected fourth-quarter earnings. The company reported earnings of $1.28 per share, excluding items, on $548 million in revenue. Analysts polled by LSEG had anticipated EPS of 66 cents and $520 million in sales. Revenues rose 20% from a year ago.
MongoDB gained 1,900 customers in the quarter, reflecting a total of 54,500.
The Alibaba office building is seen in Nanjing, Jiangsu province, China, on Aug 28, 2024.
CFOTO | Future Publishing | Getty Images
Alibaba shares surged on Wednesday after the Chinese behemoth revealed a new reasoning model it claims can rival DeepSeek’s global blockbuster R1.
Hong Kong-listed shares of Alibaba ended the Thursday session up 8.39% — hitting a new 52-week high — with the company’s New York-trading stock rising around 2.5% in premarket deals. Alibaba shares have gained nearly 71% in Hong Kong in the year to date.
The Chinese giant on Thursday unveiled QwQ-32B, its latest AI reasoning model, which it said “rivals cutting-edge reasoning model, e.g., DeepSeek-R1.”
Alibaba’s QwQ-32B operates with 32 billion parameters compared to DeepSeek’s 671 billion parameters with 37 billion parameters actively engaged during inference — the process of running live data through a trained AI model in order to generate a prediction or tackle a task.
Parameters are variables that large language models (LLMs) — AI systems that can understand and generate human language — pick up during training and use in prediction and decision-making. A lower volume of parameters typically signals higher efficiency amid increasing demand for optimized AI that consumes fewer resources.
Alibaba said its new model achieved “impressive results” and the company can “continuously improve the performance especially in math and coding.”
Both established and emerging AI players around the world are racing to produce more efficient and higher-performance models since the unexpected launch of DeepSeek’s revolutionary R1 earlier this year.
“Looking ahead, revenue growth at Cloud Intelligence Group driven by AI will continue to accelerate,” Alibaba CEO Eddie Wu said at the time.
Optimism surrounding AI developments could lead to large gains for Alibaba stock and set the company’s earnings “on a more upwardly-pointing trajectory,” Bernstein analysts said.
“The pace of innovation is incredibly fast right now. It’s really good for the world to see this happening,” Futurum Group CEO Dan Newman told CNBC’s “Squawk Box Europe” on Thursday. “When DeepSeek came out, it made everyone sort of question, was OpenAi the final answer? Would the incumbents, the Microsofts, the Googles, or the Amazons that have all made massive investments win?”
He stressed that the large language models were increasingly “becoming commoditized” as developers look to drive down costs and improve access to users.
“As we see this more efficiency, this cost coming down, we’re also going to see use going off. The training era, which is what Nvidia really built its initial AI boom off, was a big moment,” Newman said. “But the inference, the consumption of AI, is really the future and this is going to exponentially increase that volume.”