DeepSeek’s AI claims have shaken the world — but not everyone’s convinced

DeepSeek’s AI claims have shaken the world — but not everyone’s convinced


Chinese artificial intelligence firm DeepSeek rocked markets this week with claims its new AI model outperforms OpenAI’s and cost a fraction of the price to build.

The assertions — specifically that DeepSeek’s large language model cost just $5.6 million to train — have sparked concerns over the eyewatering sums that tech giants are currently spending on computing infrastructure required to train and run advanced AI workloads.

But not everyone is convinced by DeepSeek’s claims.

CNBC asked industry experts for their views on DeepSeek, and how it actually compares to OpenAI, creator of viral chatbot ChatGPT which sparked the AI revolution.

What is DeepSeek?

Last week, DeepSeek released R1, its new reasoning model that rivals OpenAI’s o1. A reasoning model is a large language model that breaks prompts down into smaller pieces and considers multiple approaches before generating a response. It is designed to process complex problems in a similar way to humans.

DeepSeek was founded in 2023 by Liang Wenfeng, co-founder of AI-focused quantitative hedge fund High-Flyer, to focus on large language models and reaching artificial general intelligence, or AGI.

AGI as a concept loosely refers to the idea of an AI that equals or surpasses human intellect on a wide range of tasks.

Much of the technology behind R1 isn’t new. What is notable, however, is that DeepSeek is the first to deploy it in a high-performing AI model with — according to the company — considerable reductions in power requirements.

“The takeaway is that there are many possibilities to develop this industry. The high-end chip/capital intensive way is one technological approach,” said Xiaomeng Lu, director of Eurasia Group’s geo-technology practice.

“But DeepSeek proves we are still in the nascent stage of AI development and the path established by OpenAI may not be the only route to highly capable AI.” 

How is it different from OpenAI?

DeepSeek has two main systems that have garnered buzz from the AI community: V3, the large language model that unpins its products, and R1, its reasoning model.

Both models are open-source, meaning their underlying code is free and publicly available for other developers to customize and redistribute.

DeepSeek’s models are much smaller than many other large language models. V3 has a total of 671 billion parameters, or variables that the model learns during training. And while OpenAI doesn’t disclose parameters, experts estimate its latest model to have at least a trillion.

In terms of performance, DeepSeek says its R1 model achieves performance comparable to OpenAI’s o1 on reasoning tasks, citing benchmarks including AIME 2024, Codeforces, GPQA Diamond, MATH-500, MMLU and SWE-bench Verified.

Read more DeepSeek coverage

In a technical report, the company said its V3 model had a training cost of only $5.6 million — a fraction of the billions of dollars that notable Western AI labs such as OpenAI and Anthropic have spent to train and run their foundational AI models. It isn’t yet clear how much DeepSeek costs to run, however.

If the training costs are accurate, though, it means the model was developed at a fraction of the cost of rival models by OpenAI, Anthropic, Google and others.

Daniel Newman, CEO of tech insight firm The Futurum Group, said these developments suggest “a massive breakthrough,” although he shed some doubt on the exact figures.

“I believe the breakthroughs of DeepSeek indicate a meaningful inflection for scaling laws and are a real necessity,” he said. “Having said that, there are still a lot of questions and uncertainties around the full picture of costs as it pertains to the development of DeepSeek.”

Meanwhile, Paul Triolio, senior VP for China and technology policy lead at advisory firm DGA Group, noted it was difficult to draw a direct comparison between DeepSeek’s model cost and that of major U.S. developers.

“The 5.6 million figure for DeepSeek V3 was just for one training run, and the company stressed that this did not represent the overall cost of R&D to develop the model,” he said. “The overall cost then was likely significantly higher, but still lower than the amount spent by major US AI companies.” 

DeepSeek wasn’t immediately available for comment when contacted by CNBC.

Comparing DeepSeek, OpenAI on price

DeepSeek and OpenAI both disclose pricing for their models’ computations on their websites.

DeepSeek says R1 costs 55 cents per 1 million tokens of inputs — “tokens” referring to each individual unit of text processed by the model — and $2.19 per 1 million tokens of output.

In comparison, OpenAI’s pricing page for o1 shows the firm charges $15 per 1 million input tokens and $60 per 1 million output tokens. For GPT-4o mini, OpenAI’s smaller, low-cost language model, the firm charges 15 cents per 1 million input tokens.

Skepticism over chips

DeepSeek’s reveal of R1 has already led to heated public debate over the veracity of its claim — not least because its models were built despite export controls from the U.S. restricting the use of advanced AI chips to China. 

DeepSeek claims it had its breakthrough using mature Nvidia clips, including H800 and A100 chips, which are less advanced than the chipmaker’s cutting-edge H100s, which can’t be exported to China.

However, in comments to CNBC last week, Scale AI CEO Alexandr Wang, said he believed DeepSeek used the banned chips — a claim that DeepSeek denies.

LinkedIn co-founder Reid Hoffman: DeepSeek AI proves this is now a 'game-on competition' with China

Nvidia has since come out and said that the GPUs that DeepSeek used were fully export-compliant.

The real deal or not?

Industry experts seem to broadly agree that what DeepSeek has achieved is impressive, although some have urged skepticism over some of the Chinese company’s claims.

“DeepSeek is legitimately impressive, but the level of hysteria is an indictment of so many,” U.S. entrepreneur Palmer Luckey, who founded Oculus and Anduril wrote on X.

“The $5M number is bogus. It is pushed by a Chinese hedge fund to slow investment in American AI startups, service their own shorts against American titans like Nvidia, and hide sanction evasion.”

Seena Rejal, chief commercial officer of NetMind, a London-headquartered startup that offers access to DeepSeek’s AI models via a distributed GPU network, said he saw no reason not to believe DeepSeek.

“Even if it’s off by a certain factor, it still is coming in as greatly efficient,” Rejal told CNBC in a phone interview earlier this week. “The logic of what they’ve explained is very sensible.”

However, some have claimed DeepSeek’s technology might not have been built from scratch.

“DeepSeek makes the same mistakes O1 makes, a strong indication the technology was ripped off,” billionaire investor Vinod Khosla said on X, without giving more details.

It’s a claim that OpenAI itself has alluded to, telling CNBC in a statement Wednesday that it is reviewing reports DeepSeek may have “inappropriately” used output data from its models to develop their AI model, a method referred to as “distillation.”

“We take aggressive, proactive countermeasures to protect our technology and will continue working closely with the U.S. government to protect the most capable models being built here,” an OpenAI spokesperson told CNBC.

Commoditization of AI

However the scrutiny surrounding DeepSeek shakes out, AI scientists broadly agree it marks a positive step for the industry.

Yann LeCun, chief AI scientist at Meta, said that DeepSeek’s success represented a victory for open-source AI models, not necessarily a win for China over the U.S. Meta is behind a popular open-source AI model called Llama.

“To people who see the performance of DeepSeek and think: ‘China is surpassing the US in AI.’ You are reading this wrong. The correct reading is: ‘Open source models are surpassing proprietary ones’,” he said in a post on LinkedIn.

“DeepSeek has profited from open research and open source (e.g. PyTorch and Llama from Meta). They came up with new ideas and built them on top of other people’s work. Because their work is published and open source, everyone can profit from it. That is the power of open research and open source.”

WATCH: Why DeepSeek is putting America’s AI lead in jeopardy

Why China's DeepSeek is putting America's AI lead in jeopardy

– CNBC’s Katrina Bishop and Hayden Field contributed to this report



Source

Astronomer HR chief Kristin Cabot resigns following Coldplay ‘kiss cam’ incident
Technology

Astronomer HR chief Kristin Cabot resigns following Coldplay ‘kiss cam’ incident

Chris Martin of Coldplay performs live at San Siro Stadium, Milan, Italy, in July 2017. Mairo Cinquetti | NurPhoto | Getty Images Days after Astronomer CEO Andy Byron resigned from the tech startup, the HR exec who was with him at the infamous Coldplay concert has left as well. “Kristin Cabot is no longer with […]

Read More
Tesla dumped 75% of its bitcoin at one of the worst times, losing out on billions
Technology

Tesla dumped 75% of its bitcoin at one of the worst times, losing out on billions

Thiago Prudencio | LightRocket | Getty Images Tesla missed on the top and bottom lines in the second quarter, but another miss was buried in its investor deck. The company’s digital assets are currently valued at $1.24 billion. That’s up substantially from $722 million a year ago. But anyone who’s been following the crypto market […]

Read More
Intel beats on revenue, issues better-than-expected forecast
Technology

Intel beats on revenue, issues better-than-expected forecast

The Intel logo is displayed on a sign in front of Intel headquarters on July 16, 2025 in Santa Clara, California. Justin Sullivan | Getty Images Intel reported second-quarter results on Thursday that beat Wall Street expectations on revenue, as the company’s new CEO Lip-Bu Tan announced significant cuts in chip factory construction. The stock […]

Read More