Chinese start-up DeepSeek has emerged as “the biggest dark horse” in the open-source large language model (LLM) arena in 2025 with its latest release DeepSeek V3.
That assessment came from Jim Fan, a senior research scientist at Nvidia and lead of its AI Agents Initiative.
“[The new AI model] shows that resource constraints force you to reinvent yourself in spectacular ways,” Fan writes.
According to 10Web analysis, DeepSeek excels in tasks such as arithmetic, math, reasoning, and coding, surpassing even some of the most renowned models like GPT-4 and LLaMA3-70B.
DeepSeek V3 comes with 671 billion parameters and was trained in around two months at a cost of US$5.58 million, using significantly fewer computing resources than models developed by bigger tech firms such as Facebook parent Meta Platforms and ChatGPT creator OpenAI.
In AI, a high number of parameters is pivotal in enabling an LLM to adapt to more complex data patterns and make precise predictions.
Open source gives public access to a software program’s source code, allowing third-party developers to modify or share its design, fix broken links or scale up its capabilities.
Sources: South China Morning Post and 10Web