Industry heavyweights from OpenAI CEO Sam Altman to former Baidu and Google scientist Andrew Ng have praised the open-source approach of DeepSeek, following its release of two advanced AI models.
Based in Hangzhou, capital of eastern Zhejiang province, DeepSeek stunned the global AI industry with its open-source reasoning model, R1. Released on January 20, the model showed capabilities comparable to closed-source models from ChatGPT creator OpenAI, but was said to be developed at significantly lower training costs.
DeepSeek said its foundation large language model, V3, released a few weeks earlier, cost only US$5.5 million to train. That statement stoked concerns that tech companies had been overspending on graphics processing units for AI training, leading to a major sell-off of AI chip supplier Nvidia’s shares last week.
OpenAI “has been on the wrong side of history here and needs to figure out a different open-source strategy”, Altman said last week in an “Ask Me Anything” session on internet forum Reddit. The US start-up has been taking a closed-source approach, keeping information such as the specific training methods and energy costs of its models tightly guarded.