How a Chinese start-up is changing how AI models are trained and outperforming OpenAI, Meta
1 week ago

How a Chinese start-up is changing how AI models are trained and outperforming OpenAI, Meta

Firstpost  

DeepSeek’s model boasts an impressive 671 billion parameters, placing it on par with some of the most advanced models globally. Yet, it was developed at a fraction of the cost incurred by giants like Meta and OpenAI, requiring only $5.58 million and 2.78 million GPU hours DeepSeek's figures are a stark contrast to Meta’s Llama 3.1, which needed 30.8 million GPU hours and more advanced hardware to train. Image Credit: Reuters Chinese start-up DeepSeek is making waves in AI developers all over the world, with the release of its latest large language model, DeepSeek V3. These figures are a stark contrast to Meta’s Llama 3.1, which needed 30.8 million GPU hours and more advanced hardware to train.

History of this topic

ModelBest launches new AI large model
11 months, 1 week ago
China's Tencent debuts large language AI model, says open for enterprise use
1 year, 4 months ago

Discover Related