55 years, 3 months ago

Meta’s Next Llama AI Models Are Training on a GPU Cluster ‘Bigger Than Anything’ Else

Meta CEO Mark Zuckerberg laid down the newest marker in generative AI training on Wednesday, saying that the next major release of the company’s Llama model is being trained on a cluster of GPUs that’s “bigger than anything” else that’s been reported. “We're training the Llama 4 models on a cluster that is bigger than 100,000 H100s, or bigger than anything that I've seen reported for what others are doing,” Zuckerberg said, referring to the Nvidia chips popular for training AI systems. On Wednesday, Zuckerberg declined to offer details on Llama 4’s potential advanced capabilities but vaguely referred to “new modalities,” “stronger reasoning,” and “much faster.” Meta’s approach to AI is proving a wild card in the corporate race for dominance. Although touted as “open source” by Meta, the Llama license does impose some restrictions on the model’s commercial use.

Wired

Discover Related