3 weeks, 3 days ago

Nvidia CEO Jensen Huang unveils Dynamo, an open-source inference framework for AI inferencing, at GTC 2025

AI chipmaker Nvidia on Tuesday unveiled Dynamo, an open-source inference framework designed to enhance the deployment of generative AI and reasoning models across large-scale, distributed environments. Announced at the GTC 2025 conference, Dynamo aims to significantly boost inference performance while reducing operational costs for AI applications. The framework supports major AI inference backends, including PyTorch, SGLang, NVIDIA TensorRT-LLM, and vLLM, providing developers and AI researchers with the flexibility to integrate Dynamo into diverse AI workflows. For enterprises seeking accelerated deployment and enterprise-grade support, Nvidia plans to include Dynamo with its NIM microservices, which are part of the NVIDIA AI Enterprise suite.

The Hindu

Discover Related