At CES 2025, NVIDIA CEO Jensen Huang introduced the latest Nemotron models, including the Llama Nemotron large language models (LLMs) and Cosmos Nemotron vision language models (VLMs), designed to enhance agentic AI capabilities and drive enterprise productivity.
The Llama Nemotron models, based on the Llama foundation models, empower developers to build AI agents for applications such as customer support, fraud detection, and supply chain optimization.
The Nemotron lineup is available in Nano, Super, and Ultra configurations to meet diverse deployment needs, from low-latency real-time applications to high-accuracy data center solutions. Optimized for computational efficiency and precision, these models support agentic AI tasks like instruction following, coding, and mathematical operations.
According to Huang, “Llama 3.1 is a complete phenomenon, with the downloads reaching 650,000 times. It has been derived and turned into other models, about 60,000 different models. It is singularly why every single enterprise and industry has been activated to start working on AI.”
Also read, Zuckerberg’s Share His New Content Moderation Plan
NVIDIA announced that these models will be accessible as downloadable resources or deployable microservices across various computing platforms, including data centers and edge devices. The Llama Nemotron and Cosmos Nemotron models will soon be available on build.nvidia.com, Hugging Face, and via the NVIDIA Developer Program.
For enterprise-grade implementations, NVIDIA will offer support through the NVIDIA AI Enterprise platform, leveraging accelerated cloud and data center infrastructure.