January 6, 2026

NVIDIA Release Open Source Models On Hugging Face

Featured image for “NVIDIA Release Open Source Models On Hugging Face”

As an AI Architect with experience designing scalable, efficient systems for enterprises and research, I’m always on the lookout for breakthroughs that democratise access to cutting-edge tools. NVIDIA’s latest announcement today has me genuinely excited, it’s a massive step forward in open-source AI. They’re expanding their “open model universe” with new models, datasets, and tools tailored for agentic AI, physical AI, autonomous vehicles (AVs), robotics, and biomedical applications. This isn’t just about releasing code; it’s about providing the building blocks for trustworthy, real-world AI systems that can reason, interact, and innovate across domains.

Backed by unprecedented scale, think 10 trillion language tokens, 500,000 robotics trajectories, 455,000 protein structures, and 100 terabytes of vehicle sensor data, these resources are freely available on platforms like Hugging Face and GitHub. As someone who architects AI pipelines, I see this as a boon for developers: easier customisation, faster prototyping, and reduced barriers to entry. Let’s dive into the key highlights, including model sizes where available, and explore how these can supercharge your next project.

NVIDIA Nemotron: Powering Agentic AI with Trust and Efficiency

Agentic AI, systems that can act autonomously and safely, are the future of interactive applications, from chatbots to enterprise assistants. NVIDIA’s Nemotron family builds on their Nemotron 3 series, introducing specialised open models for speech, retrieval-augmented generation (RAG), and safety.

Key Models and Sizes:

  • Nemotron Speech: Includes the Nemotron Speech Streaming EN 0.6B (600 million parameters), optimised for real-time, low-latency speech recognition. It delivers 10x faster performance than competitors in benchmarks.
  • Nemotron RAG: Features the Llama Nemotron Embed VL 1B v2 (1 billion parameters) for embeddings and reranking.
  • Nemotron Safety: The Llama-3.1-Nemotron-Safety-Guard-8B-v3 (8 billion parameters) for content moderation.
  • Availability: Hugging Face and GitHub.

From an architecture perspective, these models’ efficiency allows integration into hybrid systems without massive compute overhead. Companies like Bosch and ServiceNow are already adopting them!

NVIDIA Cosmos: Enabling Physical AI with Reasoning and Simulation

Physical AI—agents that understand and interact with the real world—is a holy grail for robotics and automation. NVIDIA Cosmos introduces tools for reasoning VLMs and synthetic data generation.

Key Models and Sizes:

  • Cosmos Reason 2: 2B and 8B variants, excelling at spatial-temporal understanding.
  • Cosmos Transfer/Predict 2.5: Optimised for synthetic video generation.
  • Availability: Hugging Face and GitHub.

I appreciate how Cosmos bridges simulation and reality – perfect for prototyping on standard hardware.

NVIDIA Alpamayo: Revolutionizing Autonomous Vehicles with Reasoning

Autonomous vehicles demand models that perceive and explain actions. NVIDIA Alpamayo delivers the first open, large-scale reasoning VLA for AVs.

Key Models and Sizes:

  • Alpamayo 1 (Alpamayo-R1-10B): 10 billion parameters for generalizable driving.

Tools:

  • AlpaSim simulation framework and massive datasets.

This 10B model is a sweet spot for AV research – powerful yet deployable.

NVIDIA Isaac GR00T: Foundation for Humanoid Robotics

Humanoids need models for full-body control with contextual reasoning. Isaac GR00T advances this with an open VLA foundation model.

Key Models and Sizes:

  • Isaac GR00T N1.6: 3 billion parameters for manipulation in diverse environments.

At 3B parameters, it’s efficient and accessible for robotics architects. Adopted by Franka Robotics and others!

NVIDIA Clara: Transforming Biomedical AI for Healthcare Breakthroughs

Biomedical AI promises faster drug discovery. NVIDIA Clara provides open frameworks for protein design and molecular predictions.

Key Models and Sizes:

  • La-Proteina: ~160 million parameters for generating protein structures.
  • Others like KERMT and RNAPro for drug interactions and RNA forecasting.
  • Clara’s tools lower costs in drug development – potentially designing safer proteins in days!

The Bigger Picture: Why This Matters for AI Architects

NVIDIA’s push emphasises openness and scale, with models from compact (0.6B) to mid-sized (10B), all optimised for real-world deployment. This ecosystem lets me mix and match for innovative pipelines. It’s an invitation to innovate responsibly!

If you’re building AI systems, explore on Hugging Face or GitHub. For more details: NVIDIA’s announcement.

What’s your take? How will you leverage these?

Written by

Liam Wytcherley

Share This Article: