
A personal Grace Blackwell AI supercomputer on your desk.
The leading open models built by the community, optimized and accelerated by NVIDIA's enterprise-ready inference runtime.

80B parameter AI model with hybrid reasoning, MoE architecture, support for 119 languages.

Excels in agentic coding and browser use and supports 256K context, delivering top results.

Smaller Mixture of Experts (MoE) text-only LLM for efficient AI reasoning and math

High‑efficiency LLM with hybrid Transformer‑Mamba design, excelling in reasoning and agentic tasks.
Get started with workflows and code samples to build AI applications from the ground up.

Build a custom enterprise research assistant powered by state-of-the-art models that process and synthesize multimodal data, enabling reasoning, planning, and refinement to generate comprehensive reports.

Ingest massive volumes of live or archived videos and extract insights for summarization and interactive Q&A

Power fast, accurate semantic search across multimodal enterprise data with NVIDIA’s RAG Blueprint—built on NeMo Retriever and Nemotron models—to connect your agents to trusted, authoritative sources of knowledge.

Improve safety, security, and privacy of AI systems at build, deploy and run stages.