
Cutting-edge vision-language model exceling in retrieving text and metadata from images.

Nemotron Nano 12B v2 VL enables multi-image and video understanding, along with visual Q&A and summarization capabilities.

Leading multilingual content safety model for enhancing the safety and moderation capabilities of LLMs

High‑efficiency LLM with hybrid Transformer‑Mamba design, excelling in reasoning and agentic tasks.

High efficiency model with leading accuracy for reasoning, tool calling, chat, and instruction following.

Built for agentic workflows, this model excels in coding, instruction following, and function calling

Multi-modal vision-language model that understands text/img and creates informative responses

State-of-the-art open model for reasoning, code, math, and tool calling - suitable for edge agents

Superior inference efficiency with highest accuracy for scientific and complex math reasoning, coding, tool calling, and instruction following.

High efficiency model with leading accuracy for reasoning, tool calling, chat, and instruction following.

Leading reasoning and agentic AI accuracy model for PC and edge.

Multi-modal vision-language model that understands text/img/video and creates informative responses

A bilingual Hindi-English SLM for on-device inference, tailored specifically for Hindi Language.

Leaderboard topping reward model supporting RLHF for better alignment with human preferences.

Optimized SLM for on-device inference and fine-tuned for roleplay, RAG and function calling