
Distilled version of Llama 3.3 70B using reasoning data generated by DeepSeek R1 for enhanced performance.

Build advanced AI agents for providers and patients using this developer example powered by NeMo Microservices, NVIDIA Nemotron, Riva ASR and TTS, and NVIDIA LLM NIM

Cutting-edge vision-language model exceling in retrieving text and metadata from images.

Nemotron Nano 12B v2 VL enables multi-image and video understanding, along with visual Q&A and summarization capabilities.

Leading multilingual content safety model for enhancing the safety and moderation capabilities of LLMs

Multilingual, cross-lingual embedding model for long-document QA retrieval, supporting 26 languages.

Transform your scene idea into ready-to-use 3D assets using Llama 3.1 8B, NV SANA, and Microsoft TRELLIS

High‑efficiency LLM with hybrid Transformer‑Mamba design, excelling in reasoning and agentic tasks.

High efficiency model with leading accuracy for reasoning, tool calling, chat, and instruction following.

Multilingual, cross-lingual embedding model for long-document QA retrieval, supporting 26 languages.

Multi-modal model to classify safety for input prompts as well output responses.

GPU-accelerated model optimized for providing a probability score that a given passage contains the information to answer a question.

Multimodal question-answer retrieval representing user queries as text and documents as images.

Built for agentic workflows, this model excels in coding, instruction following, and function calling

Multi-modal vision-language model that understands text/img and creates informative responses

State-of-the-art open model for reasoning, code, math, and tool calling - suitable for edge agents

Superior inference efficiency with highest accuracy for scientific and complex math reasoning, coding, tool calling, and instruction following.

A general purpose multimodal, multilingual 128 MoE model with 17B parameters.

A multimodal, multilingual 16 MoE model with 17B parameters.

Build a custom enterprise research assistant powered by state-of-the-art models that process and synthesize multimodal data, enabling reasoning, planning, and refinement to generate comprehensive reports.

High efficiency model with leading accuracy for reasoning, tool calling, chat, and instruction following.

Leading reasoning and agentic AI accuracy model for PC and edge.

Distilled version of Llama 3.1 8B using reasoning data generated by DeepSeek R1 for enhanced performance.

Power fast, accurate semantic search across multimodal enterprise data with NVIDIA’s RAG Blueprint—built on NeMo Retriever and Nemotron models—to connect your agents to trusted, authoritative sources of knowledge.

Topic control model to keep conversations focused on approved topics, avoiding inappropriate content.

Leading content safety model for enhancing the safety and moderation capabilities of LLMs

Generate detailed, structured reports on any topic using LangGraph and Llama3.3 70B NIM.

Document your github repositories with AI Agents using CrewAI and Llama3.3 70B NIM.

Multi-modal vision-language model that understands text/img/video and creates informative responses

Multilingual and cross-lingual text question-answering retrieval with long context support and optimized data storage efficiency.

Fine-tuned reranking model for multilingual, cross-lingual text question-answering retrieval, with long context support.

Advanced LLM for reasoning, math, general knowledge, and function calling

A bilingual Hindi-English SLM for on-device inference, tailored specifically for Hindi Language.

Rapidly identify and mitigate container security vulnerabilities with generative AI.

Sovereign AI model trained on Japanese language that understands regional nuances.

Sovereign AI model trained on Japanese language that understands regional nuances.

Leaderboard topping reward model supporting RLHF for better alignment with human preferences.

Advanced state-of-the-art small language model with language understanding, superior reasoning, and text generation.

Cutting-edge vision-language model exceling in high-quality reasoning from images.

Cutting-edge vision-Language model exceling in high-quality reasoning from images.

Advanced state-of-the-art small language model with language understanding, superior reasoning, and text generation.

Fine-tuned Llama 3.1 70B model for code generation, summarization, and multi-language tasks.

Sovereign AI model finetuned on Traditional Mandarin and English data using the Llama-3 architecture.

Sovereign AI model trained on Japanese language that understands regional nuances.

Optimized SLM for on-device inference and fine-tuned for roleplay, RAG and function calling

Advanced LLM for synthetic data generation, distillation, and inference for chatbots, coding, and domain-specific tasks.

Powers complex conversations with superior contextual understanding, reasoning and text generation.

Advanced state-of-the-art model with language understanding, superior reasoning, and text generation.

Advanced LLM to generate high-quality, context-aware responses for chatbots and search engines.

Powers complex conversations with superior contextual understanding, reasoning and text generation.

Advanced state-of-the-art LLM with language understanding, superior reasoning, and text generation.