High efficiency model with leading accuracy for reasoning, tool calling, chat, and instruction following.
Leading reasoning and agentic AI accuracy model for PC and edge.
The NV-EmbedCode model is a 7B Mistral-based embedding model optimized for code retrieval, supporting text, code, and hybrid queries.
Model for object detection, fine-tuned to detect charts, tables, and titles in documents.
Model for object detection, fine-tuned to detect charts, tables, and titles in documents.
Model for object detection, fine-tuned to detect charts, tables, and titles in documents.
Cutting-edge open multimodal model exceling in high-quality reasoning from images.
A lightweight, multilingual, advanced SLM text model for edge computing, resource constraint applications
Cutting-edge vision-language model exceling in retrieving text and metadata from images.
Route LLM requests to the best model for the task at hand.
Cutting-edge open multimodal model exceling in high-quality reasoning from image and audio inputs.
Multi-lingual model supporting speech-to-text recognition and translation.
Multi-lingual model supporting speech-to-text recognition and translation.
Latency-optimized language model excelling in code, math, general knowledge, and instruction-following.
Topic control model to keep conversations focused on approved topics, avoiding inappropriate content.
Industry leading jailbreak classification model for protection from adversarial attempts
Leading content safety model for enhancing the safety and moderation capabilities of LLMs
Chinese and English LLM targeting for language, coding, mathematics, reasoning, etc.
Transform PDFs into AI podcasts for engaging on-the-go audio content.
Multi-modal vision-language model that understands text/img/video and creates informative responses
Powerful mid-size code model with a 32K context length, excelling in coding in multiple languages.
Fine-tuned reranking model for multilingual, cross-lingual text question-answering retrieval, with long context support.
Model for object detection, fine-tuned to detect charts, tables, and titles in documents.
Automatic speech recognition model that transcribes speech in lower case English with record-setting accuracy and performance
Advanced AI model detects faces and identifies deep fake images.
Detects jailbreaking, bias, violence, profanity, sexual content, and unethical behavior
Advanced Small Language Model supporting RAG, summarization, classification, code, and agentic AI
Highly efficient Mixture of Experts model for RAG, summarization, entity extraction, and classification
Llama-3.1-Nemotron-70B-Instruct is a large language model customized by NVIDIA in order to improve the helpfulness of LLM generated responses.
Efficient hybrid state-space model designed for conversational and reasoning tasks.
Sovereign AI model trained on Japanese language that understands regional nuances.
Sovereign AI model trained on Japanese language that understands regional nuances.
State-of-the-art small language model delivering superior accuracy for chatbot, virtual assistants, and content generation.
Leaderboard topping reward model supporting RLHF for better alignment with human preferences.
Advanced state-of-the-art small language model with language understanding, superior reasoning, and text generation.
Cutting-edge vision-language model exceling in high-quality reasoning from images.
Cutting-edge vision-Language model exceling in high-quality reasoning from images.
Advanced state-of-the-art small language model with language understanding, superior reasoning, and text generation.
Unique language model that delivers an unmatched accuracy-efficiency performance.
Chinese and English LLM targeting for language, coding, mathematics, reasoning, etc.
Fine-tuned Llama 3.1 70B model for code generation, summarization, and multi-language tasks.
Robust image classification model for detecting and managing AI-generated content.
Sovereign AI model finetuned on Traditional Mandarin and English data using the Llama-3 architecture.
Sovereign AI model trained on Japanese language that understands regional nuances.
Cutting-edge open multimodal model exceling in high-quality reasoning from images.
State-of-the-art small language model delivering superior accuracy for chatbot, virtual assistants, and content generation.
Lightweight multilingual LLM powering AI applications in latency bound, memory/compute constrained environments
Advanced state-of-the-art LLM with language understanding, superior reasoning, and text generation.
Advanced state-of-the-art LLM with language understanding, superior reasoning, and text generation.
Grounding dino is an open vocabulary zero-shot object detection model.
ProteinMPNN is a deep learning model for predicting amino acid sequences for protein backbones.
Vision foundation model capable of performing diverse computer vision and vision language tasks.
Guardrail model to ensure that responses from LLMs are appropriate and safe
Advanced small language generative AI model for edge applications
Model for writing and interacting with code across a wide range of programming languages and tasks.
Advanced state-of-the-art model with language understanding, superior reasoning, and text generation.
Most advanced language model for reasoning, code, multilingual tasks; runs on a single GPU.
Multilingual text reranking model.
English text embedding model for question-answering retrieval.
Cutting-edge lightweight open language model exceling in high-quality reasoning.
Advanced programming model for code completion, summarization, and generation
Advanced programming model for code completion, summarization, and generation
Cutting-edge text generation model text understanding, transformation, and code generation.
Cutting-edge text generation model text understanding, transformation, and code generation.
Grades responses on five attributes helpfulness, correctness, coherence, complexity and verbosity.
Advanced text-to-image model for generating high quality images
Excels in NLP tasks, particularly in instruction-following, reasoning, and mathematics.
Advanced programming model for code generation, completion, reasoning, and instruction following.
Software programming LLM for code generation, completion, explanation, and multi-turn conversion.
Software programming LLM for code generation, completion, explanation, and multi-turn conversion.
A generative model of protein backbones for protein binder design.
Cutting-edge lightweight open language model exceling in high-quality reasoning.
Long context cutting-edge lightweight open language model exceling in high-quality reasoning.
Cutting-edge lightweight open language model exceling in high-quality reasoning.
Cutting-edge open multimodal model exceling in high-quality reasoning from images.
LLM to represent and serve the linguistic and cultural diversity of Southeast Asia
Lightweight, state-of-the-art open LLM with strong math and logical reasoning skills.
A general-purpose LLM with state-of-the-art performance in language understanding, coding, and RAG.
Optimized community model for text embedding.
Lightweight, state-of-the-art open LLM with strong math and logical reasoning skills.
An MOE LLM that follows instructions, completes requests, and generates creative text.
Powers complex conversations with superior contextual understanding, reasoning and text generation.
Advanced state-of-the-art LLM with language understanding, superior reasoning, and text generation.
Novel recurrent architecture based language model for faster inference when generating long sequences.
Cutting-edge model built on Google's Gemma-7B specialized for code generation and code completion.
GPU-accelerated model optimized for providing a probability score that a given passage contains the information to answer a question.
LLM capable of generating code from natural language and vice versa.
Run Google's DeepVariant optimized for GPU. Switch models for high accuracy on all major sequencers.
Stable Video Diffusion (SVD) is a generative diffusion model that leverages a single image as a conditioning frame to synthesize video sequences.
A fast generative text-to-image model that can synthesize photorealistic images from a text prompt in a single network evaluation
An MOE LLM that follows instructions, completes requests, and generates creative text.