Create AI agents that reason, plan, reflect and refine to produce high-quality reports based on source materials of your choice.
Develop AI powered weather analysis and forecasting application visualizing multi-layered geospatial data.
Generate exponentially large amounts of synthetic motion trajectories for robot manipulation from just a few human demonstrations.
Generates physics-aware video world states from text and image prompts for physical AI development.
Generates future frames of a physics-aware world state based on simply an image or short video prompt for physical AI development.
Simulate, test, and optimize physical AI and robotic fleets at scale in industrial digital twins before real-world deployment.
End-to-end autonomous driving stack integrating perception, prediction, and planning with sparse scene representations for efficiency and safety.
Leading reasoning and agentic AI accuracy model for PC and edge.
The NV-EmbedCode model is a 7B Mistral-based embedding model optimized for code retrieval, supporting text, code, and hybrid queries.
Distilled version of Llama 3.1 8B using reasoning data generated by DeepSeek R1 for enhanced performance.
Route LLM requests to the best model for the task at hand.
Distilled version of Qwen 2.5 32B using reasoning data generated by DeepSeek R1 for enhanced performance.
Distilled version of Qwen 2.5 14B using reasoning data generated by DeepSeek R1 for enhanced performance.
Distilled version of Qwen 2.5 7B using reasoning data generated by DeepSeek R1 for enhanced performance.
Lightweight multilingual LLM powering AI applications in latency bound, memory/compute constrained environments
This workflow shows how generative AI can generate DNA sequences that can be translated into proteins for bioengineering.
State-of-the-art, high-efficiency LLM excelling in reasoning, math, and coding.
Connect AI applications to multimodal enterprise data with a retrieval augmented generation (RAG) pipeline.
Topic control model to keep conversations focused on approved topics, avoiding inappropriate content.
This blueprint shows how generative AI and accelerated NIM microservices can design protein binders smarter and faster.
Transform PDFs into AI podcasts for engaging on-the-go audio content.
Trace and evaluate AI Agents with Weights & Biases.
Automate voice AI agents with NVIDIA NIM microservices and Pipecat.
Automate research, and generate blogs with AI Agents using LlamaIndex and Llama3.3-70B NIM LLM.
Generate detailed, structured reports on any topic using LangGraph and Llama3.3 70B NIM
Document your github repositories with AI Agents using CrewAI and Llama3.3 70B NIM.
Generates physics-aware video world states from text and image prompts for physical AI development.
Generates future frames of a physics-aware world state based on simply an image or short video prompt for physical AI development.
Multilingual and cross-lingual text question-answering retrieval with long context support and optimized data storage efficiency.
Fine-tuned reranking model for multilingual, cross-lingual text question-answering retrieval, with long context support.
Context-aware chart extraction that can detect 18 classes for chart basic elements, excluding plot elements.
Model for object detection, fine-tuned to detect charts, tables, and titles in documents.
FourCastNet predicts global atmospheric dynamics of various weather / climate variables.
Advanced AI model detects faces and identifies deep fake images.
Ingest massive volumes of live or archived videos and extract insights for summarization and interactive Q&A
Enhance and modify high-quality compositions using real-time rendering and generative AI output without affecting a hero product asset.
Efficiently refine retrieval results over multiple sources and languages.
World-class multilingual and cross-lingual question-answering retrieval.
Create intelligent virtual assistants for customer service across every industry
Advanced Small Language Model supporting RAG, summarization, classification, code, and agentic AI
Rapidly identify and mitigate container security vulnerabilities with generative AI.
Sovereign AI model trained on Japanese language that understands regional nuances.
Sovereign AI model trained on Japanese language that understands regional nuances.
Unique language model that delivers an unmatched accuracy-efficiency performance.
Robust image classification model for detecting and managing AI-generated content.
Ingest and extract highly accurate insights contained in text, graphs, charts, and tables within massive volumes of PDF documents.
Create intelligent, interactive avatars for customer service across industries
Sovereign AI model finetuned on Traditional Mandarin and English data using the Llama-3 architecture.
Sovereign AI model trained on Japanese language that understands regional nuances.
This blueprint shows how generative AI and accelerated NIM microservices can design optimized small molecules smarter and faster.
Cutting-edge MOE based LLM designed to excel in a wide array of generative AI tasks.
Cutting-edge MOE based LLM designed to excel in a wide array of generative AI tasks.
Lightweight multilingual LLM powering AI applications in latency bound, memory/compute constrained environments
Grounding dino is an open vocabulary zero-shot object detection model.
Vision foundation model capable of performing diverse computer vision and vision language tasks.
Advanced small language generative AI model for edge applications
Most advanced language model for reasoning, code, multilingual tasks; runs on a single GPU.
Multilingual text reranking model.
English text embedding model for question-answering retrieval.
Multilingual text question-answering retrieval, transforming textual information into dense vector representations.
Generates high-quality numerical embeddings from text inputs.
Visual Changenet detects pixel-level change maps between two images and outputs a semantic change segmentation mask
EfficientDet-based object detection network to detect 100 specific retail objects from an input video.
Cutting-edge open multimodal model exceling in high-quality reasoning from images.
An MOE LLM that follows instructions, completes requests, and generates creative text.
Stable Video Diffusion (SVD) is a generative diffusion model that leverages a single image as a conditioning frame to synthesize video sequences.
An MOE LLM that follows instructions, completes requests, and generates creative text.