Multi-modal vision-language model that understands text/img/video and creates informative responses
Generates physics-aware video world states from text and image prompts for physical AI development.
Generates future frames of a physics-aware world state based on simply an image or short video prompt for physical AI development.
SAM 2 is a segmentation model that enables fast, precise selection of any object in any video or image.
Context-aware chart extraction that can detect 18 classes for chart basic elements, excluding plot elements.
Model for table extraction that receives an image as input, runs OCR on the image, and returns the text within the image and its bounding boxes.
Advanced AI model detects faces and identifies deep fake images.
Shutterstock Generative 3D service for 360 HDRi generation. Trained on NVIDIA Edify using Shutterstock’s licensed creative libraries.
Advanced state-of-the-art small language model with language understanding, superior reasoning, and text generation.
Cutting-edge vision-language model exceling in high-quality reasoning from images.
Cutting-edge vision-Language model exceling in high-quality reasoning from images.
Advanced state-of-the-art small language model with language understanding, superior reasoning, and text generation.
Generates consistent characters across a series of images without requiring additional training.
Multi-modal vision-language model that understands text/img/video and creates informative responses
Robust image classification model for detecting and managing AI-generated content.
Cutting-edge open multimodal model exceling in high-quality reasoning from images.
NV-DINOv2 is a visual foundation model that generates vector embeddings for the input image.
Advanced state-of-the-art LLM with language understanding, superior reasoning, and text generation.
Advanced state-of-the-art LLM with language understanding, superior reasoning, and text generation.
An enterprise-grade text-to-image model trained on a compliant dataset produces high quality images.
Vision foundation model capable of performing diverse computer vision and vision language tasks.
AI-powered search for OpenUSD data, 3D models, images, and assets using text or image-based inputs.
Shutterstock Generative 3D service for 3D asset generation. Trained on NVIDIA Edify using Shutterstock’s licensed creative libraries
Getty Images’ API service for 4K image generation. Trained on NVIDIA Edify using Getty Images' commercially safe creative libraries.
Powers complex conversations with superior contextual understanding, reasoning and text generation.
Advanced state-of-the-art model with language understanding, superior reasoning, and text generation.
Cutting-edge text generation model text understanding, transformation, and code generation.
Cutting-edge text generation model text understanding, transformation, and code generation.
Advanced text-to-image model for generating high quality images
OCDNet and OCRNet are pre-trained models designed for optical character detection and recognition respectively.
Visual Changenet detects pixel-level change maps between two images and outputs a semantic change segmentation mask
EfficientDet-based object detection network to detect 100 specific retail objects from an input video.
Cutting-edge open multimodal model exceling in high-quality reasoning from images.
A general-purpose LLM with state-of-the-art performance in language understanding, coding, and RAG.
Powers complex conversations with superior contextual understanding, reasoning and text generation.
Advanced state-of-the-art LLM with language understanding, superior reasoning, and text generation.
Generate images and stunning visuals with realistic aesthetics.
VISTA-3D is a specialized interactive foundation model for segmenting and anotating human anatomies.
Cutting-edge text generation model text understanding, transformation, and code generation.
Stable Video Diffusion (SVD) is a generative diffusion model that leverages a single image as a conditioning frame to synthesize video sequences.
A fast generative text-to-image model that can synthesize photorealistic images from a text prompt in a single network evaluation