Cutting-edge vision-language model exceling in high-quality reasoning from images.
Cutting-edge vision-Language model exceling in high-quality reasoning from images.
Multi-modal vision-language model that understands text/images and generates informative responses
Robust image classification model for detecting and managing AI-generated content.
NV-DINOv2 is a visual foundation model that generates vector embeddings for the input image.
An enterprise-grade text-to-image model trained on a compliant dataset produces high quality images.
Vision foundation model capable of performing diverse computer vision and vision language tasks.
AI-powered search for OpenUSD data, 3D models, images, and assets using text or image-based inputs.
Shutterstock Early Access preview of Generative 3D service for 360 HDRi generation. Trained on NVIDIA Edify using Shutterstock’s licensed creative libraries.
Shutterstock Generative 3D service for 3D asset generation. Trained on NVIDIA Edify using Shutterstock’s licensed creative libraries
Getty Images’ API service for 4K image generation. Trained on NVIDIA Edify using Getty Images' commercially safe creative libraries.
GPU-accelerated model optimized for providing a probability score that a given passage contains the information to answer a question.
GPU-accelerated generation of text embeddings used for question-answering retrieval.
GPU-accelerated generation of text embeddings used for question-answering retrieval.
MAISI is a pre-trained volumetric (3D) CT Latent Diffusion Generative Model.
NV-CLIP is a multimodal embeddings model for image and text.
Advanced text-to-image model for generating high quality images
OCDNet and OCRNet are pre-trained models designed for optical character detection and recognition respectively.
Embedding model for text retrieval tasks, excelling in dense, multi-vector, and sparse retrieval.
Visual Changenet detects pixel-level change maps between two images and outputs a semantic change segmentation mask
EfficientDet-based object detection network to detect 100 specific retail objects from an input video.
Cutting-edge open multimodal model exceling in high-quality reasoning from images.
Vision language model adept at comprehending text and visual inputs to produce informative responses
GPU-accelerated generation of text embeddings.
GPU-accelerated generation of text embeddings used for question-answering retrieval.
GPU-accelerated model optimized for providing a probability score that a given passage contains the information to answer a question.
Groundbreaking multimodal model designed to understand and reason about visual elements in images.
One-shot visual language understanding model that translates images of plots into tables.
Multi-modal vision-language model that understands text/images and generates informative responses
Multi-modal model for a wide range of tasks, including image understanding and language generation.
VISTA-3D is a specialized interactive foundation model for segmenting and anotating human anatomies.
Stable Video Diffusion (SVD) is a generative diffusion model that leverages a single image as a conditioning frame to synthesize video sequences.
A fast generative text-to-image model that can synthesize photorealistic images from a text prompt in a single network evaluation