Deploy and scale models on your GPU infrastructure of choice with NVIDIA NIM inference microservices

Next-gen Qwen 3.5 VLM (400B MoE) brings advanced vision, chat, RAG, and agentic capabilities.

Vision language model that excels in understanding the physical world using structured reasoning on videos or images.

Cutting-edge vision-language model exceling in retrieving text and metadata from images.

Nemotron Nano 12B v2 VL enables multi-image and video understanding, along with visual Q&A and summarization capabilities.

Reasoning vision language model (VLM) for physical AI and robotics.

Multi-modal vision-language model that understands text/img and creates informative responses

A general purpose multimodal, multilingual 128 MoE model with 17B parameters.

A multimodal, multilingual 16 MoE model with 17B parameters.

Cutting-edge open multimodal model exceling in high-quality reasoning from images.

Cutting-edge vision-language model exceling in retrieving text and metadata from images.

Multi-modal vision-language model that understands text/img/video and creates informative responses

Advanced AI model detects faces and identifies deep fake images.

Cutting-edge vision-language model exceling in high-quality reasoning from images.

Cutting-edge vision-Language model exceling in high-quality reasoning from images.

Robust image classification model for detecting and managing AI-generated content.

Cutting-edge open multimodal model exceling in high-quality reasoning from images.

Grounding dino is an open vocabulary zero-shot object detection model.


Visual Changenet detects pixel-level change maps between two images and outputs a semantic change segmentation mask

EfficientDet-based object detection network to detect 100 specific retail objects from an input video.