NVIDIA
Explore
Models
Blueprints
GPUs
Docs
Terms of Use
Privacy Policy
Your Privacy Choices
Contact

Copyright © 2025 NVIDIA Corporation

View All Playbooks
View All Playbooks

onboarding

  • Set Up Local Network Access
  • Open WebUI with Ollama

data-science

  • Optimized JAX
  • Text to Knowledge Graph

tools

  • Comfy UI
  • DGX Dashboard
  • VS Code
  • RAG application in AI Workbench
  • Set up Tailscale on your Spark

fine-tuning

  • FLUX.1 Dreambooth LoRA Fine-tuning
  • LLaMA Factory
  • Fine-tune with NeMo
  • Fine tune with Pytorch
  • Unsloth on DGX Spark
  • Vision-Language Model Fine-tuning

use-case

  • Build and Deploy a Multi-Agent Chatbot
  • NCCL for Two Sparks
  • Connect Two Sparks
  • Video Search and Summarization

inference

  • Multi-modal Inference
  • NIM on Spark
  • NVFP4 Quantization
  • Speculative Decoding
  • TRT LLM for Inference
  • Install and Use vLLM for Inference

Video Search and Summarization

1 HR

Run the VSS Blueprint on your Spark

View on GitHub

Basic idea

Deploy NVIDIA's Video Search and Summarization (VSS) AI Blueprint to build intelligent video analytics systems that combine vision language models, large language models, and retrieval-augmented generation. The system transforms raw video content into real-time actionable insights with video summarization, Q&A, and real-time alerts. You'll set up either a completely local Event Reviewer deployment or a hybrid deployment using remote model endpoints.

What you'll accomplish

You will deploy NVIDIA's VSS AI Blueprint on NVIDIA Spark hardware with Blackwell architecture, choosing between two deployment scenarios: VSS Event Reviewer (completely local with VLM pipeline) or Standard VSS (hybrid deployment with remote LLM/embedding endpoints). This includes setting up Alert Bridge, VLM Pipeline, Alert Inspector UI, Video Storage Toolkit, and optional DeepStream CV pipeline for automated video analysis and event review.

What to know before starting

  • Working with NVIDIA Docker containers and container registries
  • Setting up Docker Compose environments with shared networks
  • Managing environment variables and authentication tokens
  • Basic understanding of video processing and analysis workflows

Prerequisites

  • NVIDIA Spark device with ARM64 architecture and Blackwell GPU
  • FastOS 1.81.38 or compatible ARM64 system
  • Driver version 580.82.09 or higher installed: nvidia-smi | grep "Driver Version"
  • CUDA version 13.0 installed: nvcc --version
  • Docker installed and running: docker --version && docker compose version
  • Access to NVIDIA Container Registry with NGC API Key
  • [Optional] NVIDIA API Key for remote model endpoints (hybrid deployment only)
  • Sufficient storage space for video processing (>10GB recommended in /tmp/)

Ancillary files

  • VSS Blueprint GitHub Repository - Main codebase and Docker Compose configurations
  • Sample CV Detection Pipeline - Reference CV pipeline for event reviewer workflow
  • VSS Official Documentation - Complete system documentation

Time & risk

  • Duration: 30-45 minutes for initial setup, additional time for video processing validation
  • Risks:
    • Container startup can be resource-intensive and time-consuming with large model downloads
    • Network configuration conflicts if shared network already exists
    • Remote API endpoints may have rate limits or connectivity issues (hybrid deployment)
  • Rollback: Stop all containers with docker compose down, remove shared network with docker network rm vss-shared-network, and clean up temporary media directories.

Resources

  • DGX Spark Documentation
  • DGX Spark Forum