NVIDIA
Explore
Models
Blueprints
GPUs
Docs
⌘KCtrl+K
View All Playbooks
View All Playbooks

onboarding

  • Set Up Local Network Access
  • Open WebUI with Ollama

data science

  • Single-cell RNA Sequencing
  • Portfolio Optimization
  • CUDA-X Data Science
  • Text to Knowledge Graph
  • Optimized JAX

tools

  • DGX Dashboard
  • Comfy UI
  • Connect Three DGX Spark in a Ring Topology
  • Connect Multiple DGX Spark through a Switch
  • RAG Application in AI Workbench
  • Set up Tailscale on Your Spark
  • VS Code

fine tuning

  • FLUX.1 Dreambooth LoRA Fine-tuning
  • LLaMA Factory
  • Fine-tune with NeMo
  • Fine-tune with Pytorch
  • Unsloth on DGX Spark

use case

  • NemoClaw with Nemotron 3 Super and Telegram on DGX Spark
  • Secure Long Running AI Agents with OpenShell on DGX Spark
  • OpenClaw 🦞
  • Live VLM WebUI
  • Install and Use Isaac Sim and Isaac Lab
  • Vibe Coding in VS Code
  • Build and Deploy a Multi-Agent Chatbot
  • Connect Two Sparks
  • NCCL for Two Sparks
  • Build a Video Search and Summarization (VSS) Agent
  • Spark & Reachy Photo Booth

inference

  • Speculative Decoding
  • Run models with llama.cpp on DGX Spark
  • vLLM for Inference
  • Nemotron-3-Nano with llama.cpp
  • SGLang for Inference
  • TRT LLM for Inference
  • NVFP4 Quantization
  • Multi-modal Inference
  • NIM on Spark
  • LM Studio on DGX Spark

Open WebUI with Ollama

15 MIN

Install Open WebUI and use Ollama to chat with models on your Spark

DGXSpark
OverviewOverviewSet up Open WebUI on Remote Spark with NVIDIA SyncSet up Open WebUI on Remote Spark with NVIDIA SyncSet Up ManuallySet Up ManuallyTroubleshootingTroubleshooting

Basic idea

Open WebUI is an extensible, self-hosted AI interface that operates entirely offline. This playbook shows you how to deploy Open WebUI with an integrated Ollama server on your DGX Spark device that lets you access the web interface from your local browser while the models run on Spark's GPU.

What you'll accomplish

You will have a fully functional Open WebUI installation running on your DGX Spark. This will be accessible through your local web browser either via NVIDIA Sync's managed SSH tunneling (recommended) or via manual setup. The setup includes integrated Ollama for model management, persistent data storage, and GPU acceleration for model inference.

What to know before starting

  • How to Set Up Local Network Access to your DGX Spark device

Prerequisites

  • DGX Spark device is set up and accessible
  • Local Network Access to your DGX Spark
  • Enough disk space for the Open WebUI container image and model downloads

Time & risk

  • Duration: 15-20 minutes for initial setup, plus model download time (varies by model size)
  • Risks:
    • Docker permission issues may require user group changes and session restart
    • Large model downloads may take significant time depending on network speed
  • Last Updated: 10/28/2025
    • Minor copyedits

Resources

  • Open WebUI Documentation
  • DGX Spark Documentation
  • DGX Spark Developer Forum
  • DGX Spark User Performance Guide
Terms of Use
Privacy Policy
Your Privacy Choices
Contact

Copyright © 2026 NVIDIA Corporation