Install Open WebUI and use Ollama to chat with models on your Spark
Open WebUI is an extensible, self-hosted AI interface that operates entirely offline. This playbook shows you how to deploy Open WebUI with an integrated Ollama server on your DGX Spark device using NVIDIA Sync. The setup creates a secure SSH tunnel that lets you access the web interface from your local browser while the models run on Spark's GPU.
You will have a fully functional Open WebUI installation running on your DGX Spark, accessible through your local web browser via NVIDIA Sync's managed SSH tunneling. The setup includes integrated Ollama for model management, persistent data storage, and GPU acceleration for model inference.