NVIDIA
Explore
Models
Blueprints
GPUs
Docs
⌘KCtrl+K
View All Playbooks
View All Playbooks

onboarding

  • Set Up Local Network Access
  • Open WebUI with Ollama

data science

  • Single-cell RNA Sequencing
  • Portfolio Optimization
  • CUDA-X Data Science
  • Text to Knowledge Graph
  • Optimized JAX

tools

  • DGX Dashboard
  • Comfy UI
  • RAG Application in AI Workbench
  • Set up Tailscale on Your Spark
  • VS Code
  • Connect Three DGX Spark in a Ring Topology
  • Connect Multiple DGX Spark through a Switch

fine tuning

  • FLUX.1 Dreambooth LoRA Fine-tuning
  • LLaMA Factory
  • Fine-tune with NeMo
  • Fine-tune with Pytorch
  • Unsloth on DGX Spark

use case

  • NemoClaw with Nemotron 3 Super and Telegram on DGX Spark
  • cuTile Kernels
  • CLI Coding Agent
  • Live VLM WebUI
  • Install and Use Isaac Sim and Isaac Lab
  • Vibe Coding in VS Code
  • Build and Deploy a Multi-Agent Chatbot
  • Connect Two Sparks
  • NCCL for Two Sparks
  • Build a Video Search and Summarization (VSS) Agent
  • Spark & Reachy Photo Booth
  • Secure Long Running AI Agents with OpenShell on DGX Spark
  • OpenClaw 🦞

inference

  • LM Studio on DGX Spark
  • Speculative Decoding
  • Run models with llama.cpp on DGX Spark
  • Nemotron-3-Nano with llama.cpp
  • SGLang for Inference
  • TRT LLM for Inference
  • NVFP4 Quantization
  • Multi-modal Inference
  • NIM on Spark
  • vLLM for Inference

CLI Coding Agent

20 MINS

Build local CLI coding agents with Ollama

Claude CodeCodexCodingLLMOllamaOpenCodeQwen
OverviewOverviewClaude CodeClaude CodeOpenCodeOpenCodeCodex CLICodex CLITroubleshootingTroubleshooting
SymptomCauseFix
ollama: command not foundOllama not installed or PATH not updatedRerun curl -fsSL https://ollama.com/install.sh | sh and open a new shell
ollama launch reports unknown commandOllama is older than v0.15Update Ollama: curl -fsSL https://ollama.com/install.sh | sh
Model load fails with version error or HTTP 412Ollama version is too old for the modelUpdate Ollama: curl -fsSL https://ollama.com/install.sh | sh
model not found when launching an agentModel was not pulledRun ollama pull qwen3.6 and retry
connection refused to localhost:11434Ollama service not runningStart with ollama serve or sudo systemctl start ollama
ollama launch <agent> exits immediatelyAgent integration failed to initializeRe-run ollama launch <agent>; if it persists, check journalctl -u ollama
Slow responses or OOM errorsModel variant too large for GPU memorySwitch to qwen3.6:35b-a3b-nvfp4 or close other GPU workloads

NOTE

DGX Spark uses a Unified Memory Architecture (UMA), which enables dynamic memory sharing between the GPU and CPU. If you see memory pressure, flush the buffer cache with:

sudo sh -c 'sync; echo 3 > /proc/sys/vm/drop_caches'

Resources

  • Ollama Documentation
  • Ollama Launch Method
  • Qwen3.6 Model Page
  • Claude Code + Ollama Guide
  • OpenCode Ollama Provider
  • Codex + Ollama Guide
  • DGX Spark Documentation
  • DGX Spark Forum
Terms of Use
Privacy Policy
Your Privacy Choices
Contact

Copyright © 2026 NVIDIA Corporation