Use Ollama on DGX Spark to run a local coding model and connect a CLI coding agent. This
playbook supports three options: Claude Code, OpenCode, and Codex CLI. Each
agent is wired up with Ollama's built-in launch method (ollama launch <agent>), so you
can work without environment variables, provider config files, or external cloud APIs.
Pick the tab that matches the CLI agent you want to use:
You will run a local coding model (Qwen3.6) on your DGX Spark with Ollama, launch your chosen CLI agent against it with a single command, and complete a small coding task end-to-end.
ollama launch)qwen3.6:latest (35B-a3b, MoE) — ~24GB, 256K contextqwen3.6:35b-a3b-nvfp4 — ~22GB, NVIDIA FP4 build tuned for Blackwell (DGX Spark)qwen3.6:35b-a3b-q8_0 — ~39GB, higher-quality quantqwen3.6:35b-a3b-bf16 — ~71GB, full precision (fits Spark's unified memory)ollama launch~/.ollama/modelsollama launch method and upgraded the default model to Qwen3.6