NVIDIA
Explore
Models
Blueprints
GPUs
Docs
View All Playbooks
View All Playbooks

onboarding

  • Set Up Local Network Access
  • Open WebUI with Ollama

data science

  • CUDA-X Data Science
  • Optimized JAX
  • Text to Knowledge Graph

tools

  • Comfy UI
  • DGX Dashboard
  • VS Code
  • RAG application in AI Workbench
  • Set up Tailscale on your Spark

fine tuning

  • FLUX.1 Dreambooth LoRA Fine-tuning
  • LLaMA Factory
  • Fine-tune with NeMo
  • Fine tune with Pytorch
  • Unsloth on DGX Spark
  • Vision-Language Model Fine-tuning

use case

  • Vibe Coding in VS Code
  • Build and Deploy a Multi-Agent Chatbot
  • NCCL for Two Sparks
  • Connect Two Sparks
  • Video Search and Summarization

inference

  • Multi-modal Inference
  • NIM on Spark
  • NVFP4 Quantization
  • Speculative Decoding
  • TRT LLM for Inference
  • Install and Use vLLM for Inference
Terms of Use
Privacy Policy
Your Privacy Choices
Contact

Copyright © 2025 NVIDIA Corporation

Vibe Coding in VS Code

30 MIN

Use DGX Spark as a local or remote Vibe Coding assistant with Ollama and Continue

Step 1
Install Ollama

Install the latest version of Ollama using the following command:

curl -fsSL https://ollama.com/install.sh | sh

Once the service is running, pull the desired model:

ollama pull gpt-oss:120b

Step 2
(Optional) Enable Remote Access

To allow remote connections (e.g., from a workstation using VSCode and Continue), modify the Ollama systemd service:

sudo systemctl edit ollama

Add the following lines beneath the commented section:

[Service]
Environment="OLLAMA_HOST=0.0.0.0:11434"
Environment="OLLAMA_ORIGINS=*"

Reload and restart the service:

sudo systemctl daemon-reload
sudo systemctl restart ollama

If using a firewall, open port 11434:

sudo ufw allow 11434/tcp

Verify that the workstation can connect to your DGX Spark's Ollama server:

curl -v http://YOUR_SPARK_IP:11434/api/version

Replace YOUR_SPARK_IP with your DGX Spark's IP address. If the connection fails please see the Troubleshooting tab.

Step 3
Install VSCode

For DGX Spark (ARM-based), download and install VSCode: Navigate to https://code.visualstudio.com/download and download the Linux ARM64 version of VSCode. After the download completes note the downloaded package name. Use it in the next command in place of DOWNLOADED_PACKAGE_NAME.

sudo dpkg -i DOWNLOADED_PACKAGE_NAME

If using a remote workstation, install VSCode appropriate for your system architecture.

Step 4
Install Continue.dev Extension

Open VSCode and install Continue.dev from the Marketplace:

  • Go to Extensions view in VSCode
  • Search for Continue published by Continue.dev and install the extension. After installation, click the Continue icon on the right-hand bar.

Step 5
Local Inference Setup

  • Click Or, configure your own models
  • Click Click here to view more providers
  • Choose Ollama as the Provider
  • For Model, select Autodetect
  • Test inference by sending a test prompt.

Your downloaded model will now be the default (e.g., gpt-oss:120b) for inference.

Step 6
Setting up a Workstation to Connect to the DGX Spark' Ollama Server

To connect a workstation running VSCode to a remote DGX Spark instance the following must be completed on that workstation:

  • Install Continue as instructed in Step 4
  • Click on the Continue icon on the left pane
  • Click Or, configure your own models
  • Click Click here to view more providers
  • Select Ollama as the Provider
  • Select Autodetect as the Model.

Continue will fail to detect the model as it is attempting to connect to a locally hosted Ollama server.

  • Find the gear icon in the upper right corner of the Continue window and click on it.
  • On the left pane, click Models
  • Next to the first dropdown menu under Chat click the gear icon.
  • Continue's config.yaml will open. Take note of your DGX Spark's IP address.
  • Replace the configuration with the following. YOUR_SPARK_IP should be replaced with your DGX Spark's IP.
name: Config
version: 1.0.0
schema: v1

assistants:
  - name: default
    model: OllamaSpark

models:
  - name: OllamaSpark
    provider: ollama
    model: gpt-oss:120b
    apiBase: http://YOUR_SPARK_IP:11434
    title: gpt-oss:120b
    roles:
      - chat
      - edit
      - autocomplete

Replace YOUR_SPARK_IP with the IP address of your DGX Spark.
Add additional model entries for any other Ollama models you wish to host remotely.

Resources

  • DGX Spark Documentation
  • Ollama Documentation
  • VSCode
  • Continue.dev
  • DGX Spark Forum