LM Studio on DGX Spark
Deploy LM Studio and serve LLMs on a Spark device
Install llmster on the DGX Spark
llmster is LM Studio's terminal native, headless LM Studio ‘daemon’.
You can install it on servers, cloud instances, machines with no GUI, or just on your computer. This is useful for running LM Studio in headless mode on DGX Spark, then connecting to it from your laptop via the API.
On your Spark, install llmster by running:
curl -fsSL https://lmstudio.ai/install.sh | bash
For Windows:
irm https://lmstudio.ai/install.ps1 | iex
Once installed, follow the instructions in your terminal output to add lms to your PATH. Interact with LM Studio using the lms CLI or the SDK / LM Studio V1 REST API (new with enhanced features) / OpenAI-compatible REST API.
Download Required Ancillary Files
Run the following curl commands in your local terminal to download files required to complete later steps in this playbook. You may choose from Python, JavaScript, or Bash.
# JavaScript
curl -L -O https://raw.githubusercontent.com/lmstudio-ai/docs/main/_assets/nvidia-spark-playbook/js/run.js
# Python
curl -L -O https://raw.githubusercontent.com/lmstudio-ai/docs/main/_assets/nvidia-spark-playbook/py/run.py
# Bash
curl -L -O https://raw.githubusercontent.com/lmstudio-ai/docs/main/_assets/nvidia-spark-playbook/bash/run.sh
Start the LM Studio API Server
Use lms, LM Studio's CLI to start the server from your terminal. Enable local network access, which allows the LM Studio API server running on your machine to be accessed by all other devices on the same local network (make sure they are trusted devices). To do this, run the following command:
lms server start --bind 0.0.0.0 --port 1234
Test the connectivity between your laptop and your Spark, run the following command in your local terminal
curl http://<SPARK_IP>:1234/api/v1/models
where <SPARK_IP> is your device's IP address." You can find your Spark’s IP address by running this on your Spark:
hostname -I
Download a model to your Spark
As an example, let's download and run gpt-oss 120B, one of the best open source models from OpenAI. This model is too large for many laptops due to memory limitations, which makes this a fantastic use case for the Spark.
lms get openai/gpt-oss-120b
This download will take a while due to its large size. Verify that the model has been successfully downloaded by listing your models:
lms ls
Load the model
Load the model on your Spark so that it is ready to respond to requests from your laptop.
lms load openai/gpt-oss-120b
Set up a simple program that uses LM Studio SDK on the laptop
Install the LM Studio SDKs and use a simple script to send a prompt to your Spark and validate the response. To get started quickly, we provide simple scripts below for Python, Javascript, and Bash. Download the scripts from the Overview page of this playbook and run the corresponding command from the directory containing it.
NOTE
Within each script, replace <SPARK_IP> with the IP address of your DGX Spark on your local network.
Javascript
Pre-reqs: User has installed npm and node
npm install @lmstudio/sdk
node run.js
Python
Pre-reqs: User has installed uv
uv run --script run.py
Bash
Pre-reqs: User has installed jq and curl
bash run.sh
Next Steps
Try downloading and serving different models from the LM Studio model catalog
Cleanup and rollback
Remove and uninstall LM Studio completely if needed. Note that LM Studio stores models separately from the application. Uninstalling LM Studio will not remove downloaded models unless you explicitly delete them.
If you want to remove the entire LM Studio application, quit LM Studio from the tray first then move the application to trash.
To uninstall llmster, remove the folder ~/.lmstudio/llmster.
To remove downloaded models, delete the contents of ~/.lmstudio/models/.