Explore
Models
Blueprints
GPUs
Docs
⌘K
Ctrl+K
?
Login
5 results for
Filters (1)
Models (0)
Blueprints (0)
Other (5)
Sort By
score:DESC
Best Match
DGX Station
20 MIN
Serve Qwen3-235B with vLLM
Set up vLLM server with Qwen3-235B on DGX Station
Playbook
vLLM
+1
1mo
Items per page
24
1
1
of 1 pages
DGX Spark
30 MIN
LM Studio on DGX Spark
Deploy LM Studio and serve LLMs on a Spark device; use LM Link to access models remotely.
Playbook
Inference
+3
2mo
DGX Spark
30 MIN
Nemotron-3-Nano with llama.cpp
Run Nemotron-3-Nano-30B model using llama.cpp on DGX Spark
Playbook
Nemotron
+3
4mo
DGX Spark
30 MIN
Run models with llama.cpp on DGX Spark
Build llama.cpp with CUDA and serve models via an OpenAI-compatible API (Nemotron 3 Nano Omni as example)
Playbook
DGX Spark
+3
3w
DGX Spark
60 MIN
cuTile Kernels
Run cuTile kernel benchmarks, FMHA implementation, and LLM inference on DGX Spark and B300
Playbook
FMHA
+10
1d