NVIDIA
Explore
Models
Blueprints
GPUs
Docs
⌘KCtrl+K
View All Playbooks
View All Playbooks

onboarding

  • Set Up Local Network Access
  • Open WebUI with Ollama

data science

  • Single-cell RNA Sequencing
  • Portfolio Optimization
  • CUDA-X Data Science
  • Text to Knowledge Graph
  • Optimized JAX

tools

  • DGX Dashboard
  • Comfy UI
  • RAG Application in AI Workbench
  • Set up Tailscale on Your Spark
  • VS Code
  • Connect Three DGX Spark in a Ring Topology
  • Connect Multiple DGX Spark through a Switch

fine tuning

  • FLUX.1 Dreambooth LoRA Fine-tuning
  • LLaMA Factory
  • Fine-tune with NeMo
  • Fine-tune with Pytorch
  • Unsloth on DGX Spark

use case

  • NemoClaw with Nemotron 3 Super and Telegram on DGX Spark
  • Run Hermes Agent with Local Models
  • cuTile Kernels
  • CLI Coding Agent
  • Live VLM WebUI
  • Install and Use Isaac Sim and Isaac Lab
  • Vibe Coding in VS Code
  • Build and Deploy a Multi-Agent Chatbot
  • Connect Two Sparks
  • NCCL for Two Sparks
  • Build a Video Search and Summarization (VSS) Agent
  • Spark & Reachy Photo Booth
  • Secure Long Running AI Agents with OpenShell on DGX Spark
  • OpenClaw 🦞

inference

  • Speculative Decoding
  • Run models with llama.cpp on DGX Spark
  • Nemotron-3-Nano with llama.cpp
  • SGLang for Inference
  • TRT LLM for Inference
  • NVFP4 Quantization
  • Multi-modal Inference
  • NIM on Spark
  • LM Studio on DGX Spark
  • vLLM for Inference

Run Hermes Agent with Local Models

30 MIN

Install and run the Hermes self-improving AI agent on DGX Spark.

AI AgentLLMNous Research
View on GitHub
OverviewOverviewInstructionsInstructionsTroubleshootingTroubleshooting

Basic idea

Hermes Agent is a self-improving AI agent built by Nous Research. It runs as a terminal TUI on your machine and, through a built-in gateway, can also be reached from messaging platforms like Telegram, Discord, and Slack. It creates skills from experience, improves them during use, persists memory across sessions, and can run scheduled tasks via its built-in cron.

Running Hermes and its LLM fully on your DGX Spark keeps your conversations and data private and avoids ongoing cloud API costs. DGX Spark is well suited for this: it runs Linux, is designed to stay on, and has 128GB memory, so you can serve large local models for better reasoning quality and connect to the agent from your phone over Telegram while the heavy work runs locally.

What you'll accomplish

You will have Hermes installed on your DGX Spark and connected to a local LLM served by Ollama. You can chat with the agent from the DGX Spark terminal and from Telegram on your phone or laptop. The gateway runs as a system service, so the agent stays reachable across reboots without anyone logging in.

  • Install Ollama and pull a local model
  • Install Hermes and configure it against the local Ollama endpoint
  • Set up a Telegram bot so you can message Hermes from any Telegram client
  • Resume past sessions, switch models, update, and uninstall using the hermes CLI

Popular use cases

  • Personal assistant from your phone: Chat with Hermes via Telegram while the model runs on your Spark — manage email drafts, summarize docs, or answer questions on the go.
  • Multi-step task automation: Ask the agent to walk you through configurations (e.g., setting up email); on non-trivial tasks Hermes can autonomously persist a reusable skill for next time.
  • Scheduled checks: Use the built-in cron to watch a product price online or run a daily check, and have results delivered to your Telegram home channel.
  • Reasoning-visible problem solving: Use /reasoning show in the TUI to follow the agent's intermediate reasoning on complex problems.

What to know before starting

  • Basic use of the Linux terminal and a text editor
  • Familiarity with Ollama or willingness to follow the Ollama on Spark playbook first
  • A Telegram account if you want to use the messaging gateway
  • Awareness of the security considerations below

Important: security and risks

AI agents that can execute commands and reach external services introduce real risks. Read the upstream guidance, especially the dedicated security topics: Hermes Agent — Security.

Main risks:

  1. Data exposure: Personal information or files on your DGX Spark may be leaked through agent actions or messaging channels.
  2. Unauthorized access: A Telegram bot left open to anyone who finds it can be misused; a model endpoint exposed beyond localhost can be abused.

You cannot eliminate all risk; proceed at your own risk. Recommended security measures:

  • Restrict the Telegram bot by entering one or more numeric Telegram user IDs at the "Allowed user IDs" prompt during install. Leaving this blank allows anyone who finds the bot to use it.
  • Keep the Ollama endpoint bound to localhost only; do not expose http://<spark-ip>:11434 to your LAN or the public internet without strong authentication.
  • Run Hermes on a Spark dedicated to this purpose where possible, and only place files on it that the agent is allowed to access.
  • Monitor activity: Periodically review the gateway service logs (sudo journalctl -u <hermes-gateway-unit> -e) and the Hermes session history.

Prerequisites

  • DGX Spark running Linux, connected to your network
  • Terminal (SSH or local) access to the Spark
  • curl and git installed (verified in Step 1 of the instructions)
  • Interactive terminal access for the setup wizard and any sudo password prompts. Non-interactive SSH is supported with the config-command fallback in the Instructions tab.
  • Enough disk and GPU memory for the Ollama model you plan to serve (the playbook uses qwen3.6:27b as the example; pick a smaller model if you want a faster first install)
  • A Telegram account and the ability to create a bot via @BotFather if you plan to use the messaging gateway

Time and risk

  • Duration: About 30 minutes for install and first-time setup; model download time depends on size and network speed.
  • Risk level: Medium — the agent can execute commands, persist skills, and is reachable from Telegram. Risk increases if you skip the allowed-user-IDs restriction or expose the local model endpoint beyond localhost. Always follow the security measures above.
  • Rollback: Run hermes uninstall (with sudo if you installed the gateway as a system service) to remove Hermes, the gateway service, and the shell-profile entry. The data directory ~/.hermes may still be present afterward; remove it manually if you want a full reset (see the Cleanup and Troubleshooting tabs). Uninstall Ollama separately if desired.
  • Last Updated: 2026-05-08
    • First Publication

Resources

  • Hermes Agent Documentation
  • Hermes Agent — Security
  • Hermes Agent GitHub Repository
  • Nous Research
  • DGX Spark Documentation
  • DGX Spark Forum
Terms of Use
Privacy Policy
Your Privacy Choices
Contact

Copyright © 2026 NVIDIA Corporation