Transform unstructured text into interactive knowledge graphs with LLM inference and graph visualization
This playbook is for DGX Station. In a terminal, clone the repository and navigate to the project directory.
git clone https://github.com/NVIDIA/dgx-station-playbooks
cd dgx-station-playbooks/nvidia/station-txt2kg/assets
The default backend is vLLM (supported on DGX Station). The script starts services and waits for the vLLM backend to be ready (model load can take 30+ minutes; progress is shown in the terminal). To use Ollama instead, run ./start.sh --ollama.
./start.sh
# Optional: ./start.sh --ollama # Use ArangoDB + Ollama instead of vLLM
# Optional: ./start.sh --no-wait # Skip waiting for vLLM readiness
The script will:
If you started with Ollama (./start.sh --ollama), pull the Llama model:
docker exec ollama-compose ollama pull llama3.1:405b
Browse available models at https://ollama.com/search. With the default vLLM stack, the model is loaded automatically by the vLLM container.
Open your browser and navigate to:
http://localhost:3001
You can also access:
The web UI defaults to local (vLLM or Ollama). If the backend is still loading, a banner and the model selector will show “Initializing…” until the backend is ready.
Future Enhancement: GraphRAG capabilities with vector-based KNN search for entity retrieval are planned.
Stop all services (use the same flags as when you started):
# Stop services (default: vLLM stack)
./stop.sh
# If you started with Ollama: ./stop.sh --ollama
# Remove containers and volumes (optional)
# From assets dir: docker compose -f deploy/compose/docker-compose.vllm.yml down -v
# Or with Ollama: docker compose -f deploy/compose/docker-compose.yml down -v
# Remove downloaded Ollama models (Ollama only)
# docker exec ollama-compose ollama rm llama3.1:405b
./start.sh --ollama for ArangoDB + Ollama.