NVIDIA
Explore
Models
Blueprints
GPUs
Docs
Terms of Use
Privacy Policy
Your Privacy Choices
Contact

Copyright © 2025 NVIDIA Corporation

View All Playbooks
View All Playbooks

onboarding

  • Set Up Local Network Access
  • Open WebUI with Ollama

data-science

  • Optimized JAX
  • Text to Knowledge Graph

tools

  • Comfy UI
  • DGX Dashboard
  • VS Code
  • RAG application in AI Workbench
  • Set up Tailscale on your Spark

fine-tuning

  • FLUX.1 Dreambooth LoRA Fine-tuning
  • LLaMA Factory
  • Fine-tune with NeMo
  • Fine tune with Pytorch
  • Unsloth on DGX Spark
  • Vision-Language Model Fine-tuning

use-case

  • Build and Deploy a Multi-Agent Chatbot
  • NCCL for Two Sparks
  • Connect Two Sparks
  • Video Search and Summarization

inference

  • Multi-modal Inference
  • NIM on Spark
  • NVFP4 Quantization
  • Speculative Decoding
  • TRT LLM for Inference
  • Install and Use vLLM for Inference

Text to Knowledge Graph

30 MIN

Transform unstructured text into interactive knowledge graphs with LLM inference and graph visualization

View on GitHub

Basic idea

This playbook demonstrates how to build and deploy a comprehensive knowledge graph generation and visualization solution that serves as a reference for knowledge graph extraction. The unified memory architecture enables running larger, more accurate models that produce higher-quality knowledge graphs and deliver superior downstream GraphRAG performance.

This txt2kg playbook transforms unstructured text documents into structured knowledge graphs using:

  • Knowledge Triple Extraction: Using Ollama with GPU acceleration for local LLM inference to extract subject-predicate-object relationships
  • Graph Database Storage: ArangoDB for storing and querying knowledge triples with relationship traversal
  • GPU-Accelerated Visualization: Three.js WebGPU rendering for interactive 2D/3D graph exploration

Future Enhancements: Vector embeddings and GraphRAG capabilities are planned enhancements.

What you'll accomplish

You will have a fully functional system capable of processing documents, generating and editing knowledge graphs, and providing querying, accessible through an interactive web interface. The setup includes:

  • Local LLM Inference: Ollama for GPU-accelerated LLM inference with no API keys required
  • Graph Database: ArangoDB for storing and querying triples with relationship traversal
  • Interactive Visualization: GPU-accelerated graph rendering with Three.js WebGPU
  • Modern Web Interface: Next.js frontend with document management and query interface
  • Fully Containerized: Reproducible deployment with Docker Compose and GPU support

Prerequisites

  • DGX Spark with latest NVIDIA drivers
  • Docker installed and configured with NVIDIA Container Toolkit
  • Docker Compose

Time & risk

  • Duration:

    • 2-3 minutes for initial setup and container deployment
    • 5-10 minutes for Ollama model download (depending on model size)
    • Immediate document processing and knowledge graph generation
  • Risks:

    • GPU memory requirements depend on chosen Ollama model size
    • Document processing time scales with document size and complexity
  • Rollback: Stop and remove Docker containers, delete downloaded models if needed

Resources

  • DGX Spark Documentation
  • Ollama Documentation
  • ArangoDB Documentation
  • DGX Spark Forum