NCCL for Two Sparks

30 MIN

Install and test NCCL on two Sparks

Basic idea

NCCL (NVIDIA Collective Communication Library) enables high-performance GPU-to-GPU communication across multiple nodes. This walkthrough sets up NCCL for multi-node distributed training on DGX Spark systems with Blackwell architecture. You'll configure networking, build NCCL from source with Blackwell support, and validate communication between nodes.

What you'll accomplish

You'll have a working multi-node NCCL environment that enables high-bandwidth GPU communication across DGX Spark systems for distributed training workloads, with validated network performance and proper GPU topology detection.

What to know before starting

  • Working with Linux network configuration and netplan
  • Basic understanding of MPI (Message Passing Interface) concepts
  • SSH key management and passwordless authentication setup

Prerequisites

  • Two DGX Spark systems
  • Completed the Connect two Sparks playbook
  • NVIDIA driver installed: nvidia-smi
  • CUDA toolkit available: nvcc --version
  • Root/sudo privileges: sudo whoami

Time & risk

  • Duration: 30 minutes for setup and validation
  • Risk level: Medium - involves network configuration changes
  • Rollback: The NCCL & NCCL Tests repositories can be deleted from DGX Spark