Fine tune with Pytorch
1 HR
Use Pytorch to fine-tune models locally
Basic idea
This playbook guides you through setting up and using Pytorch for fine-tuning large language models on NVIDIA Spark devices.
What you'll accomplish
You'll establish a complete fine-tuning environment for large language models (1-70B parameters) on your NVIDIA Spark device. By the end, you'll have a working installation that supports parameter-efficient fine-tuning (PEFT) and supervised fine-tuning (SFT).
What to know before starting
- Previous experience with fine-tuning in Pytorch
- Working with Docker
Prerequisites
Recipes are specifically for DIGITS SPARK. Please make sure that OS and drivers are latest.
Ancillary files
ALl files required for fine-tuning are included in the folder in the GitHub repository here.
Time & risk
- Time estimate: 30-45 mins for setup and runing fine-tuning. Fine-tuning run time varies depending on model size
- Risks: Model downloads can be large (several GB), ARM64 package compatibility issues may require troubleshooting.