NVIDIA
Explore
Models
Blueprints
GPUs
Docs
⌘KCtrl+K
Terms of Use
Privacy Policy
Your Privacy Choices
Contact

Copyright © 2025 NVIDIA Corporation

meta

llama-3.1-8b-instruct

Run Anywhere

Advanced state-of-the-art model with language understanding, superior reasoning, and text generation.

meta

llama-3.1-8b-instruct

Run Anywhere

Advanced state-of-the-art model with language understanding, superior reasoning, and text generation.

ChatLanguage GenerationRun-on-RTXText-to-TextCode Generation

The following are key system requirements and supported features to consider when self-hosting the llama-3.1-8b-instruct model.

GPU Memory Requirements

PrecisionMinimum GPU MemoryRecommended GPU Memory
bf1616 GB33 GB
fp88 GB16 GB

Deploying this NIM with less than the recommended amount of GPU memory requires setting the environment variable NIM_RELAX_MEM_CONSTRAINTS=1

Feature Support

FeatureSupported
LoRA Customization✅
Fine-tuning Customization✅
Tool Calling✅
TensorRT-LLM Local Engine Building✅

Links

  • Documentation
  • NVIDIA NGC Catalog