meta/llama-3.2-90b-vision-instruct

RUN ANYWHERE

Cutting-edge vision-Language model exceling in high-quality reasoning from images.

By running the below commands, you accept the NVIDIA AI Enterprise Terms of Use and the NVIDIA Community Models License.

Pull and run meta/llama-3.2-90b-vision-instruct using Docker (this will download the full model and run it in your local environment)

$ docker login nvcr.io Username: $oauthtoken Password: <PASTE_API_KEY_HERE>

Pull and run the NVIDIA NIM with the command below. This will download the optimized model for your infrastructure.

export NGC_API_KEY=<PASTE_API_KEY_HERE> export LOCAL_NIM_CACHE=~/.cache/nim mkdir -p "$LOCAL_NIM_CACHE" docker run -it --rm \ --gpus all \ --shm-size=16GB \ -e NGC_API_KEY \ -v "$LOCAL_NIM_CACHE:/opt/nim/.cache" \ -u $(id -u) \ -p 8000:8000 \ nvcr.io/nim/meta/llama-3.2-90b-vision-instruct:latest

You can now make a local API call using this curl command:

curl -X 'POST' \ 'http://0.0.0.0:8000/v1/chat/completions' \ -H 'accept: application/json' \ -H 'Content-Type: application/json' \ -d '{ "model": "meta/llama-3.2-90b-vision-instruct", "messages": [{"role":"user", "content":[ {"type": "text", "text": "Describe this image"}, { "type": "image_url", "image_url": {"url": "https://assets.ngc.nvidia.com/products/api-catalog/phi-3-5-vision/example1b.jpg"} } ]}], "max_tokens": 256 }'

For more details on getting started with this NIM, visit the NVIDIA VLM NIM Docs.