Explore
Models
Blueprints
GPUs
Docs
⌘K
Ctrl+K
?
Login
Models
Deploy and scale models on your GPU infrastructure of choice with NVIDIA NIM inference microservices
Optimized by NVIDIA
Launch from Hugging Face
Beta
Filters
2 models
Sort By
dateCreated:DESC
Most Recent
Microsoft
phi-4-mini-flash-reasoning
Lightweight reasoning model for applications in latency bound, memory/compute constrained environments
edge
+4
413K
7mo
NVIDIA
llama-3.1-nemotron-nano-4b-v1.1
State-of-the-art open model for reasoning, code, math, and tool calling - suitable for edge agents
edge
+4
98.6K
8mo
Items per page
24
1
1
of 1 pages