Explore
Models
Blueprints
GPUs
Docs
⌘K
Ctrl+K
?
Login
26 results for
Filters
Models (26)
Blueprints (0)
Other (0)
Sort By
score:DESC
Best Match
DeepSeek AI
deepseek-r1-distill-qwen-7b
Distilled version of Qwen 2.5 7B using reasoning data generated by DeepSeek R1 for enhanced performance.
Model
coding
+3
9mo
OpenAI
gpt-oss-20b
Smaller Mixture of Experts (MoE) text-only LLM for efficient AI reasoning and math
Model
text-to-text
+3
7mo
Marin
marin-8b-instruct
State-of-the-art open model trained on open datasets, excelling in reasoning, math, and science.
Model
Reasoning
+4
9mo
NVIDIA
llama-3.1-nemotron-nano-4b-v1.1
State-of-the-art open model for reasoning, code, math, and tool calling - suitable for edge agents
Model
edge
+4
8mo
NVIDIA
llama-3.1-nemotron-ultra-253b-v1
Superior inference efficiency with highest accuracy for scientific and complex math reasoning, coding, tool calling, and instruction following.
Model
chat
+4
7mo
Sarvamai
sarvam-m
Multilingual, hybrid-reasoning model optimized for Indian language tasks, programming, mathematical reasoning capabilities.
Model
coding
+6
7mo
DeepSeek AI
deepseek-r1-distill-qwen-14b
Distilled version of Qwen 2.5 14B using reasoning data generated by DeepSeek R1 for enhanced performance.
Model
coding
+4
9mo
DeepSeek AI
deepseek-r1-distill-qwen-32b
Distilled version of Qwen 2.5 32B using reasoning data generated by DeepSeek R1 for enhanced performance.
Model
coding
+4
9mo
OpenAI
gpt-oss-120b
Mixture of Experts (MoE) reasoning LLM (text-only) designed to fit within 80GB GPU.
Model
text-to-text
+3
7mo
NVIDIA
llama-3.1-nemotron-nano-8b-v1
Leading reasoning and agentic AI accuracy model for PC and edge.
Model
chat
+4
8mo
NVIDIA
llama-3.3-nemotron-super-49b-v1
High efficiency model with leading accuracy for reasoning, tool calling, chat, and instruction following.
Model
chat
+4
7mo
NVIDIA
llama-3.3-nemotron-super-49b-v1.5
High efficiency model with leading accuracy for reasoning, tool calling, chat, and instruction following.
Model
chat
+4
7mo
Mistral AI
magistral-small-2506
High performance reasoning model optimized for efficiency and edge deployment
Model
coding
+4
7mo
Microsoft
phi-4-mini-flash-reasoning
Lightweight reasoning model for applications in latency bound, memory/compute constrained environments
Model
edge
+4
7mo
Qwen
qwen3-235b-a22b
Advanced reasoing MOE mode excelling at reasoning, multilingual tasks, and instruction following
Model
chat
+3
7mo
Qwen
qwq-32b
Powerful reasoning model capable of thinking and reasoning, can achieve significantly enhanced performance in downstream tasks, especially hard problems.
Model
coding
+3
8mo
Meta
llama-3.3-70b-instruct
Advanced LLM for reasoning, math, general knowledge, and function calling
Model
Reasoning
+5
8mo
Tiiuae
falcon3-7b-instruct
Instruction tuned LLM achieving SoTA performance on reasoning, math and general knowledge capabilities
Model
Coding
+6
9mo
DeepSeek AI
deepseek-r1-distill-llama-8b
Distilled version of Llama 3.1 8B using reasoning data generated by DeepSeek R1 for enhanced performance.
Model
Distillation
+5
7mo
Baichuan AI
baichuan2-13b-chat
Support Chinese and English chat, coding, math, instruction following, solving quizzes
Model
Chinese Language Generation
+3
9mo
Mistral AI
mistral-small-24b-instruct
Latency-optimized language model excelling in code, math, general knowledge, and instruction-following.
Model
code
+4
8mo
Microsoft
phi-3-mini-128k-instruct
Lightweight, state-of-the-art open LLM with strong math and logical reasoning skills.
Model
chat
+4
9mo
Microsoft
phi-3-mini-4k-instruct
Lightweight, state-of-the-art open LLM with strong math and logical reasoning skills.
Model
chat
+4
9mo
Qwen
qwen2-7b-instruct
Chinese and English LLM targeting for language, coding, mathematics, reasoning, etc.
Model
Chinese Language Generation
+3
9mo
Items per page
24
1
1
2
2
of 2 pages