
A context‑aware safety model that applies reasoning to enforce domain‑specific policies.

Vision language model that excels in understanding the physical world using structured reasoning on videos or images.

State-of-the-art 685B reasoning LLM with sparse attention, long context, and integrated agentic tools.

Open, efficient MoE model with 1M context, excelling in coding, reasoning, instruction following, tool calling, and more

State-of-the-art open code model with deep reasoning, 256k context, and unmatched efficiency.

Open reasoning model with 256K context window, native INT4 quantization and enhanced tool use

Open Mixture of Experts LLM (230B, 10B active) for reasoning, coding, and tool-use/agent workflows

DeepSeek-V3.1: hybrid inference LLM with Think/Non-Think modes, stronger agents, 128K context, strict function calling.

Follow-on version of Kimi-K2-Instruct with longer context window and enhanced reasoning capabilities

80B parameter AI model with hybrid reasoning, MoE architecture, support for 119 languages.

ByteDance open-source LLM with long-context, reasoning, and agentic intelligence.

DeepSeek V3.1 Instruct is a hybrid AI model with fast reasoning, 128K context, and strong tool use.

High‑efficiency LLM with hybrid Transformer‑Mamba design, excelling in reasoning and agentic tasks.

Reasoning vision language model (VLM) for physical AI and robotics.

Smaller Mixture of Experts (MoE) text-only LLM for efficient AI reasoning and math

Mixture of Experts (MoE) reasoning LLM (text-only) designed to fit within 80GB GPU.

High efficiency model with leading accuracy for reasoning, tool calling, chat, and instruction following.

Advanced reasoing MOE mode excelling at reasoning, multilingual tasks, and instruction following

High efficiency model with leading accuracy for reasoning, tool calling, chat, and instruction following.

Lightweight reasoning model for applications in latency bound, memory/compute constrained environments

State-of-the-art open mixture-of-experts model with strong reasoning, coding, and agentic capabilities

An MOE LLM that follows instructions, completes requests, and generates creative text.

An MOE LLM that follows instructions, completes requests, and generates creative text.

State-of-the-art, high-efficiency LLM excelling in reasoning, math, and coding.

Powerful, multimodal language model designed for enterprise applications, including software development, data analysis, and reasoning.

High performance reasoning model optimized for efficiency and edge deployment

Superior inference efficiency with highest accuracy for scientific and complex math reasoning, coding, tool calling, and instruction following.

Small language model fine-tuned for improved reasoning, coding, and instruction-following

Advanced state-of-the-art model with language understanding, superior reasoning, and text generation.

Distilled version of Llama 3.1 8B using reasoning data generated by DeepSeek R1 for enhanced performance.

Advanced LLM for code generation, reasoning, and fixing across popular programming languages.

State-of-the-art open model for reasoning, code, math, and tool calling - suitable for edge agents

Latency-optimized language model excelling in code, math, general knowledge, and instruction-following.

Leading reasoning and agentic AI accuracy model for PC and edge.

Advanced LLM for reasoning, math, general knowledge, and function calling

Powers complex conversations with superior contextual understanding, reasoning and text generation.

Updated version of DeepSeek-R1 with enhanced reasoning, coding, math, and reduced hallucination.

Cutting-edge vision-Language model exceling in high-quality reasoning from images.

Cutting-edge vision-language model exceling in high-quality reasoning from images.

Cutting-edge open multimodal model exceling in high-quality reasoning from images.

Cutting-edge open multimodal model exceling in high-quality reasoning from image and audio inputs.

Lightweight, state-of-the-art open LLM with strong math and logical reasoning skills.

Instruction tuned LLM achieving SoTA performance on reasoning, math and general knowledge capabilities

Distilled version of Qwen 2.5 14B using reasoning data generated by DeepSeek R1 for enhanced performance.

Distilled version of Qwen 2.5 32B using reasoning data generated by DeepSeek R1 for enhanced performance.

Distilled version of Qwen 2.5 7B using reasoning data generated by DeepSeek R1 for enhanced performance.

Advanced state-of-the-art LLM with language understanding, superior reasoning, and text generation.

Long context cutting-edge lightweight open language model exceling in high-quality reasoning.

Advanced state-of-the-art LLM with language understanding, superior reasoning, and text generation.

Lightweight, state-of-the-art open LLM with strong math and logical reasoning skills.

Chinese and English LLM targeting for language, coding, mathematics, reasoning, etc.

Cutting-edge lightweight open language model exceling in high-quality reasoning.

Chinese and English LLM targeting for language, coding, mathematics, reasoning, etc.

Cutting-edge lightweight open language model exceling in high-quality reasoning.

Cutting-edge lightweight open language model exceling in high-quality reasoning.

State-of-the-art open model trained on open datasets, excelling in reasoning, math, and science.

Advanced state-of-the-art small language model with language understanding, superior reasoning, and text generation.

Advanced state-of-the-art LLM with language understanding, superior reasoning, and text generation.

Powers complex conversations with superior contextual understanding, reasoning and text generation.

Advanced state-of-the-art small language model with language understanding, superior reasoning, and text generation.

Excels in NLP tasks, particularly in instruction-following, reasoning, and mathematics.

Cutting-edge open multimodal model exceling in high-quality reasoning from images.