
Open, efficient MoE model with 1M context, excelling in coding, reasoning, instruction following, tool calling, and more

State-of-the-art open code model with deep reasoning, 256k context, and unmatched efficiency.

Open Mixture of Experts LLM (230B, 10B active) for reasoning, coding, and tool-use/agent workflows

Follow-on version of Kimi-K2-Instruct with longer context window and enhanced reasoning capabilities

Excels in agentic coding and browser use and supports 256K context, delivering top results.

State-of-the-art open mixture-of-experts model with strong reasoning, coding, and agentic capabilities

Supports Chinese and English languages to handle tasks including chatbot, content generation, coding, and translation.

High performance reasoning model optimized for efficiency and edge deployment

Superior inference efficiency with highest accuracy for scientific and complex math reasoning, coding, tool calling, and instruction following.

Small language model fine-tuned for improved reasoning, coding, and instruction-following

Distilled version of Llama 3.1 8B using reasoning data generated by DeepSeek R1 for enhanced performance.

Built for agentic workflows, this model excels in coding, instruction following, and function calling

Instruction tuned LLM achieving SoTA performance on reasoning, math and general knowledge capabilities

Distilled version of Qwen 2.5 14B using reasoning data generated by DeepSeek R1 for enhanced performance.

Distilled version of Qwen 2.5 32B using reasoning data generated by DeepSeek R1 for enhanced performance.

Distilled version of Qwen 2.5 7B using reasoning data generated by DeepSeek R1 for enhanced performance.

Support Chinese and English chat, coding, math, instruction following, solving quizzes

Chinese and English LLM targeting for language, coding, mathematics, reasoning, etc.

Chinese and English LLM targeting for language, coding, mathematics, reasoning, etc.

Powerful mid-size code model with a 32K context length, excelling in coding in multiple languages.

Advanced LLM for synthetic data generation, distillation, and inference for chatbots, coding, and domain-specific tasks.