
State-of-the-art 685B reasoning LLM with sparse attention, long context, and integrated agentic tools.

Open, efficient MoE model with 1M context, excelling in coding, reasoning, instruction following, tool calling, and more

Open reasoning model with 256K context window, native INT4 quantization and enhanced tool use

Open Mixture of Experts LLM (230B, 10B active) for reasoning, coding, and tool-use/agent workflows

Multilingual, cross-lingual embedding model for long-document QA retrieval, supporting 26 languages.

Qwen3-Next Instruct blends hybrid attention, sparse MoE, and stability boosts for ultra-long context AI.

Follow-on version of Kimi-K2-Instruct with longer context window and enhanced reasoning capabilities

ByteDance open-source LLM with long-context, reasoning, and agentic intelligence.

Excels in agentic coding and browser use and supports 256K context, delivering top results.

Multilingual, cross-lingual embedding model for long-document QA retrieval, supporting 26 languages.

Multilingual and cross-lingual text question-answering retrieval with long context support and optimized data storage efficiency.

Fine-tuned reranking model for multilingual, cross-lingual text question-answering retrieval, with long context support.

Long context cutting-edge lightweight open language model exceling in high-quality reasoning.