
Leading multilingual content safety model for enhancing the safety and moderation capabilities of LLMs

Multilingual, cross-lingual embedding model for long-document QA retrieval, supporting 26 languages.

Multilingual 7B LLM, instruction-tuned on all 24 EU languages for stable, culturally aligned output.

Multilingual, cross-lingual embedding model for long-document QA retrieval, supporting 26 languages.

Fine-tuned reranking model for multilingual, cross-lingual text question-answering retrieval, with long context support.

Advanced reasoing MOE mode excelling at reasoning, multilingual tasks, and instruction following

Multilingual and cross-lingual text question-answering retrieval with long context support and optimized data storage efficiency.

A general purpose multimodal, multilingual 128 MoE model with 17B parameters.

A multimodal, multilingual 16 MoE model with 17B parameters.

High performance reasoning model optimized for efficiency and edge deployment

Latency-optimized language model excelling in code, math, general knowledge, and instruction-following.

Natural and expressive voices in multiple languages. For voice agents and brand ambassadors.

State-of-the-art, multilingual model tailored to all 24 official European Union languages.

Lightweight multilingual LLM powering AI applications in latency bound, memory/compute constrained environments

NVIDIA DGX Cloud trained multilingual LLM designed for mission critical use cases in regulated industries including financial services, government, heavy industry

Multilingual LLM with emphasis on European languages supporting regulated use cases including financial services, government, heavy industry

A lightweight, multilingual, advanced SLM text model for edge computing, resource constraint applications

Efficient multimodal model excelling at multilingual tasks, image understanding, and fast-responses

High accuracy and optimized performance for transcription in 25 languages

Robust Speech Recognition via Large-Scale Weak Supervision.

Lightweight multilingual LLM powering AI applications in latency bound, memory/compute constrained environments