NVIDIA
Explore
Models
Blueprints
GPUs
Docs
⌘KCtrl+K
Terms of Use
Privacy Policy
Your Privacy Choices
Contact

Copyright © 2026 NVIDIA Corporation

12 results for

Filters (1)

  • Free Endpoint
    2
  • Partner Endpoint
    10
  • Download Available
    10
  • Code Generation
    2
  • Image-to-Text
    1
  • Fireworks AI
    7
  • Deep Infra
    7
  • Together AI
    7
  • GMI Cloud
    6
  • CoreWeave
    5
  • Mistral AI
    3
  • Qwen
    2
  • NVIDIA
    2
  • OpenAI
    2
  • Moonshotai
    1
  • reasoning
  • NVIDIA
    Downloadable

    nemotron-3-nano-30b-a3b

    Open, efficient MoE model with 1M context, excelling in coding, reasoning, instruction following, tool calling, and more
    Model
    chat
    11.98M
    3mo
    NVIDIA
    Downloadable

    nemotron-3-super-120b-a12b

    Open, efficient hybrid Mamba-Transformer MoE with 1M context, excelling in agentic reasoning, coding, planning, tool calling, and more
    Model
    chat
    3.11M
    1w
    Z.ai
    Downloadable

    glm-5

    GLM-5 744B MoE enables efficient reasoning for complex systems and long-horizon agentic tasks.
    Model
    MoE
    10.65M
    1mo
    OpenAI
    Downloadable

    gpt-oss-120b

    Mixture of Experts (MoE) reasoning LLM (text-only) designed to fit within 80GB GPU.
    Model
    reasoning
    37.62M
    7mo
    OpenAI
    Downloadable

    gpt-oss-20b

    Smaller Mixture of Experts (MoE) text-only LLM for efficient AI reasoning and math
    Model
    reasoning
    8.36M
    7mo
    Moonshotai
    Downloadable

    kimi-k2.5

    1T multimodal MoE for high‑capacity video and image understanding with efficient inference.
    Model
    Multimodal
    20.31M
    1mo
    Mistral AI
    Downloadable

    mistral-small-4-119b-2603

    Hybrid MoE model unifying instruct, reasoning, and coding with multimodal input and 256k context
    Model
    chat
    938
    2d
    Mistral AI
    Downloadable

    mixtral-8x22b-instruct-v0.1

    An MOE LLM that follows instructions, completes requests, and generates creative text.
    Model
    chat
    4.98M
    8mo
    Mistral AI
    Downloadable

    mixtral-8x7b-instruct-v0.1

    An MOE LLM that follows instructions, completes requests, and generates creative text.
    Model
    chat
    718K
    8mo
    Qwen
    Downloadable

    qwen3-next-80b-a3b-thinking

    80B parameter AI model with hybrid reasoning, MoE architecture, support for 119 languages.
    Model
    chat
    4.43M
    6mo
    Qwen
    Free Endpoint

    qwen3.5-122b-a10b

    122B MoE LLM (10B active) for coding, reasoning, multimodal chat. Agent-ready.
    Model
    chat
    1.73M
    1w
    Stepfun-ai
    Free Endpoint

    step-3.5-flash

    200B open-source reasoning engine with sparse MoE powering frontier agentic AI.
    Model
    chat
    7.89M
    1mo
    Items per page
    of 1 pages