NVIDIA
Explore
Models
Blueprints
GPUs
Docs
⌘KCtrl+K
Terms of Use
Privacy Policy
Your Privacy Choices
Contact

Copyright © 2026 NVIDIA Corporation

19 results for

Filters

  • Free Endpoint
    7
  • Download Available
    12
  • Image-to-Text
    4
  • Code Generation
    2
  • Qwen
    5
  • Mistral AI
    3
  • NVIDIA
    3
  • Meta
    2
  • OpenAI
    2
  • Qwen
    Downloadable

    qwen3.5-397b-a17b

    Next-gen Qwen 3.5 VLM (400B MoE) brings advanced vision, chat, RAG, and agentic capabilities.
    Model
    chat
    8.02M
    3w
    NVIDIA
    Downloadable

    nemotron-3-nano-30b-a3b

    Open, efficient MoE model with 1M context, excelling in coding, reasoning, instruction following, tool calling, and more
    Model
    chat
    12.23M
    3mo
    NVIDIA
    Downloadable

    nemotron-3-super-120b-a12b

    Open, efficient hybrid Mamba-Transformer MoE with 1M context, excelling in agentic reasoning, coding, planning, tool calling, and more
    Model
    chat
    329K
    4d
    Qwen
    Free Endpoint

    qwen3-coder-480b-a35b-instruct

    Excels in agentic coding and browser use and supports 256K context, delivering top results.
    Model
    agentic coding
    3.91M
    6mo
    Z.ai
    Downloadable

    glm-5

    GLM-5 744B MoE enables efficient reasoning for complex systems and long-horizon agentic tasks.
    Model
    MoE
    9.8M
    1mo
    Meta
    DownloadableFree Endpoint

    llama-4-scout-17b-16e-instruct

    A multimodal, multilingual 16 MoE model with 17B parameters.
    Model
    language generation
    156K
    8mo
    OpenAI
    Downloadable

    gpt-oss-120b

    Mixture of Experts (MoE) reasoning LLM (text-only) designed to fit within 80GB GPU.
    Model
    reasoning
    41.01M
    7mo
    OpenAI
    Downloadable

    gpt-oss-20b

    Smaller Mixture of Experts (MoE) text-only LLM for efficient AI reasoning and math
    Model
    reasoning
    8.46M
    7mo
    AI21 Labs
    Free Endpoint

    jamba-1.5-mini-instruct

    Cutting-edge MOE based LLM designed to excel in a wide array of generative AI tasks.
    Model
    chat
    571K
    9mo
    Moonshotai
    Downloadable

    kimi-k2.5

    1T multimodal MoE for high‑capacity video and image understanding with efficient inference.
    Model
    Multimodal
    22.84M
    1mo
    Meta
    Free Endpoint

    llama-4-maverick-17b-128e-instruct

    A general purpose multimodal, multilingual 128 MoE model with 17B parameters.
    Model
    chat
    3.25M
    8mo
    Mistral AI
    Downloadable

    mixtral-8x22b-instruct-v0.1

    An MOE LLM that follows instructions, completes requests, and generates creative text.
    Model
    chat
    4.96M
    8mo
    Mistral AI
    Downloadable

    mixtral-8x7b-instruct-v0.1

    An MOE LLM that follows instructions, completes requests, and generates creative text.
    Model
    chat
    750K
    8mo
    Qwen
    Downloadable

    qwen3-next-80b-a3b-instruct

    Qwen3-Next Instruct blends hybrid attention, sparse MoE, and stability boosts for ultra-long context AI.
    Model
    chat
    11.94M
    5mo
    Qwen
    Downloadable

    qwen3-next-80b-a3b-thinking

    80B parameter AI model with hybrid reasoning, MoE architecture, support for 119 languages.
    Model
    chat
    4.24M
    6mo
    Qwen
    Free Endpoint

    qwen3.5-122b-a10b

    122B MoE LLM (10B active) for coding, reasoning, multimodal chat. Agent-ready.
    Model
    chat
    1.49M
    1w
    Stepfun-ai
    Free Endpoint

    step-3.5-flash

    200B open-source reasoning engine with sparse MoE powering frontier agentic AI.
    Model
    chat
    7.8M
    1mo
    Mistral AI
    Free Endpoint

    mistral-large-3-675b-instruct-2512

    A state-of-the-art general purpose MoE VLM ideal for chat, agentic and instruction based use cases.
    Model
    chat
    6.69M
    3mo
    DGX Spark
    30 MIN

    CUDA-X Data Science

    Install and use NVIDIA cuML and NVIDIA cuDF to accelerate UMAP, HDBSCAN, pandas and more with zero code changes
    Playbook
    pandas
    4mo
    Items per page
    of 1 pages