Mixture of Experts (MoE) reasoning LLM (text-only) designed to fit within 80GB GPU.
Deploy this model now on your endpoint provider of choice