
Mixture of Experts (MoE) reasoning LLM (text-only) designed to fit within 80GB GPU.

An edge computing AI model which accepts text, audio and image input, ideal for resource-constrained environments

An edge computing AI model which accepts text, audio and image input, ideal for resource-constrained environments


Cutting-edge open multimodal model exceling in high-quality reasoning from images.

A lightweight, multilingual, advanced SLM text model for edge computing, resource constraint applications

Powerful mid-size code model with a 32K context length, excelling in coding in multiple languages.

Advanced small language generative AI model for edge applications

Estimate gaze angles of a person in a video and redirect to make it frontal.

Cutting-edge text generation model text understanding, transformation, and code generation.

Cutting-edge text generation model text understanding, transformation, and code generation.