Multi-modal model to classify safety for input prompts as well output responses.
An edge computing AI model which accepts text, audio and image input, ideal for resource-constrained environments
An edge computing AI model which accepts text, audio and image input, ideal for resource-constrained environments
Generates physics-aware video world states for physical AI development using text prompts and multiple spatial control inputs derived from real-world data or simulation.
Multimodal question-answer retrieval representing user queries as text and documents as images.
Updated version of DeepSeek-R1 with enhanced reasoning, coding, math, and reduced hallucination.
State-of-the-art model for Polish language processing tasks such as text generation, Q&A, and chatbots.
Advanced reasoing MOE mode excelling at reasoning, multilingual tasks, and instruction following
FLUX.1-schnell is a distilled image generation model, producing high quality images at fast speeds
State-of-the-art, multilingual model tailored to all 24 official European Union languages.
SOTA LLM pre-trained for instruction following and proficiency in Indonesian language and its dialects.
Efficient multimodal model excelling at multilingual tasks, image understanding, and fast-responses
Generalist model to generate future world state as videos from text and image prompts to create synthetic training data for robots and autonomous vehicles.
Generates future frames of a physics-aware world state based on simply an image or short video prompt for physical AI development.
End-to-end autonomous driving stack integrating perception, prediction, and planning with sparse scene representations for efficiency and safety.
Leading reasoning and agentic AI accuracy model for PC and edge.
Distilled version of Llama 3.1 8B using reasoning data generated by DeepSeek R1 for enhanced performance.
Distilled version of Qwen 2.5 32B using reasoning data generated by DeepSeek R1 for enhanced performance.
Distilled version of Qwen 2.5 14B using reasoning data generated by DeepSeek R1 for enhanced performance.
Distilled version of Qwen 2.5 7B using reasoning data generated by DeepSeek R1 for enhanced performance.
Lightweight multilingual LLM powering AI applications in latency bound, memory/compute constrained environments
State-of-the-art, high-efficiency LLM excelling in reasoning, math, and coding.
Topic control model to keep conversations focused on approved topics, avoiding inappropriate content.
Leading content safety model for enhancing the safety and moderation capabilities of LLMs
Context-aware chart extraction that can detect 18 classes for chart basic elements, excluding plot elements.
FourCastNet predicts global atmospheric dynamics of various weather / climate variables.
Advanced AI model detects faces and identifies deep fake images.
Sovereign AI model trained on Japanese language that understands regional nuances.
Sovereign AI model trained on Japanese language that understands regional nuances.
Unique language model that delivers an unmatched accuracy-efficiency performance.
Robust image classification model for detecting and managing AI-generated content.
Sovereign AI model finetuned on Traditional Mandarin and English data using the Llama-3 architecture.
Sovereign AI model trained on Japanese language that understands regional nuances.
Cutting-edge MOE based LLM designed to excel in a wide array of generative AI tasks.
Cutting-edge MOE based LLM designed to excel in a wide array of generative AI tasks.
Lightweight multilingual LLM powering AI applications in latency bound, memory/compute constrained environments
Grounding dino is an open vocabulary zero-shot object detection model.
Advanced small language generative AI model for edge applications
EfficientDet-based object detection network to detect 100 specific retail objects from an input video.
LLM to represent and serve the linguistic and cultural diversity of Southeast Asia
An MOE LLM that follows instructions, completes requests, and generates creative text.
Stable Video Diffusion (SVD) is a generative diffusion model that leverages a single image as a conditioning frame to synthesize video sequences.
An MOE LLM that follows instructions, completes requests, and generates creative text.