
An MOE LLM that follows instructions, completes requests, and generates creative text.
Mixtral 8x22B is MistralAI's latest open model. It sets a new standard for performance and efficiency within the AI community. It is a sparse Mixture-of-Experts (SMoE) model that uses only 39B active parameters out of 141B, offering unparalleled cost efficiency for its size.
Mixtral 8x22B comes with the following strengths:
This model is not owned or developed by NVIDIA. This model has been developed and built to a third-party’s requirements for this application and use case; see Mixtral 8x22b's Model Card.
By using this software or model, you are agreeing to the terms and conditions of the license, acceptable use policy and Mistral's privacy policy. Mixtral-8x22B is released under the Apache 2.0 license
Mixtral 8x22B Instruct Model Card on Hugging Face
Cheaper, Better, Faster, Stronger | Mistral AI
Architecture Type: Transformer
Network Architecture: Sparse Mixture of GPT-based experts
Model Version: 0.1
Input Format: Text
Input Parameters: Temperature, Top P, Max Output Tokens
Output Format: Text
Output Parameters: None
Supported Hardware Platform(s): Hopper, Ampere, Turing, Ada
Supported Operating System(s): Linux
Engine: Triton
Test Hardware: Other