An MOE LLM that follows instructions, completes requests, and generates creative text.
An MOE LLM that follows instructions, completes requests, and generates creative text.
Mixtral 8x22B is MistralAI's latest open model. It sets a new standard for performance and efficiency within the AI community. It is a sparse Mixture-of-Experts (SMoE) model that uses only 39B active parameters out of 141B, offering unparalleled cost efficiency for its size.
Mixtral 8x22B comes with the following strengths:
This model is not owned or developed by NVIDIA. This model has been developed and built to a third-party’s requirements for this application and use case; see Mixtral 8x22b's Model Card.
By using this software or model, you are agreeing to the terms and conditions of the license, acceptable use policy and Mistral's privacy policy. Mixtral-8x22B is released under the Apache 2.0 license
Mixtral 8x22B Instruct Model Card on Hugging Face
Cheaper, Better, Faster, Stronger | Mistral AI
Architecture Type: Transformer
Network Architecture: Sparse Mixture of GPT-based experts
Model Version: 0.1
Input Format: Text
Input Parameters: Temperature, Top P, Max Output Tokens
Output Format: Text
Output Parameters: None
Supported Hardware Platform(s): Hopper, Ampere, Turing, Ada
Supported Operating System(s): Linux
Engine: Triton
Test Hardware: Other