Most advanced language model for reasoning, code, multilingual tasks; runs on a single GPU.
Mistral-NeMo is a Large Language Model (LLM) composed of 12B parameters. This model leads accuracy on popular benchmarks across common sense reasoning, coding, math, multilingual and multi-turn chat tasks; it significantly outperforms existing models smaller or similar in size.
This model is ready for commercial use.
This model was a jointly trained by Mistral and NVIDIA.
Your use of this API is governed by the NVIDIA API Trial Service Terms of Use; and the use of this model is governed by the NVIDIA AI Foundation Models Community License. Mistral NeMo-12B is released under the Apache 2.0 license.
Mistral NeMo 12B Blogpost
Architecture Type: Transformer
Network Architecture: Mistral
Model Version: 0.1
This transformer model has the following characteristics:
Input
Output
Engine: TensorRT-LLM
Test Hardware: H100
When downloaded or used in accordance with our terms of service, developers should work with their internal model team to ensure this model meets requirements for the relevant industry and use case and addresses unforeseen product misuse. Please report security vulnerabilities or NVIDIA AI Concerns here.