Model for writing and interacting with code across a wide range of programming languages and tasks.
Codestral Mamba is trained on a diverse dataset of 80+ programming languages, including the most popular ones, such as Python, Java, C, C++, JavaScript, and Bash. It also performs well on more specific ones like Swift and Fortran. This broad language base ensures Codestral Mamba can assist developers in various coding environments and projects.
This model is ready for commercial use and testing purposes.
This model is not owned or developed by NVIDIA. This model has been developed and built to a third-party’s requirements for this application and use case; see Codestral's Mamba 7B Hugging Face Model Card.
By using this software or model, you are agreeing to the terms and conditions of the license, acceptable use policy and Mistral's privacy policy. Mamba-Codestral-7B-v0.1 is released under the Apache 2.0 license.
Mamba Codestral 7B blogpost
Architecture Type: Transformer
Network Architecture: Mamba Codestral 7B v0.1
Model Version: 0.1
Input Format: Text
Input Parameters: Max Tokens, Temperature, Top P
Max Input Tokens: 4096
Output Format: Text
Output Parameters: None
Max Output Tokens: 4096
Supported Hardware Platform(s): NVIDIA Ampere, NVIDIA Hopper, NVIDIA Turing
Supported Operating System(s): Linux
Engine: TRT-LLM
Test Hardware: L40S
NVIDIA believes Trustworthy AI is a shared responsibility and we have established policies and practices to enable development for a wide array of AI applications. When downloaded or used in accordance with our terms of service, developers should work with their internal model team to ensure this model meets requirements for the relevant industry and use case and addresses unforeseen product misuse. Please report security vulnerabilities or NVIDIA AI Concerns here.