Mistral technology

AI models

We release the world’s most capable open models, enabling frontier AI innovation.

Developer platform

Our portable developer platform serves our open and optimized models for building fast and intelligent applications. Get started for free!

Premier models

State of the art models across a variety of sizes, available to experiment under the Mistral Research License and take to production with the Commercial License.

Mistral Large 24.11

Top-tier reasoning for high-complexity tasks and sophisticated problems.

  • Multi-lingual and high proficiency in coding
  • 128k token context window
  • Deployable anywhere (on-prem / VPC / API)
Pixtral Large

Vision-capable large model with frontier reasoning capabilities.

  • Vision-capable large model
  • Analyze, search, review, and better understand images
  • Deployable anywhere (on-prem / VPC / API)
Mistral Small 24.09

Enterprise-grade small model.

  • The most powerful model in its size
  • Available under the Mistral Research License
  • 128k token context window
  • Cost-efficient and fast model for a wide array of use cases such as translation, summarization, and sentiment analysis
Codestral

State-of-the-art Mistral model trained specifically for code tasks.

  • Trained on 80+ programming languages (incl. Python, Java, C, C++, PHP, Bash)
  • Optimized for low latency: Way smaller than competitive coding models
  • Context window of 32K tokens
Ministral 8B 24.10

Our most powerful edge model. Successor to Mistral 7B.
Sets the benchmark in commonsense, reasoning, and function-calling in the sub-10B category.

  • 128k token context window
  • Ideal for on-device computing and edge use cases
  • Highly capable in reasoning, function-calling, and common sense
Ministral 3B 24.10

Our most efficient edge model.
The most capable in its category, ideal for low-power, low-latency on-device computing and edge use cases.

  • 128k token context window
  • Highly capable in function-calling for agentic workflows
  • Ideal for lossless quantization and tailoring specific use cases
Mistral Embed

State-of-the-art semantic for extracting representation of text extracts.

Mistral Moderation 24.11

A classifier service for text content moderation.

  • 8K token context window
  • Supports 9 policies for undersirable content
  • Suitable for user and model generated content
Try it on la Plateforme

Get started with the API

Get a commercial license

For POCs, testing, and production

Download weights

For research only

For more details on the various pricing options, check out our pricing page here: See pricing.

Free models

Free to use under the Apache 2.0 license.

Latest models
Pixtral 12B

Vision-capable small model.

  • Analyze, search, review, and better understand images
  • Available under the Apache 2.0 license
  • Deploy in your own environment so you don’t have to upload your files to a third party provider
Mathstral

Variant of Mistral-7B, optimized for solving advanced mathematics problems.

  • Context window of 32K tokens
  • 7B parameters
  • Available under Apache 2.0 license
Codestral Mamba

A Mamba2 language model designed for coding tasks.

  • Context window of 256K tokens
  • 7.3B parameters
  • Available under Apache 2.0 license
Mistral NeMo

A state-of-the-art 12B small model built in collaboration with NVIDIA.

  • The most powerful model in its size category
  • Available under Apache 2.0 license
  • Multi-lingual (incl. European languages, Chinese, Japanese, Korean, Hindi, Arabic)
  • Large context window of 128K tokens
Mistral 8 X 22B

Mixtral 8x22B set a new standard for performance and efficiency, with only 39B active parameters out of 141B, offering unparalleled cost efficiency for its size. It is natively capable of function calling, which enables application development and tech stack modernisation at scale.

    Mistral 8 X 7B

    A high-quality sparse mixture of experts (SMoE) with open weights. Matches or outperforms GPT3.5 on most standard benchmarks, particularly in multilingual capabilities and code.

      Mistral 7B

      The first Mistral model, engineered for superior performance and efficiency. The model leverages grouped-query attention (GQA) for faster inference, coupled with sliding window attention (SWA) to effectively handle sequences of arbitrary length with a reduced inference cost.

        These models also have commercial licenses for business purposes: Explore commercial licenses.

        Fine-tune our models

        We allow you to fine-tune our models in an easy, effective & cost-efficient way, and thus use smaller and better-suited models to solve your specific use cases. Fine-tuning can be done with our open-source fine-tuning code as well as on La Plateforme with our efficient Fine-tuning API.

        Fine-tune Mistral models on La Plateforme with Mistral Fine-tuning API

        Leverage Mistral’s unique expertise in training models by using our highly efficient fine-tuning service to specialize both our open-source and commercial models.

        Use Mistral fine-tuning code

        Benefit from Mistral fine-tuning code to perform fine-tuning on Mistral open-source models on your own.

        Pay-as-you-go pricing

        Price in $
        Price in €
        Premier models
        ModelAPI NameDescriptionInput (/M tokens)Output (/M tokens)
        Mistral Large 24.11mistral-large-latestTop-tier reasoning for high-complexity tasks and sophisticated problems.$2$6
        Pixtral Largepixtral-large-latestVision-capable large model with frontier reasoning capabilities.$2$6
        Mistral Small 24.09mistral-small-latestCost-efficient, fast, and reliable option for use cases such as translation, summarization, and sentiment analysis.$0.2$0.6
        Codestralcodestral-latestState-of-the-art Mistral model trained specifically for code tasks.$0.2$0.6
        Ministral 8B 24.10ministral-8b-latestPowerful model for on-device use cases.$0.1$0.1
        Ministral 3B 24.10ministral-3b-latestMost efficient edge model.$0.04$0.04
        Mistral Embedmistral-embedState-of-the-art semantic for extracting representation of text extracts.$0.1
        Mistral Moderation 24.11mistral-moderation-latestA classifier service for text content moderation.$0.1
        ModelAPI NameDescriptionInput (/M tokens)Output (/M tokens)
        Mistral Large 24.11mistral-large-latestTop-tier reasoning for high-complexity tasks and sophisticated problems.1.8€5.4€
        Pixtral Largepixtral-large-latestVision-capable large model with frontier reasoning capabilities.1.8€5.4€
        Mistral Small 24.09mistral-small-latestCost-efficient, fast, and reliable option for use cases such as translation, summarization, and sentiment analysis.0.18€0.54€
        Codestralcodestral-latestState-of-the-art Mistral model trained specifically for code tasks.0.18€0.54€
        Ministral 8B 24.10ministral-8b-latestPowerful model for on-device use cases.0.09€0.09€
        Ministral 3B 24.10ministral-3b-latestMost efficient edge model.0.04€0.04€
        Mistral Embedmistral-embedState-of-the-art semantic for extracting representation of text extracts.0.09€
        Mistral Moderation 24.11mistral-moderation-latestA classifier service for text content moderation.0.09€

        Models used with the Batch API cost 50% lower than prices shown above.

        Free models
        ModelAPI NameDescriptionInput (/M tokens)Output (/M tokens)
        Pixtral 12Bpixtral-12bVision-capable small model.$0.15$0.15
        Mistral NeMomistral-nemoState-of-the-art Mistral model trained specifically for code tasks.$0.15$0.15
        Mistral 7Bopen-mistral-7bA 7B transformer model, fast-deployed and easily customisable.$0.25$0.25
        Mixtral 8x7Bopen-mixtral-8x7bA 7B sparse Mixture-of-Experts (SMoE). Uses 12.9B active parameters out of 45B total.$0.7$0.7
        Mixtral 8x22Bopen-mixtral-8x22bMixtral 8x22B is currently the most performant open model. A 22B sparse Mixture-of-Experts (SMoE). Uses only 39B active parameters out of 141B.$2$6
        ModelAPI NameDescriptionInput (/M tokens)Output (/M tokens)
        Pixtral 12Bpixtral-12bVision-capable small model.0.13€0.13€
        Mistral NeMomistral-nemoState-of-the-art Mistral model trained specifically for code tasks.0.13€0.13€
        Mistral 7Bopen-mistral-7bA 7B transformer model, fast-deployed and easily customisable.0.2€0.2€
        Mixtral 8x7Bopen-mixtral-8x7bA 7B sparse Mixture-of-Experts (SMoE). Uses 12.9B active parameters out of 45B total.0.65€0.65€
        Mixtral 8x22Bopen-mixtral-8x22bMixtral 8x22B is currently the most performant open model. A 22B sparse Mixture-of-Experts (SMoE). Uses only 39B active parameters out of 141B.1.9€5.6€
        Fine-tuning
        ModelOne-off training (/M tokens)StorageInput (/M tokens)Output (/M tokens)
        Mistral NeMo$1$2 per month per model$0.15$0.15
        Mistral Large 24.11$9$4 per month per model$2$6
        Mistral Small$3$2 per month per model$0.2$0.6
        Codestral$3$2 per month per model$0.2$0.6
        ModelOne-off training (/M tokens)StorageInput (/M tokens)Output (/M tokens)
        Mistral NeMo0.9€1.8€ per month per model0.13€0.13€
        Mistral Large 24.118.2€3.8€ per month per model1.8€5.4€
        Mistral Small2.7€1.8€ per month per model0.18€0.54€
        Codestral2.7€1.8€ per month per model0.18€0.54€

        Tokens are numerical representations of words or parts of words. On average, one token is roughly equivalent to 4 characters or 0.75 words in English.

        Fine-tuning pricing explained

        Mistral AI provides a fine-tuning API through La Plateforme, making it easy to fine-tune our open-source and commercial models. There are three costs related to fine-tuning:

        • One-off training: Price per token on the data you want to fine-tune our standard models on; minimum fee per fine-tuning job of $4
        • Inference: Price per input/output token when using the fine-tuned model(s)
        • Storage: Price per month per model for storage (irrespective of model usage; models can be deleted any time)

        Deploy anywhere

        La Plateforme
        La Plateforme

        Get started with Mistral models in a few clicks via our developer platform hosted on Mistral’s infrastructure and build your own applications and services. Our servers are hosted in EU.

        Cloud platforms
        Cloud platforms

        Access our models via your preferred cloud provider and use your cloud credits. Our open models are currently available via our cloud partners (GCP, AWS, Azure, IBM, Snowflake, NVIDIA, Outscale).
        Mistral models are available on Azure AI Studio, AWS Bedrock, Google Cloud Model Garden, IBM Watsonx, and Snowflake.

        Self-deployment
        Self-deployment

        Deploy Mistral models on virtual cloud or on-prem. Self-deployment offers more advanced levels of customization and control. Your data stays within your walls. Try deploying our open models, and contact our team to deploy our optimized models similarly.

        La Plateforme

        Access our latest products via our developer platform, hosted in Europe

        from mistralai.client import MistralClient
        from mistralai.models.chat_completion import ChatMessage
        
        api_key = os.environ["MISTRAL_API_KEY"]
        model = "mistral-tiny"
        
        client = MistralClient(api_key=api_key)
        
        messages = [
            ChatMessage(role="user",
            content="Who is the most renowned French painter?")
        ]
        
        Built for developers

        La Plateforme is developers’ preferred way to access all Mistral Al’s models. Hosted and served on Mistral’s infrastructure, in Europe.

        • Our best models at the best price: Get access to our models at an unmatched price/performance point
        • Guides & community: Use our guides and community forums to build your own application and services
        • Secure by design: Your data are encrypted at rest (AES256) and in transit (TLS 1.2+); our servers are in the EU
        • Get started for free! No credit card required
        from mistralai.client import MistralClient
        from mistralai.models.chat_completion import ChatMessage
        
        api_key = os.environ["MISTRAL_API_KEY"]
        model = "mistral-tiny"
        
        client = MistralClient(api_key=api_key)
        
        messages = [
            ChatMessage(role="user",
            content="Who is the most renowned French painter?")
        ]
        

        Model licenses

        We distribute two categories of models:

        • Models licensed under Apache 2.0 that are no-string-attached contributions to the community and industry
        • Models with a double license: Mistral Research License / Mistral Commercial License. These models can be used for non-commercial purposes with no strings attached. A commercial license must be bought for commercial use cases requiring self-deployment. These models may also be used directly through our portable platform or distributors, with pay-as-you-go and reserved offerings.

        Apache 2.0Mistral Research LicenseMistral Commercial License
        Access to weights
        Deployment for research purposes and individual usage
        Creation of derivatives (e.g. fine-tuning) for research purposes and individual usage The same license applies to derivatives The same license applies to derivatives
        Deployment for commercial purposes (internal & external use cases) Requires Mistral Commercial License
        Creation and usage of derivatives (e.g. fine-tuning) for commercial use cases Requires Mistral Commercial License
        Custom terms & support (self-deployment)