Browse and compare AI models across providers, modalities, and use cases.
Showing 20 of 56 models
Our cutting-edge language model for code completion released end of July 2025, Codestral specializes in low-latency, high-frequency tasks such as fill-in-the-middle (FIM) and code generation.
Pricing
Input: $0.30 / 1M tokensOutput: $0.90 / 1M tokens
Context
128.0K
Our cutting-edge language model for coding with the first version released May 2024
Context
32.0K
Our cutting-edge language model for coding with the second version released January 2025, Codestral specializes in low-latency, high-frequency tasks such as fill-in-the-middle (FIM), code correction and test generation. Learn more on our blog post
Pricing
Input: $0.30 / 1M tokensOutput: $0.90 / 1M tokens
Context
256.0K
Our cutting-edge language model for coding released in January 2025, Codestral specializes in low-latency, high-frequency tasks such as fill-in-the-middle (FIM), code correction and test generation. Learn more in our blog post
Context
256.0K
Our state-of-the-art semantic for extracting representation of code extracts
Context
8.0K
Our state-of-the-art semantic for extracting representation of code extracts
Pricing
Input: $0.15 / 1M tokensOutput: $0.15 / 1M tokens
Context
8.0K
Our first mamba 2 open source model released July 2024. Learn more on our blog post
Context
256.0K
Our frontier code agents model for solving software engineering tasks; excels at using tools to explore codebases, editing multiple files and power software engineering agents.
Pricing
Input: $0.40 / 1M tokensOutput: $2.00 / 1M tokens
Context
256.0K
An enterprise grade text model, that excels at using tools to explore codebases, editing multiple files and power software engineering agents.
Pricing
Input: $0.40 / 1M tokensOutput: $2.00 / 1M tokens
Context
128.0K
An update to our open source model that excels at using tools to explore codebases, editing multiple files and power software engineering agents. Learn more in our blog post
Pricing
Input: $0.10 / 1M tokensOutput: $0.30 / 1M tokens
Context
128.0K
Our open source model that excels at using tools to explore codebases, editing multiple files and power software engineering agents.
Pricing
Input: $0.10 / 1M tokensOutput: $0.30 / 1M tokens
Context
256.0K
Our frontier-class multimodal reasoning model update of September 2025.
Pricing
Input: $2.00 / 1M tokensOutput: $5.00 / 1M tokens
Context
128.0K
Our small multimodal reasoning model update of September 2025.
Pricing
Input: $0.50 / 1M tokensOutput: $1.50 / 1M tokens
Context
128.0K
Ministral 3 14B is the largest model in the Ministral 3 family, offering state-of-the-art capabilities and performance comparable to its larger Mistral Small 3.2 24B counterpart. Optimized for local deployment, it delivers high performance across diverse hardware, including local setups.
Pricing
Input: $0.20 / 1M tokensOutput: $0.20 / 1M tokens
Context
256.0K
Ministral 3 3B is the smallest and most efficient model in the Ministral 3 family, offering robust language and vision capabilities in a compact package. Designed for edge deployment, it delivers high performance across diverse hardware, including local setups.
Pricing
Input: $0.10 / 1M tokensOutput: $0.10 / 1M tokens
Context
256.0K
Ministral 3 8B is a powerful and efficient model in the Ministral 3 family, offering best-in-class text and vision capabilities. Built for edge deployment, it delivers high performance across diverse hardware, including local setups.
Pricing
Input: $0.15 / 1M tokensOutput: $0.15 / 1M tokens
Context
256.0K
World’s best edge model. Learn more on our blog post
Pricing
Input: $0.04 / 1M tokensOutput: $0.04 / 1M tokens
Context
131.0K
Powerful edge model with extremely high performance/price ratio. Learn more on our blog post
Pricing
Input: $0.10 / 1M tokensOutput: $0.10 / 1M tokens
Context
131.0K
Powerful edge model with extremely high performance/price ratio. Learn more in our blog post
Context
128.0K