🚀 New model available: DeepSeek-V3.1 🚀
Developed by Mistral AI, a leading French research lab, Mistral is a family of open-source AI models built for multilingual excellence, advanced reasoning, and cost-effective performance. These models excel at complex reasoning, mathematics, coding, and specialized tasks while offering complete transparency and deployment freedom through open-source licensing.
Mistral Small 3.2 delivers breakthrough efficiency with native fluency in European languages, while specialized variants handle specific needs: Devstral for coding, Voxtral for audio processing, and Mixtral for high-performance tasks. With Apache 2.0 licensing, extensive context windows up to 128K tokens, and comprehensive customization options, Mistral provides enterprise-grade capabilities without vendor lock-in.
Perfect for building multilingual applications, coding assistants, and reasoning systems where you need both powerful performance and complete control over your AI deployment.
Mistral-Small-3.2-24B-Instruct is a drop-in upgrade over the 3.1 release, with markedly better instruction following, roughly half the infinite-generation errors, and a more robust function-calling interface—while otherwise matching or slightly improving on all previous text and vision benchmarks.
Price per 1M input tokens
$0.05
Price per 1M output tokens
$0.10
Release Date
06/23/2025
Context Size
128,000
Quantization
fp8
# Assume openai>=1.0.0
from openai import OpenAI
# Create an OpenAI client with your deepinfra token and endpoint
openai = OpenAI(
api_key="$DEEPINFRA_TOKEN",
base_url="https://api.deepinfra.com/v1/openai",
)
chat_completion = openai.chat.completions.create(
model="mistralai/Mistral-Small-3.2-24B-Instruct-2506",
messages=[{"role": "user", "content": "Hello"}],
)
print(chat_completion.choices[0].message.content)
print(chat_completion.usage.prompt_tokens, chat_completion.usage.completion_tokens)
# Hello! It's nice to meet you. Is there something I can help you with, or would you like to chat?
# 11 25
Developed by Mistral AI, a leading French research lab, Mistral is a family of open-source AI models built for multilingual excellence, advanced reasoning, and cost-effective performance
Model | Context | $ per 1M input tokens | $ per 1M output tokens | Actions |
---|---|---|---|---|
Mistral-Small-3.2-24B-Instruct-2506 | 125k | $0.05 | $0.10 | |
Mistral-Small-3.1-24B-Instruct-2503 | 125k | $0.05 | $0.10 | |
Mistral-Small-24B-Instruct-2501 | 32k | $0.05 | $0.08 | |
Mistral-7B-Instruct-v0.3 | 32k | $0.028 | $0.054 | |
Mistral-Nemo-Instruct-2407 | 128k | $0.02 | $0.04 | |
Mixtral-8x7B-Instruct-v0.1 | 32k | $0.08 | $0.24 | |
Devstral-Small-2507 | 125k | $0.07 | $0.28 |
Voxtral is a family of audio models with state-of-the-art speech to text capabilities.
Model | $ per minute of audio input | Actions |
---|---|---|
Voxtral-Small-24B-2507 | $0.00300 | |
Voxtral-Mini-3B-2507 | $0.00100 |
© 2025 Deep Infra. All rights reserved.