Mistral Small 3 is a powerful 24-billion-parameter language model from mistralai, specifically engineered for low-latency performance across a wide range of common AI tasks. Released under the permissive Apache 2.0 license, this model offers both pre-trained and instruction-tuned versions, making it highly versatile for efficient local deployment and integration into various applications. Achieving an impressive 81% accuracy on the MMLU benchmark, Mistral Small 3 demonstrates competitive performance against significantly larger models such as Llama 3.3 70B and Qwen 32B. Crucially, it operates at three times the speed on equivalent hardware, providing a significant advantage for applications requiring rapid responses. With a generous 32K token context window and a 4K token max output, it supports complex interactions. This model is available for free access and is priced at $0.03 per 1M input tokens and $0.11 per 1M output tokens. Leverage Mistral Small 3 for tasks like chat, code generation, and translation. It supports advanced capabilities including function calling, code interpretation, and streaming output. Note that it does not support image generation or internet access.
✅ Best For
🚀 Capabilities
❌ Limitations
Specifications
| Provider | mistralai |
| Context Window | 32,768 tokens |
| Max Output | 16,384 tokens |
| Minimum Plan | Economy |
Pricing
| Input Price | $0.0500 / 1M tokens |
| Output Price | $0.0800 / 1M tokens |
💡 With PRO subscription, cost is reduced by 20%