We're releasing Mistral Small 3!
- 24B params, 81% MMLU
- Latency optimized: 150 tokens/s
- Competitive with Llama-3.3 70B, Qwen-2.5 32B, GPT4o-mini
- Apache 2.0
https://mistral.ai/news/mistral-small-3/
- 24B params, 81% MMLU
- Latency optimized: 150 tokens/s
- Competitive with Llama-3.3 70B, Qwen-2.5 32B, GPT4o-mini
- Apache 2.0
https://mistral.ai/news/mistral-small-3/
Comments
https://huggingface.co/mistralai/Ministral-8B-Instruct-2410
https://huggingface.co/mistralai/Mistral-Small-24B-Base-2501
https://huggingface.co/mistralai/Mistral-Small-24B-Instruct-2501
- Ollama: https://ollama.com/library/mistral-small
- Kaggle: https://kaggle.com/models/mistral-ai/mistral-small-24b
- Fireworks: https://fireworks.ai/models/fireworks/mistral-small-24b-instruct-2501
- Together: https://together.ai/blog/mistral-small-3-api-now-available-on-together-ai-a-new-category-leader-in-small-models
And many more soon!
Very random question from someone not really too deep on this stuff, but why did you compared it vs Gemma and not Gemini 1.5 Flash?