unsloth/Magistral-Small-2509

Warm
Public
Vision
24B
FP8
32768
Sep 17, 2025
License: apache-2.0
Hugging Face
Overview

Magistral Small 1.2 Overview

Magistral Small 1.2 is a 24 billion parameter multimodal language model from Mistral AI, based on Mistral Small 3.2 (2506). It has been fine-tuned with added reasoning capabilities through Supervised Fine-Tuning (SFT) and Reinforcement Learning (RL), making it an efficient model for complex analytical tasks. A key update in this version is the integration of a vision encoder, allowing it to process and reason based on multimodal inputs, including images.

Key Capabilities

  • Advanced Reasoning: Capable of generating long chains of reasoning traces before providing an answer, encapsulated by [THINK] and [/THINK] special tokens.
  • Multimodality: Features a vision encoder for analyzing images and reasoning from visual content, alongside text.
  • Multilingual Support: Supports dozens of languages, including English, French, German, Japanese, Chinese, and Arabic.
  • Extended Context Window: Offers a 128k context window, with good performance up to 40k tokens.
  • Improved Performance: Demonstrates significantly better performance compared to Magistral Small 1.1 across various benchmarks like AIME24, AIME25, GPQA Diamond, and Livecodebench.
  • Enhanced Output Quality: Provides better LaTeX and Markdown formatting, shorter answers for easy prompts, and reduced likelihood of infinite generation loops.

Good For

  • Applications requiring robust reasoning and problem-solving.
  • Multimodal tasks involving both text and image analysis.
  • Deployments on resource-constrained hardware like a single RTX 4090 or a 32GB RAM MacBook, once quantized.
  • Developers looking for an open-licensed model (Apache 2.0) for commercial and non-commercial use.