lmstudio-community/magistral-small-2506-mlx-bf16

Warm
Public
24B
FP8
32768
License: apache-2.0
Hugging Face
Overview

Model Overview

The magistral-small model, developed by mistral-ai, is a 24 billion parameter language model. This particular version is an MLX quantization of the original bfloat16 magistral-small model, specifically optimized for efficient execution on Apple Silicon.

Key Characteristics

  • Creator: mistral-ai
  • Original Model: magistral-small (bfloat16 version)
  • Quantization: MLX, provided by the LM Studio team using mlx_lm.
  • Hardware Optimization: Designed for Apple Silicon, leveraging the MLX framework developed by the Apple Machine Learning Research team.

Use Cases

This model is particularly well-suited for developers and researchers who:

  • Require a powerful 24 billion parameter language model.
  • Are working within the Apple ecosystem and need models optimized for Apple Silicon.
  • Seek efficient local inference capabilities on their Apple hardware, benefiting from the MLX framework's performance advantages.