ailexleon/WeirdCompound-v1.7-24b-mlx-fp16

TEXT GENERATIONConcurrency Cost:2Model Size:24BQuant:FP8Ctx Length:32kPublished:Dec 25, 2025Architecture:Transformer Cold

The ailexleon/WeirdCompound-v1.7-24b-mlx-fp16 model is a 24 billion parameter language model converted to the MLX format, originally from FlareRebellion/WeirdCompound-v1.7-24b. This model is optimized for efficient deployment and inference on Apple silicon using the MLX framework, supporting a context length of 32768 tokens. Its primary use case is general language generation and understanding tasks within the MLX ecosystem.

Loading preview...

Overview

The ailexleon/WeirdCompound-v1.7-24b-mlx-fp16 is a 24 billion parameter language model, specifically a conversion of the original FlareRebellion/WeirdCompound-v1.7-24b model into the MLX format. This conversion was performed using mlx-lm version 0.28.3, making it suitable for efficient inference on Apple silicon.

Key Capabilities

  • MLX Compatibility: Fully optimized for the MLX framework, enabling high-performance execution on Apple devices.
  • Large Parameter Count: With 24 billion parameters, it offers robust language understanding and generation capabilities.
  • Extensive Context Window: Supports a context length of 32768 tokens, allowing for processing and generating longer texts.

Good for

  • Developers working with Apple silicon who require a powerful, locally runnable language model.
  • Applications demanding a large context window for complex tasks like summarization, content creation, or detailed question answering.
  • Experimentation and deployment of large language models within the MLX ecosystem.