gjdeboer/AgentFlow_Slime_Agentic_Qwen2.5_7B-mlx-fp16

TEXT GENERATIONConcurrency Cost:1Model Size:7.6BQuant:FP8Ctx Length:32kPublished:Mar 13, 2026Architecture:Transformer Cold

The gjdeboer/AgentFlow_Slime_Agentic_Qwen2.5_7B-mlx-fp16 model is a 7.6 billion parameter Qwen2.5 architecture, converted to the MLX format for efficient deployment on Apple silicon. This model is derived from LMIS-ORG/AgentFlow_Slime_Agentic_Qwen2.5_7B and supports a 32,768 token context length. It is specifically designed for agentic applications, leveraging its Qwen2.5 base for robust language understanding and generation within an MLX environment.

Loading preview...

Model Overview

The gjdeboer/AgentFlow_Slime_Agentic_Qwen2.5_7B-mlx-fp16 is a 7.6 billion parameter language model based on the Qwen2.5 architecture. It has been specifically converted to the MLX format, enabling optimized performance on Apple silicon devices. This conversion was performed from the original LMIS-ORG/AgentFlow_Slime_Agentic_Qwen2.5_7B using mlx-lm version 0.29.1.

Key Characteristics

  • Architecture: Qwen2.5 base model, known for strong general-purpose language capabilities.
  • Parameter Count: 7.6 billion parameters, offering a balance between performance and computational efficiency.
  • Context Length: Supports a substantial context window of 32,768 tokens, suitable for complex, multi-turn interactions and detailed document processing.
  • MLX Format: Optimized for Apple silicon, providing efficient inference on compatible hardware.
  • Agentic Focus: The "Agentic" designation suggests fine-tuning or design considerations for tasks involving autonomous agents, tool use, or complex reasoning workflows.

Intended Use Cases

This model is particularly well-suited for:

  • Agent-based applications: Developing intelligent agents that require robust language understanding and generation.
  • Local inference on Apple silicon: Users with Apple hardware can leverage the MLX format for efficient local deployment.
  • Complex conversational AI: Its large context window makes it suitable for maintaining long-running dialogues and understanding intricate user requests.
  • Research and development: Experimenting with agentic workflows and Qwen2.5 capabilities in an MLX environment.