alexgusevski/Llama3.3-8B-Instruct-Thinking-Claude-4.5-Opus-High-Reasoning-mlx-fp16

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:8kPublished:Jan 12, 2026License:apache-2.0Architecture:Transformer Open Weights Warm

The alexgusevski/Llama3.3-8B-Instruct-Thinking-Claude-4.5-Opus-High-Reasoning-mlx-fp16 model is an 8 billion parameter instruction-tuned language model, converted to MLX format from DavidAU's Llama3.3-8B-Instruct-Thinking-Claude-4.5-Opus-High-Reasoning. This model is designed for high-reasoning tasks, leveraging the Llama3.3 architecture and fine-tuning inspired by Claude 4.5 Opus. It is optimized for local deployment on Apple Silicon via MLX, making it suitable for advanced instruction-following and complex problem-solving.

Loading preview...

Model Overview

The alexgusevski/Llama3.3-8B-Instruct-Thinking-Claude-4.5-Opus-High-Reasoning-mlx-fp16 is an 8 billion parameter instruction-tuned language model. It is a conversion of the DavidAU/Llama3.3-8B-Instruct-Thinking-Claude-4.5-Opus-High-Reasoning model into the MLX format, specifically using mlx-lm version 0.29.1.

Key Characteristics

  • Architecture: Based on the Llama3.3 family, providing a robust foundation for language understanding and generation.
  • Parameter Count: Features 8 billion parameters, balancing performance with computational efficiency.
  • Fine-tuning: The original model was fine-tuned with a focus on "Thinking" and "High-Reasoning," drawing inspiration from Claude 4.5 Opus, suggesting enhanced capabilities in complex problem-solving and logical inference.
  • MLX Conversion: Optimized for Apple Silicon (Macs with M-series chips) through its MLX format, enabling efficient local inference.

Intended Use Cases

This model is particularly well-suited for:

  • Advanced Instruction Following: Excelling in tasks that require precise and nuanced responses to user instructions.
  • Reasoning-Heavy Applications: Ideal for scenarios demanding strong logical deduction, analysis, and problem-solving abilities.
  • Local Deployment: Its MLX format makes it a strong candidate for developers looking to run high-performance LLMs directly on Apple Silicon hardware.