alexgusevski/CapybaraHermes-2.5-Mistral-7B-mlx-fp16
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Jan 12, 2026License:apache-2.0Architecture:Transformer Open Weights Cold
The alexgusevski/CapybaraHermes-2.5-Mistral-7B-mlx-fp16 model is a 7 billion parameter language model, converted to the MLX format from argilla/CapybaraHermes-2.5-Mistral-7B. This model leverages the Mistral architecture and is specifically designed for efficient inference on Apple silicon via the MLX framework. It maintains a 4096-token context length, making it suitable for general-purpose language generation and understanding tasks within the MLX ecosystem.
Loading preview...