verque/qwen3-8b-karma-v3-mlx-fp16
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Feb 2, 2026License:apache-2.0Architecture:Transformer Open Weights Cold

The verque/qwen3-8b-karma-v3-mlx-fp16 is an 8 billion parameter language model, converted to MLX format from verque-app/qwen3-8b-karma-v3. This model is designed for efficient deployment and inference on Apple Silicon, leveraging the MLX framework. It provides a robust base for general language tasks, offering a balance between performance and resource utilization for local machine learning applications.

Loading preview...