alexgusevski/OpenThinker2-32B-mlx-fp16
TEXT GENERATIONConcurrency Cost:2Model Size:32.8BQuant:FP8Ctx Length:32kPublished:Apr 6, 2025License:apache-2.0Architecture:Transformer Open Weights Cold
OpenThinker2-32B-mlx-fp16 is a 32.8 billion parameter language model, converted by alexgusevski to the MLX format from the original open-thoughts/OpenThinker2-32B. This model is specifically designed for efficient deployment and inference on Apple Silicon using the MLX framework, offering a context length of 131072 tokens. Its primary use case is general-purpose language generation and understanding, leveraging the performance benefits of MLX for local execution.
Loading preview...