meshllm/llama-3.2-1b-instruct-parity-bf16-mlx

TEXT GENERATIONConcurrency Cost:1Model Size:1BQuant:BF16Ctx Length:32kPublished:Apr 6, 2026License:llama3.2Architecture:Transformer Cold

The meshllm/llama-3.2-1b-instruct-parity-bf16-mlx model is a 1 billion parameter instruction-tuned Llama 3.2 variant, converted to the MLX bf16 format. Developed by meshllm, this model serves as a parity testing artifact for the mesh-llm ecosystem. It is specifically designed to validate consistent behavior against its f16 GGUF counterpart, making it suitable for MLX-based development and testing workflows.

Loading preview...

Overview

meshllm/llama-3.2-1b-instruct-parity-bf16-mlx is a 1 billion parameter instruction-tuned language model based on the Llama 3.2 architecture. This specific version is an MLX artifact, converted from the meta-llama/Llama-3.2-1B-Instruct checkpoint into the bf16 (bfloat16) format. Its primary purpose is to serve as a parity testing model within the mesh-llm ecosystem.

Key Characteristics

  • Architecture: Llama 3.2
  • Parameter Count: 1 billion (1B)
  • Format: MLX bf16 (bfloat16)
  • Origin: Converted from meta-llama/Llama-3.2-1B-Instruct
  • Context Length: 32768 tokens

Intended Use and Validation

This model is explicitly designed for parity testing against its intended pair, meshllm/llama-3.2-1b-instruct-parity-f16-gguf. It has been validated using the mesh-llm exact smoke suite, confirming expected behavior including shared strict lowercase drift on specific prompts (blue/green/red) and clean agreement on other exact prompts. Developers can use this model to ensure consistent output and behavior across different quantization and framework conversions within the mesh-llm framework, particularly for MLX-based applications.