kairawal/Llama-3.2-1B-Instruct-PT-SynthDolly-1A-E8

TEXT GENERATIONConcurrency Cost:1Model Size:1BQuant:BF16Ctx Length:32kPublished:Apr 6, 2026License:apache-2.0Architecture:Transformer Open Weights Cold

The kairawal/Llama-3.2-1B-Instruct-PT-SynthDolly-1A-E8 is a 1 billion parameter Llama-3.2-Instruct model developed by kairawal. This model was fine-tuned using Unsloth and Huggingface's TRL library, enabling faster training. It is designed for instruction-following tasks, leveraging its Llama-3.2 base architecture and efficient fine-tuning process. The model has a context length of 32768 tokens, making it suitable for applications requiring processing of moderately long inputs.

Loading preview...

Model Overview

The kairawal/Llama-3.2-1B-Instruct-PT-SynthDolly-1A-E8 is a 1 billion parameter instruction-tuned language model. It is based on the Llama-3.2-Instruct architecture and was developed by kairawal. A key characteristic of this model is its efficient training process, which utilized Unsloth and Huggingface's TRL library, resulting in a 2x faster fine-tuning compared to standard methods.

Key Capabilities

  • Instruction Following: Fine-tuned to understand and execute instructions effectively.
  • Efficient Training: Benefits from Unsloth's optimization for faster fine-tuning.
  • Llama-3.2 Base: Inherits the foundational capabilities of the Llama-3.2 architecture.
  • Context Length: Supports a context window of 32768 tokens, allowing for processing of substantial input lengths.

Good For

  • Applications requiring a compact yet capable instruction-following model.
  • Scenarios where efficient deployment and inference of a 1 billion parameter model are crucial.
  • Developers looking for a Llama-3.2 based model that has undergone optimized fine-tuning.