kairawal/Llama-3.2-3B-Instruct-DA-SynthDolly-1A-E1

TEXT GENERATIONConcurrency Cost:1Model Size:3.2BQuant:BF16Ctx Length:32kPublished:Apr 9, 2026License:apache-2.0Architecture:Transformer Open Weights Cold

The kairawal/Llama-3.2-3B-Instruct-DA-SynthDolly-1A-E1 is a 3.2 billion parameter instruction-tuned Llama model developed by kairawal. Finetuned from unsloth/llama-3.2-3b-Instruct, this model was trained using Unsloth and Huggingface's TRL library for accelerated performance. It is designed for general instruction-following tasks, leveraging its efficient training methodology.

Loading preview...

Model Overview

The kairawal/Llama-3.2-3B-Instruct-DA-SynthDolly-1A-E1 is a 3.2 billion parameter instruction-tuned language model. Developed by kairawal, this model is finetuned from the unsloth/llama-3.2-3b-Instruct base model.

Key Characteristics

  • Efficient Training: This model was trained significantly faster using the Unsloth library in conjunction with Huggingface's TRL library, indicating an optimization for training speed and resource utilization.
  • Instruction-Tuned: As an instruction-tuned model, it is designed to follow natural language instructions effectively, making it suitable for a variety of conversational and task-oriented applications.
  • Llama Architecture: Built upon the Llama 3.2 architecture, it benefits from the foundational capabilities of this model family.
  • License: The model is released under the Apache 2.0 license, allowing for broad use and distribution.

Good For

  • General Instruction Following: Ideal for applications requiring a model to understand and execute commands given in natural language.
  • Resource-Efficient Deployment: Its 3.2 billion parameter size, combined with efficient training, suggests it could be suitable for scenarios where computational resources are a consideration.
  • Experimentation with Unsloth: Developers interested in models trained with Unsloth for speed and efficiency may find this a useful reference or starting point.