gjyotin305/Llama-3.2-3B-Instruct_old_sft_alpaca_005
TEXT GENERATIONConcurrency Cost:1Model Size:3.2BQuant:BF16Ctx Length:32kPublished:Jan 9, 2026License:apache-2.0Architecture:Transformer Open Weights Warm
gjyotin305/Llama-3.2-3B-Instruct_old_sft_alpaca_005 is a 3.2 billion parameter instruction-tuned Llama-3.2 model developed by gjyotin305, fine-tuned from unsloth/Llama-3.2-3B-Instruct. This model was trained using Unsloth and Huggingface's TRL library, enabling faster fine-tuning. It is designed for general instruction-following tasks, leveraging its Llama-3.2 architecture and 32768 token context length.
Loading preview...
Model Overview
gjyotin305/Llama-3.2-3B-Instruct_old_sft_alpaca_005 is an instruction-tuned language model based on the Llama-3.2 architecture, developed by gjyotin305. This model, with approximately 3.2 billion parameters and a 32768 token context length, was fine-tuned from the unsloth/Llama-3.2-3B-Instruct base model.
Key Characteristics
- Architecture: Llama-3.2-3B, providing a robust foundation for language understanding and generation.
- Fine-tuning: Utilizes the Unsloth library in conjunction with Huggingface's TRL library, which facilitated a 2x faster training process.
- Instruction-Tuned: Optimized for following instructions and engaging in conversational tasks, making it suitable for various NLP applications.
- License: Distributed under the Apache-2.0 license, allowing for broad usage and modification.
Potential Use Cases
- General Instruction Following: Capable of responding to a wide range of prompts and instructions.
- Chatbots and Conversational AI: Its instruction-tuned nature makes it suitable for dialogue systems.
- Text Generation: Can be used for creative writing, summarization, and other text generation tasks where a compact yet capable model is desired.