haji80mr-uoft/maj-semi-wtype-Llama-tuned-Lora-merged-gpt5

TEXT GENERATIONConcurrency Cost:1Model Size:3.2BQuant:BF16Ctx Length:32kPublished:Apr 16, 2026License:apache-2.0Architecture:Transformer Open Weights Cold

The haji80mr-uoft/maj-semi-wtype-Llama-tuned-Lora-merged-gpt5 is a 3.2 billion parameter Llama-based instruction-tuned model developed by haji80mr-uoft. This model was finetuned from unsloth/llama-3.2-3b-instruct-unsloth-bnb-4bit using Unsloth and Huggingface's TRL library, enabling faster training. It is designed for general instruction-following tasks, leveraging its Llama architecture and efficient finetuning process.

Loading preview...

Model Overview

The haji80mr-uoft/maj-semi-wtype-Llama-tuned-Lora-merged-gpt5 is a 3.2 billion parameter language model developed by haji80mr-uoft. It is based on the Llama architecture and was finetuned from the unsloth/llama-3.2-3b-instruct-unsloth-bnb-4bit model. A key characteristic of this model's development is its efficient training process, which was accelerated using Unsloth and Huggingface's TRL library, reportedly achieving 2x faster training.

Key Capabilities

  • Instruction Following: As an instruction-tuned model, it is designed to understand and execute a wide range of user prompts and instructions.
  • Efficient Training: Benefits from the Unsloth framework, which optimizes the finetuning process for speed.
  • Llama Architecture: Inherits the robust capabilities and general-purpose language understanding of the Llama model family.

Good For

  • Applications requiring a compact yet capable instruction-tuned model.
  • Scenarios where efficient deployment and inference of a 3.2B parameter model are crucial.
  • General text generation and conversational AI tasks that align with instruction-following paradigms.