shabul/qwen2.5-3b-dolly-finetuned
The shabul/qwen2.5-3b-dolly-finetuned model is a 3.1 billion parameter Qwen2.5-3B-Instruct model, fine-tuned by Shabul Abdul using LoRA on the Dolly-15k dataset. This fine-tune adapts the base model's output style towards more direct, concise, and human-annotated responses. It is particularly optimized for tasks like brainstorming, open question answering, summarization, and creative writing, offering a lightweight instruction-following capability.
Loading preview...
What is shabul/qwen2.5-3b-dolly-finetuned?
This model is a LoRA fine-tune of the Qwen/Qwen2.5-3B-Instruct base model, developed by Shabul Abdul. It was trained on the databricks/databricks-dolly-15k dataset, which consists of 15,000 human-written instruction/response pairs covering a wide range of tasks including brainstorming, classification, QA, summarization, and creative writing.
Key Characteristics & Training
- Base Model:
Qwen/Qwen2.5-3B-Instruct(3.1 billion parameters). - Fine-tuning Method: LoRA (rank 8, alpha 16), updating only 0.216% of the model's weights.
- Dataset:
databricks/databricks-dolly-15k, with 13,500 training and 1,500 validation examples. - Hardware: Trained on an Apple M5 MacBook Pro with 24 GB unified memory using Apple MLX, completing in approximately 13 minutes.
- Performance: Achieved a 46.9% reduction in validation loss (from 2.725 to 1.446) during training.
What makes this model different?
This fine-tune specifically adjusts the output style of the already capable Qwen2.5-3B-Instruct model. It aims for:
- Tone: More direct and less verbose responses.
- Length: Calibrated to match the average response length found in the Dolly dataset.
- Structure: Favors plain paragraphs over markdown-heavy formatting.
- Efficiency: Achieves significant style adaptation without altering 99.784% of the base model's weights, making it a lightweight modification.
Should you use this model?
This model is suitable for use cases requiring instruction-following with a preference for concise, direct, and human-like responses. It excels in:
- Brainstorming
- Open Question Answering
- Summarization
- Creative Writing
For highly domain-specific tasks, the fine-tuning pipeline can be easily adapted with custom datasets.