akshayballal/Qwen2.5-7B-Instruct-SFT-Pubmed-16bit-DFT
akshayballal/Qwen2.5-7B-Instruct-SFT-Pubmed-16bit-DFT is a 7.6 billion parameter instruction-tuned causal language model developed by akshayballal. This model is a fine-tuned variant of Qwen2.5-7B-Instruct, specifically optimized for tasks related to PubMed content. It was trained using Unsloth and Huggingface's TRL library, enabling faster fine-tuning.
Loading preview...
Overview
akshayballal/Qwen2.5-7B-Instruct-SFT-Pubmed-16bit-DFT is a 7.6 billion parameter instruction-tuned language model. It is a specialized fine-tune of the Qwen2.5-7B-Instruct base model, developed by akshayballal.
Key Characteristics
- Base Model: Fine-tuned from unsloth/qwen2.5-7b-instruct-unsloth-bnb-4bit.
- Training Efficiency: Leverages Unsloth and Huggingface's TRL library for accelerated training, achieving 2x faster fine-tuning.
- Context Length: Supports a substantial context window of 131,072 tokens.
Primary Use Case
This model is specifically fine-tuned for applications involving PubMed content, suggesting its strength in biomedical and scientific text understanding and generation. Developers can utilize this model for tasks requiring deep knowledge or processing of medical literature.