pmking27/PrathameshLLM-2B
TEXT GENERATIONConcurrency Cost:1Model Size:2.6BQuant:BF16Ctx Length:8kLicense:apache-2.0Architecture:Transformer0.0K Open Weights Warm
pmking27/PrathameshLLM-2B is a 2.6 billion parameter causal language model developed by pmking27, fine-tuned from Google's Gemma-2B architecture. This model was trained using Unsloth and Huggingface's TRL library, achieving 2x faster training. It is designed for instruction-following tasks, demonstrated through its use with an Alpaca prompt template for question answering based on provided context.
Loading preview...
pmking27/PrathameshLLM-2B: A Faster-Trained Gemma Model
pmking27/PrathameshLLM-2B is a 2.6 billion parameter language model developed by pmking27, built upon the google/gemma-2b architecture. A key differentiator for this model is its training methodology: it was fine-tuned significantly faster using the Unsloth library in conjunction with Huggingface's TRL library.
Key Capabilities
- Instruction Following: The model is demonstrated to follow instructions effectively, particularly when provided with a context for question answering, using an Alpaca-style prompt format.
- Contextual Question Answering: It can process given textual contexts and generate relevant answers to questions posed in various languages, as shown with a Marathi language example.
- Efficient Training: Leveraging Unsloth, this model achieved a 2x speedup during its fine-tuning process, indicating potential for rapid iteration and deployment.
Good For
- Developers seeking a compact, instruction-tuned model: Its 2.6B parameter size makes it suitable for applications where computational resources are a consideration.
- Applications requiring contextual information extraction: The model's ability to answer questions based on provided context makes it useful for tasks like summarization, information retrieval, and chatbot development.
- Experimentation with faster fine-tuning techniques: The use of Unsloth highlights its potential for researchers and developers interested in optimizing LLM training workflows.