pmking27/PrathameshLLM-2B

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:2.6BQuant:BF16Ctx Length:8kLicense:apache-2.0Architecture:Transformer0.0K Open Weights Warm

pmking27/PrathameshLLM-2B is a 2.6 billion parameter causal language model developed by pmking27, fine-tuned from Google's Gemma-2B architecture. This model was trained using Unsloth and Huggingface's TRL library, achieving 2x faster training. It is designed for instruction-following tasks, demonstrated through its use with an Alpaca prompt template for question answering based on provided context.

Loading preview...

pmking27/PrathameshLLM-2B: A Faster-Trained Gemma Model

pmking27/PrathameshLLM-2B is a 2.6 billion parameter language model developed by pmking27, built upon the google/gemma-2b architecture. A key differentiator for this model is its training methodology: it was fine-tuned significantly faster using the Unsloth library in conjunction with Huggingface's TRL library.

Key Capabilities

  • Instruction Following: The model is demonstrated to follow instructions effectively, particularly when provided with a context for question answering, using an Alpaca-style prompt format.
  • Contextual Question Answering: It can process given textual contexts and generate relevant answers to questions posed in various languages, as shown with a Marathi language example.
  • Efficient Training: Leveraging Unsloth, this model achieved a 2x speedup during its fine-tuning process, indicating potential for rapid iteration and deployment.

Good For

  • Developers seeking a compact, instruction-tuned model: Its 2.6B parameter size makes it suitable for applications where computational resources are a consideration.
  • Applications requiring contextual information extraction: The model's ability to answer questions based on provided context makes it useful for tasks like summarization, information retrieval, and chatbot development.
  • Experimentation with faster fine-tuning techniques: The use of Unsloth highlights its potential for researchers and developers interested in optimizing LLM training workflows.