activeDap/gemma-2b_hh_helpful

Loading
Public
2.5B
BF16
8192
License: apache-2.0
Hugging Face
Overview

Model Overview

activeDap/gemma-2b_hh_helpful is a 2.5 billion parameter language model derived from Google's Gemma-2b. It has undergone Supervised Fine-Tuning (SFT) using the activeDap/sft-hh-data dataset, which focuses on helpfulness. The fine-tuning process involved 20 steps, achieving a final training loss of 2.1454 over 9.15 seconds.

Key Capabilities

  • Instruction Following: Fine-tuned to generate responses that align with user instructions, particularly in a helpful assistant style.
  • Conversational AI: Optimized for prompt-completion tasks, making it suitable for dialogue systems.
  • Efficient Inference: As a 2.5B parameter model, it offers a balance between performance and computational efficiency.

Training Details

The model was trained for 1 epoch with a per-device batch size of 16 (total batch size 64 across 4 GPUs) and a learning rate of 2e-05. It utilized a cosine learning rate scheduler, BF16 mixed precision, and a maximum sequence length of 512 tokens. The training focused on assistant-only loss, ensuring the model learns to generate relevant and helpful outputs.

Good For

  • Developing helpful AI assistants.
  • Instruction-based text generation.
  • Applications requiring a smaller, yet capable, fine-tuned language model for conversational tasks.