abideen/gemma-7b-openhermes

TEXT GENERATIONConcurrency Cost:1Model Size:8.5BQuant:FP8Ctx Length:8kPublished:Feb 21, 2024License:cc-by-nc-4.0Architecture:Transformer0.0K Open Weights Cold

The abideen/gemma-7b-openhermes is an 8.5 billion parameter language model, a fine-tuned variant of Google's Gemma 7B. It was further trained using QLoRA on the OpenHermes-2.5 preference dataset, enhancing its conversational capabilities. This model is designed for generating English-language text in response to various inputs, such as questions or prompts, making it suitable for chat-based applications.

Loading preview...

Model Overview

abideen/gemma-7b-openhermes is an 8.5 billion parameter language model derived from Google's Gemma 7B. It has undergone further fine-tuning using QLoRA on the OpenHermes-2.5 preference dataset, which is designed to improve conversational quality and alignment. The model utilizes a specific chat template for conversational use, which can be applied via the tokenizer's built-in functionality.

Key Capabilities

  • Conversational AI: Optimized for generating coherent and contextually relevant responses in chat-based interactions.
  • Text Generation: Capable of producing English-language text from diverse inputs like questions, prompts, or documents.
  • Fine-tuned Performance: Leverages the OpenHermes-2.5 dataset to enhance its ability to follow instructions and engage in dialogue.

Evaluation Highlights

Evaluations across various benchmarks indicate its performance:

  • Nous Benchmark: Achieved an average of 22.29 on Agieval and 32.00 on GPT4ALL tasks, with a TruthfulQA average of 38.90.
  • OpenLLM Benchmark: Demonstrated an average of 73.5% across tasks including arc_challenge, hellaswag, gsm8k, winogrande, and mmlu, with mmlu at 53.62% accuracy.

Training Details

The model was trained with a learning rate of 5e-07, a total batch size of 8, and 1000 training steps, utilizing an Adam optimizer and a cosine learning rate scheduler. It was built with Axolotl, indicating a focus on efficient fine-tuning practices.