nbeerbower/gemma2-gutenberg-27B

Warm
Public
27B
FP8
32768
2
License: gemma
Hugging Face
Overview

Model Overview

nbeerbower/gemma2-gutenberg-27B is a 27 billion parameter instruction-tuned language model built upon the robust Google Gemma-2-27B-IT architecture. This model distinguishes itself through its specific fine-tuning process, which involved applying the ORPO (Odds Ratio Preference Optimization) method. The training was conducted over three epochs using an 80GB A100 GPU, leveraging the jondurbin/gutenberg-dpo-v0.1 dataset.

Key Capabilities

  • Instruction Following: Optimized to understand and execute complex instructions effectively.
  • High-Quality Text Generation: Benefits from DPO-style training on a curated dataset, enhancing output quality.
  • Gemma-2 Foundation: Inherits the strong base capabilities of the Gemma-2 series, including a 32K context length.

Good For

  • Applications requiring precise and coherent text generation based on user prompts.
  • Tasks where instruction adherence and output quality are paramount.
  • Developers looking for a Gemma-2 variant with enhanced DPO-tuned performance.