chujiezheng/Starling-LM-7B-beta-ExPO

Warm
Public
7B
FP8
4096
Apr 26, 2024
License: apache-2.0
Hugging Face
Overview

Starling-LM-7B-beta-ExPO Overview

This model, developed by chujiezheng, is a 7 billion parameter language model derived from Nexusflow/Starling-LM-7B-beta and openchat/openchat-3.5-0106. Its core innovation lies in the Extrapolated (ExPO) training method, as detailed in the "Weak-to-Strong Extrapolation Expedites Alignment" paper. By extrapolating with an alpha of 0.5 from SFT and DPO/RLHF checkpoints, the model aims for enhanced alignment with human preferences.

Key Capabilities & Performance

The Starling-LM-7B-beta-ExPO model demonstrates notable improvements in alignment and performance:

  • Superior Human Preference Alignment: Achieves higher win rates on the AlpacaEval 2.0 benchmark across various comparisons, indicating better alignment with human judgments.
  • Improved MT-Bench Scores: Shows increased scores on the MT-Bench benchmark, suggesting enhanced conversational and instruction-following abilities.
  • Extrapolation Method: Leverages a novel extrapolation technique to boost performance from existing SFT and DPO/RLHF checkpoints.

When to Use This Model

This model is particularly well-suited for applications requiring strong alignment with human preferences and robust performance in conversational and instruction-following tasks. Its improved benchmark scores suggest it can be a strong candidate for:

  • Chatbots and Conversational AI: Where nuanced understanding and human-like responses are critical.
  • Instruction Following: For tasks that benefit from models that accurately interpret and execute complex instructions.
  • Preference-Aligned Applications: Any use case where aligning with human feedback and preferences is a primary objective.