chujiezheng/Starling-LM-7B-beta-ExPO

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Apr 26, 2024License:apache-2.0Architecture:Transformer0.0K Open Weights Warm

The chujiezheng/Starling-LM-7B-beta-ExPO is a 7 billion parameter language model based on Nexusflow/Starling-LM-7B-beta and openchat/openchat-3.5-0106, utilizing an extrapolated (ExPO) training method. This model achieves superior alignment with human preference by extrapolating from SFT and DPO/RLHF checkpoints. It demonstrates improved win rates on the AlpacaEval 2.0 benchmark and higher MT-Bench scores compared to its base models and other 7B-class LLMs.

Loading preview...

Starling-LM-7B-beta-ExPO Overview

This model, developed by chujiezheng, is a 7 billion parameter language model derived from Nexusflow/Starling-LM-7B-beta and openchat/openchat-3.5-0106. Its core innovation lies in the Extrapolated (ExPO) training method, as detailed in the "Weak-to-Strong Extrapolation Expedites Alignment" paper. By extrapolating with an alpha of 0.5 from SFT and DPO/RLHF checkpoints, the model aims for enhanced alignment with human preferences.

Key Capabilities & Performance

The Starling-LM-7B-beta-ExPO model demonstrates notable improvements in alignment and performance:

  • Superior Human Preference Alignment: Achieves higher win rates on the AlpacaEval 2.0 benchmark across various comparisons, indicating better alignment with human judgments.
  • Improved MT-Bench Scores: Shows increased scores on the MT-Bench benchmark, suggesting enhanced conversational and instruction-following abilities.
  • Extrapolation Method: Leverages a novel extrapolation technique to boost performance from existing SFT and DPO/RLHF checkpoints.

When to Use This Model

This model is particularly well-suited for applications requiring strong alignment with human preferences and robust performance in conversational and instruction-following tasks. Its improved benchmark scores suggest it can be a strong candidate for:

  • Chatbots and Conversational AI: Where nuanced understanding and human-like responses are critical.
  • Instruction Following: For tasks that benefit from models that accurately interpret and execute complex instructions.
  • Preference-Aligned Applications: Any use case where aligning with human feedback and preferences is a primary objective.