Nexusflow/Starling-LM-7B-beta
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Mar 19, 2024License:apache-2.0Architecture:Transformer0.3K Open Weights Warm
Starling-LM-7B-beta is a 7 billion parameter language model developed by The Nexusflow Team, fine-tuned from Openchat-3.5-0106 (based on Mistral-7B-v0.1) using Reinforcement Learning from AI Feedback (RLAIF). This model leverages a new reward model, Starling-RM-34B, and the Nectar ranking dataset to achieve an improved 8.12 score on MT Bench with GPT-4 as a judge. It is optimized for generating helpful and harmless responses, making it suitable for general conversational AI applications.
Loading preview...