jieliu/Storm-7B
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Apr 25, 2024License:apache-2.0Architecture:Transformer0.0K Open Weights Cold

Storm-7B is a 7 billion parameter language model developed by Jie Liu, Zhanhui Zhou, Jiaheng Liu, Xingyuan Bu, Chao Yang, Han-Sen Zhong, and Wanli Ouyang, fine-tuned from openchat-3.5-0106. It utilizes iterative length-regularized Direct Preference Optimization (iLR-DPO) to achieve performance comparable to GPT-4 on AlpacaEval 2.0, specifically optimized to improve response quality without increasing verbosity. This model is designed for high-quality, concise conversational AI applications, demonstrating a 50.5% length-controlled win rate against GPT-4 Preview.

Loading preview...