ewqr2130/alignment-handbook-zephyr-7b-sft-full-dpo-5e7-cont1
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Jan 15, 2024License:apache-2.0Architecture:Transformer Open Weights Cold
The ewqr2130/alignment-handbook-zephyr-7b-sft-full-dpo-5e7-cont1 is a 7 billion parameter language model developed by ewqr2130, featuring a 4096-token context length. This model is a continuation of the Zephyr-7B-SFT series, likely fine-tuned for specific alignment tasks using Direct Preference Optimization (DPO). Its primary strength lies in its specialized alignment, making it suitable for applications requiring nuanced and controlled text generation.
Loading preview...