jondurbin/bagel-dpo-7b-v0.5
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Apr 15, 2024License:apache-2.0Architecture:Transformer0.0K Open Weights Cold

The jondurbin/bagel-dpo-7b-v0.5 is a 7 billion parameter language model developed by jondurbin, fine-tuned from Mistral-7B-v0.2. This model incorporates a Direct Preference Optimization (DPO) pass using the bagel v0.5 dataset, enhancing its ability to follow instructions and generate nuanced responses. It is particularly optimized for diverse tasks including complex question answering, summarization, function calling, and creative writing, supporting a 4096 token context length.

Loading preview...