jondurbin/bagel-dpo-7b-v0.4
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Feb 3, 2024License:apache-2.0Architecture:Transformer0.0K Open Weights Cold

jondurbin/bagel-dpo-7b-v0.4 is a 7 billion parameter language model developed by jondurbin, fine-tuned from Mistral-7B-v0.1 using Direct Preference Optimization (DPO). This model is optimized for diverse instruction following, creative writing, and specialized tasks like function calling and contextual question answering, leveraging a broad array of SFT and DPO datasets. It supports multiple prompt formats including Llama-2, Alpaca, Vicuna, and ChatML, making it adaptable for various conversational and generative AI applications.

Loading preview...