jondurbin/bagel-7b-v0.4
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Feb 4, 2024License:apache-2.0Architecture:Transformer0.0K Open Weights Cold

The jondurbin/bagel-7b-v0.4 is a 7 billion parameter Mistral-7b based language model fine-tuned by jondurbin. This pre-DPO version is optimized for diverse tasks including roleplay, complex reasoning, code generation, and function calling, leveraging a wide array of SFT datasets. It features a unique multi-format prompt training approach to enhance generalization across various instruction types, making it versatile for applications requiring nuanced understanding and structured outputs.

Loading preview...