HenryJJ/dolphin-2.6-mistral-7b-dpo-orca-v3
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Jan 14, 2024License:apache-2.0Architecture:Transformer Open Weights Cold
HenryJJ/dolphin-2.6-mistral-7b-dpo-orca-v3 is a 7 billion parameter auto-regressive language model, fine-tuned by HenryJJ using DPO from cognitivecomputations/dolphin-2.6-mistral-7b. It was trained for 1200 steps on the Intel/orca_dpo_pairs dataset with a 1024 context window. This model is optimized for instruction following and conversational tasks, utilizing a ChatML prompt format.
Loading preview...