HenryJJ/dolphin-2.6-mistral-7b-dpo-orca
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Jan 12, 2024License:apache-2.0Architecture:Transformer Open Weights Cold
HenryJJ/dolphin-2.6-mistral-7b-dpo-orca is a 7 billion parameter auto-regressive language model, fine-tuned by HenryJJ using DPO (Direct Preference Optimization) from cognitivecomputations/dolphin-2.6-mistral-7b. It leverages the Mistral 7B architecture and was trained on the Intel/orca_dpo_pairs dataset for 1200 steps with a 1024 token context window. This model is designed for English language tasks, focusing on instruction following and chat applications, utilizing the ChatML prompt format.
Loading preview...