HenryJJ/dolphin-2.6-mistral-7b-dpo-orca-v1
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Jan 14, 2024License:apache-2.0Architecture:Transformer Open Weights Cold
HenryJJ/dolphin-2.6-mistral-7b-dpo-orca-v1 is a 7 billion parameter auto-regressive language model, fine-tuned by HenryJJ using DPO from cognitivecomputations/dolphin-2.6-mistral-7b. It was trained on the Intel/orca_dpo_pairs dataset for 1200 steps with a 1024-token context window. This model is optimized for instruction following and conversational tasks, leveraging the Mistral architecture for efficient performance.
Loading preview...