HenryJJ/dolphin-2.6-mistral-7b-dpo-orca-v2
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Jan 14, 2024License:apache-2.0Architecture:Transformer Open Weights Cold

HenryJJ/dolphin-2.6-mistral-7b-dpo-orca-v2 is a 7 billion parameter auto-regressive language model, fine-tuned by HenryJJ using DPO (Direct Preference Optimization) on the Intel/orca_dpo_pairs dataset. Based on the Llama 2 transformer architecture, this model is optimized for instruction following and conversational tasks, leveraging a 4096 token context window. It is designed to provide helpful and compliant AI assistance, trained for 1200 steps.

Loading preview...