dphn/dolphin-2.6-mistral-7b-dpo-laser
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Jan 1, 2024License:apache-2.0Architecture:Transformer0.1K Open Weights Cold

The dphn/dolphin-2.6-mistral-7b-dpo-laser is a 7 billion parameter language model developed by @ehartford and @fernandofernandes, based on the Mistral-7b architecture with a 4096 token context length. This model incorporates a specialized DPO (Direct Preference Optimization) training method enhanced by a novel LASER (Layer-Selective Rank Reduction) technique, utilizing Random Matrix Theory for optimal rank calculation. It is designed to provide robust and highly compliant outputs, demonstrating improved performance over previous Dolphin-DPO versions on benchmarks like MMLU and GSM-8k.

Loading preview...