abideen/AlphaMonarch-daser
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Mar 16, 2024License:cc-by-nc-4.0Architecture:Transformer0.0K Open Weights Cold
AlphaMonarch-daser is a 7 billion parameter language model developed by abideen, fine-tuned using a combination of LaserQlora and Dora techniques. This model is a DPO fine-tuned version of mlabonne/NeuralMonarch-7B, utilizing the argilla/OpenHermes2.5-dpo-binarized-alpha preference dataset. It demonstrates improved performance over AlphaMonarch-dora on the YALL leaderboard, despite being trained on only half of the projections. The model is optimized for general language tasks, leveraging its DPO fine-tuning for enhanced conversational and instruction-following capabilities.
Loading preview...