SanjiWatsuki/Lelantos-DPO-7B
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Jan 12, 2024License:cc-by-nc-4.0Architecture:Transformer0.0K Open Weights Cold

SanjiWatsuki/Lelantos-DPO-7B is a 7 billion parameter language model developed by SanjiWatsuki, fine-tuned using Direct Preference Optimization (DPO). This model demonstrates strong performance across various benchmarks, achieving an average score of 58.54% across AGIEval, GPT4All, TruthfulQA, and Bigbench. It is particularly suited for general-purpose language understanding and generation tasks where robust performance on common reasoning and knowledge-based evaluations is critical.

Loading preview...