kyujinpy/Sakura-SOLRCA-Instruct-DPO
TEXT GENERATIONConcurrency Cost:1Model Size:10.7BQuant:FP8Ctx Length:4kPublished:Dec 24, 2023License:cc-by-nc-sa-4.0Architecture:Transformer Open Weights Warm
Sakura-SOLRCA-Instruct-DPO is a 10.7 billion parameter instruction-tuned causal language model developed by Kyujin Han and the LLM research consortium of Media Group Saramgwasup and Marker. This model, fine-tuned using the DPO method on the Intel/orca_dpo_pairs dataset, demonstrates strong performance across various benchmarks, achieving an average score of 74.05 on the Open LLM Leaderboard. It is designed for general-purpose instruction following and reasoning tasks, offering competitive capabilities for its size.
Loading preview...
Popular Sampler Settings
Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.
temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p