bardsai/jaskier-7b-dpo
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:8kPublished:Jan 10, 2024License:apache-2.0Architecture:Transformer Open Weights Cold
bardsai/jaskier-7b-dpo is a 7 billion parameter language model developed by bards.ai, fine-tuned from mindy-labs/mindy-7b-v2 (a Mistral7B derivative) using Direct Preference Optimization (DPO). This model is specifically trained on the Intel/orca_dpo_pairs dataset, aiming to enhance conversational quality and alignment. With an 8192-token context length, it is designed for general conversational applications, particularly those benefiting from DPO-tuned responses.
Loading preview...
Popular Sampler Settings
Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.
temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p