mlabonne/UltraMerge-7B
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:8kPublished:Mar 21, 2024License:cc-by-nc-4.0Architecture:Transformer0.0K Open Weights Cold

UltraMerge-7B is an experimental 7 billion parameter DPO fine-tune of automerger/YamShadow-7B, developed by mlabonne. This model is trained on a diverse set of DPO datasets including mlabonne/truthy-dpo-v0.1 and mlabonne/ultrafeedback-binarized-preferences-cleaned, making it suitable for general-purpose conversational AI tasks. It features an 8192 token context length, offering robust performance for applications requiring extended conversational memory.

Loading preview...

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p