allenai/tulu-2-dpo-13b
TEXT GENERATIONConcurrency Cost:1Model Size:13BQuant:FP8Ctx Length:4kPublished:Nov 13, 2023License:ai2-impact-license-low-riskArchitecture:Transformer0.0K Warm

allenai/tulu-2-dpo-13b is a 13 billion parameter language model developed by AllenAI, fine-tuned from Llama 2 using Direct Preference Optimization (DPO). It is designed as a helpful assistant, excelling in chat-based interactions and offering a strong alternative to Llama 2 13B Chat. This model demonstrates enhanced alignment and performance on benchmarks like MT-Bench and AlpacaEval.

Loading preview...

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p