allenai/tulu-v2.5-dpo-13b-stackexchange
TEXT GENERATIONConcurrency Cost:1Model Size:13BQuant:FP8Ctx Length:4kPublished:Jun 11, 2024License:apache-2.0Architecture:Transformer Open Weights Cold

The allenai/tulu-v2.5-dpo-13b-stackexchange is a 13 billion parameter language model developed by AllenAI, fine-tuned from Llama-2-13b-hf. This model is specifically trained using DPO on 500k samples from the StackExchange paired dataset, optimizing it for helpful assistant-like responses. It is part of the Tulu V2.5 series, focusing on learning from preference feedback to enhance conversational quality.

Loading preview...