allenai/tulu-v2.5-dpo-13b-stackexchange-60k
TEXT GENERATIONConcurrency Cost:1Model Size:13BQuant:FP8Ctx Length:4kPublished:Jun 11, 2024License:apache-2.0Architecture:Transformer0.0K Open Weights Cold

The allenai/tulu-v2.5-dpo-13b-stackexchange-60k model is a 13 billion parameter language model developed by AllenAI, fine-tuned from meta-llama/Llama-2-13b-hf. It is part of the Tulu V2.5 series, specifically trained using Direct Preference Optimization (DPO) on a 60k subsample of the StackExchange paired dataset. This model is designed to act as a helpful assistant, specializing in generating responses aligned with user preferences, particularly in technical Q&A contexts.

Loading preview...