allenai/llama-3-tulu-2-dpo-8b
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:8kPublished:Jun 20, 2024License:apache-2.0Architecture:Transformer0.0K Open Weights Cold

allenai/llama-3-tulu-2-dpo-8b is an 8 billion parameter language model developed by AllenAI, fine-tuned from Meta Llama 3. It is trained as a helpful assistant using a mix of public, synthetic, and human datasets, further aligned with Direct Preference Optimization (DPO) on the UltraFeedback dataset. This model is primarily English-language and excels in assistant-like conversational tasks.

Loading preview...