allenai/tulu-v2.5-ppo-13b-nectar-60k
TEXT GENERATIONConcurrency Cost:1Model Size:13BQuant:FP8Ctx Length:4kPublished:Jun 11, 2024License:apache-2.0Architecture:Transformer Open Weights Cold
The allenai/tulu-v2.5-ppo-13b-nectar-60k model is a 13 billion parameter language model developed by AllenAI, fine-tuned from Llama-2-13b-hf. It is part of the Tulu V2.5 series, trained using PPO on a 60k subsample of the Nectar dataset, specifically designed to function as a helpful assistant. This model focuses on learning from preference feedback, leveraging a reward model trained on the same Nectar split to enhance its conversational capabilities.
Loading preview...