allenai/tulu-v2.5-ppo-13b-uf-mean-70b-uf-rm-mixed-prompts
TEXT GENERATIONConcurrency Cost:1Model Size:13BQuant:FP8Ctx Length:4kPublished:Jun 11, 2024License:apache-2.0Architecture:Transformer Open Weights Cold

The allenai/tulu-v2.5-ppo-13b-uf-mean-70b-uf-rm-mixed-prompts is a 13 billion parameter language model developed by AllenAI, fine-tuned from Llama-2-13b-hf. This model is part of the Tulu V2.5 series, trained using PPO with a 70B UltraFeedback Reward Model and a mixture of prompts, focusing on acting as a helpful assistant. It is optimized for instruction following and general conversational tasks, building upon the Tulu 2 suite's alignment methods. The model has a context length of 4096 tokens and is intended for English language applications.

Loading preview...