allenai/llama-3-tulu-v2.5-8b-uf-mean-8b-uf-rm
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:8kPublished:Oct 14, 2024License:apache-2.0Architecture:Transformer0.0K Open Weights Cold

The allenai/llama-3-tulu-v2.5-8b-uf-mean-8b-uf-rm is an 8 billion parameter language model developed by AllenAI, built upon Meta's Llama 3 architecture. It is a Tulu V2.5 series model, specifically fine-tuned using Proximal Policy Optimization (PPO) with an 8B reward model on the UltraFeedback dataset, designed to function as a helpful assistant. This model is optimized for conversational AI and instruction following, demonstrating strong performance in areas like mathematical reasoning.

Loading preview...