UCLA-AGI/Llama-3-Instruct-8B-SPPO-Iter2
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:8kPublished:Jun 25, 2024License:apache-2.0Architecture:Transformer Open Weights Warm

UCLA-AGI/Llama-3-Instruct-8B-SPPO-Iter2 is an 8 billion parameter instruction-tuned language model developed by UCLA-AGI. It is based on the Meta-Llama-3-8B-Instruct architecture and fine-tuned using Self-Play Preference Optimization (SPPO) at its second iteration. This model is optimized for improved alignment and performance, demonstrating enhanced win rates on the AlpacaEval Leaderboard compared to its predecessor.

Loading preview...