princeton-nlp/Llama-3-Base-8B-SFT-CPO
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:8kPublished:Jul 6, 2024Architecture:Transformer Warm

princeton-nlp/Llama-3-Base-8B-SFT-CPO is an 8 billion parameter language model developed by Princeton NLP, based on the Llama-3 architecture with an 8192 token context length. This model is fine-tuned using Supervised Fine-Tuning (SFT) and Preference Optimization (CPO) as detailed in the SimPO research, focusing on improving response quality without a reference-free reward. It is designed for general language understanding and generation tasks, leveraging advanced preference optimization techniques.

Loading preview...