haoranxu/Llama-3-Instruct-8B-SimPO
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:8kPublished:Jun 7, 2024License:llama3Architecture:Transformer0.0K Warm

The haoranxu/Llama-3-Instruct-8B-SimPO model is an 8 billion parameter instruction-tuned language model, fine-tuned from Meta-Llama-3-8B-Instruct. It leverages the SimPO training method on the princeton-nlp/llama3-ultrafeedback dataset, enhancing its ability to follow instructions and generate high-quality responses. This model is designed for general-purpose conversational AI and instruction-following tasks, offering an 8192 token context window.

Loading preview...