dfurman/Llama-3-8B-Orpo-v0.1
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:8kPublished:Apr 26, 2024License:llama3Architecture:Transformer0.0K Warm
dfurman/Llama-3-8B-Orpo-v0.1 is an 8 billion parameter language model fine-tuned by dfurman using the ORPO method on 4k samples from the mlabonne/orpo-dpo-mix-40k dataset. Based on Meta-Llama-3-8B, this model utilizes an 8k context window and follows the ChatML template. It demonstrates improved performance in HellaSwag and Winogrande benchmarks compared to its base model, making it suitable for conversational AI applications.
Loading preview...