anakin87/gemma-2b-orpo
TEXT GENERATIONConcurrency Cost:1Model Size:2.6BQuant:BF16Ctx Length:8kPublished:Mar 24, 2024License:gemma-terms-of-useArchitecture:Transformer0.0K Warm

anakin87/gemma-2b-orpo is a 2.6 billion parameter language model fine-tuned from Google's Gemma-2B using the ORPO (Odds Ratio Preference Optimization) training paradigm. This model combines supervised fine-tuning and preference alignment for improved performance with faster training and reduced memory usage. It demonstrates strong performance for its size on benchmarks like Nous and Open LLM Leaderboard, making it suitable for general language generation tasks where efficiency is key.

Loading preview...