OpenLLM-Ro/RoGemma2-9b-Instruct-DPO
TEXT GENERATIONConcurrency Cost:1Model Size:9BQuant:FP8Ctx Length:16kPublished:Oct 10, 2024License:cc-by-nc-4.0Architecture:Transformer0.0K Open Weights Cold
OpenLLM-Ro/RoGemma2-9b-Instruct-DPO is a 9 billion parameter instruction-tuned generative text model developed by OpenLLM-Ro, specifically designed for the Romanian language. This model is a human-aligned instruct variant, fine-tuned using Direct Preference Optimization (DPO) on various Romanian datasets. It excels in Romanian language tasks, offering strong performance in areas like machine translation (EN-RO) and semantic textual similarity (STS) in few-shot settings, making it suitable for research and assistant-like chat applications in Romanian.
Loading preview...