Aratako/Llama-Gemma-2-27b-ORPO-iter3
TEXT GENERATIONConcurrency Cost:2Model Size:27BQuant:FP8Ctx Length:32kPublished:Dec 16, 2024License:llama3.1Architecture:Transformer0.0K Cold

Aratako/Llama-Gemma-2-27b-ORPO-iter3 is a 27 billion parameter instruction-tuned causal language model developed by Aratako. It is built upon a Llama and Gemma 2 base, further refined using ORPO (Optimized Reward Policy Optimization) after initial CPO_SimPO instruction tuning. This model is designed for general instruction following, leveraging its sophisticated fine-tuning process for enhanced performance.

Loading preview...