nbeerbower/Merlina-ORPO-12B
TEXT GENERATIONConcurrency Cost:1Model Size:12BQuant:FP8Ctx Length:32kLicense:apache-2.0Architecture:Transformer0.0K Open Weights Cold

Merlina-ORPO-12B is a 12 billion parameter language model developed by nbeerbower, based on the same training run as schneewolflabs/A0l-12B. This model distinguishes itself by utilizing a custom ORPO (Odds Ratio Preference Optimization) implementation with a beta value of 0.1. It is designed for tasks benefiting from advanced preference optimization techniques.

Loading preview...