jack009064/Affine-mmh2-5EptJ5DkkearraPC65QFsPbkHkB1BZnNfoeJ5iLKeNXJGUR2
TEXT GENERATIONConcurrency Cost:2Model Size:32BQuant:FP8Ctx Length:32kPublished:Mar 27, 2026Architecture:Transformer Cold
The jack009064/Affine-mmh2-5EptJ5DkkearraPC65QFsPbkHkB1BZnNfoeJ5iLKeNXJGUR2 model is a 32 billion parameter language model fine-tuned using Direct Preference Optimization (DPO). This model leverages the TRL framework for training, focusing on aligning its responses with human preferences. It is designed for general text generation tasks, offering improved conversational quality through preference-based learning.
Loading preview...