Radiantloom/radintloom-mistral-7b-fusion-dpo
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Feb 20, 2024License:apache-2.0Architecture:Transformer0.0K Open Weights Cold

Radiantloom/radintloom-mistral-7b-fusion-dpo is a 7 billion parameter causal language model developed by Radiantloom, fine-tuned using Direct Preference Optimization (DPO). It is an enhanced version of the Radiantloom Mistral 7B Fusion model, designed for improved performance through preference learning. This model is built on the Mistral architecture and supports a 4096-token context length, making it suitable for general language generation tasks where preference alignment is beneficial.

Loading preview...