sonthenguyen/OpenHermes-2.5-Mistral-7B-mt-bench-DPO-corrupted
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Feb 4, 2024License:apache-2.0Architecture:Transformer Open Weights Cold

The sonthenguyen/OpenHermes-2.5-Mistral-7B-mt-bench-DPO-corrupted model is a 7 billion parameter causal language model, fine-tuned using DPO (Direct Preference Optimization) on the OpenHermes-2.5-Mistral-7B base. This model was trained with specific LoRA configurations and optimized for conversational tasks, leveraging a maximum context length of 4096 tokens. Its training methodology suggests a focus on aligning model outputs with human preferences, making it suitable for interactive AI applications.

Loading preview...