mlabonne/NeuralHermes-2.5-Mistral-7B
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Nov 29, 2023License:apache-2.0Architecture:Transformer0.2K Open Weights Cold
NeuralHermes-2.5-Mistral-7B by mlabonne is a 7 billion parameter language model fine-tuned with Direct Preference Optimization (DPO) on the teknium/OpenHermes-2.5-Mistral-7B base. This model demonstrates improved performance across benchmarks like AGIEval, GPT4All, and TruthfulQA, making it suitable for general conversational AI and reasoning tasks. It leverages a ChatML-formatted dataset for DPO, enhancing its instruction-following capabilities.
Loading preview...