NousResearch/Nous-Hermes-2-Mistral-7B-DPO
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Feb 18, 2024License:apache-2.0Architecture:Transformer0.2K Open Weights Warm

NousResearch/Nous-Hermes-2-Mistral-7B-DPO is a 7 billion parameter instruction-tuned language model developed by NousResearch, based on the Mistral architecture. This model was fine-tuned using Direct Preference Optimization (DPO) from Teknium's OpenHermes-2.5-Mistral-7B, demonstrating improved performance across AGIEval, BigBench Reasoning, GPT4All, and TruthfulQA benchmarks. It is optimized for general-purpose conversational AI and instruction following, leveraging a 4096-token context length.

Loading preview...