shahzebnaveed/NeuralHermes-2.5-Mistral-7B
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Feb 15, 2024License:apache-2.0Architecture:Transformer Open Weights Cold

NeuralHermes-2.5-Mistral-7B by shahzebnaveed is a 7 billion parameter language model based on the teknium/OpenHermes-2.5-Mistral-7B architecture. It has been fine-tuned using Direct Preference Optimization (DPO) on the Intel/orca_dpo_pairs dataset, reformatting with the ChatML template. This model aims to improve performance through an RLHF-inspired process, similar to Intel/neural-chat-7b-v3-1. It is suitable for general conversational AI tasks, leveraging its Mistral base and DPO fine-tuning.

Loading preview...