openaccess-ai-collective/openhermes-2_5-dpo-no-robots
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Nov 27, 2023License:apache-2.0Architecture:Transformer0.0K Open Weights Cold

The openaccess-ai-collective/openhermes-2_5-dpo-no-robots is a 7 billion parameter language model, fine-tuned using Reinforcement Learning (RL) with Direct Preference Optimization (DPO). It is based on teknium/OpenHermes-2.5-Mistral-7B and specifically optimized on a preference dataset derived from HuggingFace's 'no_robots' dataset. This model is designed for tasks requiring nuanced understanding of human preferences, particularly in conversational AI where avoiding 'robotic' responses is crucial.

Loading preview...