hkust-nlp/Mistral-Small-24B-SimpleRL-Zoo
TEXT GENERATIONConcurrency Cost:2Model Size:24BQuant:FP8Ctx Length:32kPublished:Mar 24, 2025License:apache-2.0Architecture:Transformer Open Weights Cold

The hkust-nlp/Mistral-Small-24B-SimpleRL-Zoo model is a 24 billion parameter language model developed by hkust-nlp, based on the Mistral architecture. It features a substantial 32768-token context window, making it suitable for processing extensive inputs. This model is specifically fine-tuned using SimpleRL, indicating an optimization for reinforcement learning from human feedback (RLHF) techniques to enhance its conversational and instruction-following capabilities. Its primary strength lies in generating coherent and contextually relevant responses across a wide range of general-purpose language tasks.

Loading preview...