CompassioninMachineLearning/PretrainingBasellama3kv3_plus3khelpfullnessGRPO1epoch
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:8kPublished:Mar 11, 2026License:apache-2.0Architecture:Transformer Open Weights Cold

The CompassioninMachineLearning/PretrainingBasellama3kv3_plus3khelpfullnessGRPO1epoch is an 8 billion parameter Llama-based language model developed by CompassioninMachineLearning, fine-tuned from the PretrainingBasellama3kv3 model. This model was trained for enhanced helpfulness using the TRL library and Unsloth, which accelerated its training process. It features an 8192 token context length and is designed for general language understanding and generation tasks.

Loading preview...