CompassioninMachineLearning/PretrainingBasellama3kv3_plus3kcodingGRPO1epoch
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:8kPublished:Mar 11, 2026License:apache-2.0Architecture:Transformer Open Weights Cold

The CompassioninMachineLearning/PretrainingBasellama3kv3_plus3kcodingGRPO1epoch is an 8 billion parameter Llama-based language model developed by CompassioninMachineLearning. This model was finetuned from CompassioninMachinelearning/PretrainingBasellama3kv3, leveraging Unsloth and Huggingface's TRL library for accelerated training. It is designed for general language understanding and generation tasks, with a focus on efficient development and deployment.

Loading preview...