BRlkl/distill-sft-grpo-4_70-full
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Mar 27, 2026Architecture:Transformer Warm

BRlkl/distill-sft-grpo-4_70-full is a 4 billion parameter instruction-tuned causal language model developed by BRlkl. It is a fine-tuned version of the GRPO-4_70 base model, optimized for conversational text generation through Supervised Fine-Tuning (SFT). This model is suitable for general text generation tasks, providing coherent and contextually relevant responses.

Loading preview...