xiaolesu/OsmosisProofling-SFT-NT-GRPO-TK-V2
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Apr 10, 2026License:apache-2.0Architecture:Transformer Open Weights Cold

The xiaolesu/OsmosisProofling-SFT-NT-GRPO-TK-V2 is an 8 billion parameter causal language model developed by xiaolesu, fine-tuned from Qwen/Qwen3-8B. This model is specifically fine-tuned on the OsmosisProofling-SFT dataset, demonstrating a validation perplexity of 1.4252. It is built using the Axolotl framework and incorporates Liger plugin features like rope, rms_norm, and glu_activation. This model is optimized for tasks aligned with its fine-tuning dataset, showing strong performance in terms of loss and perplexity.

Loading preview...