xiaolesu/OsmosisProofling-SFT-NT-GRPO-NT
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Apr 5, 2026License:apache-2.0Architecture:Transformer Open Weights Cold

The xiaolesu/OsmosisProofling-SFT-NT-GRPO-NT is an 8 billion parameter language model fine-tuned from Qwen/Qwen3-8B. It was trained using the xiaolesu/OsmosisProofling-v3-SFT dataset with a context length of 32768 tokens. This model demonstrates a validation loss of 0.3543 and a perplexity of 1.4252, indicating strong performance on the fine-tuning task. It is optimized for tasks related to the specific dataset it was trained on.

Loading preview...