shubhamrgandhi/qwen3-8b-full-sft-prm-opus-distill-32k-lr5e6_rejection-sample_think
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Mar 27, 2026License:otherArchitecture:Transformer Cold
The shubhamrgandhi/qwen3-8b-full-sft-prm-opus-distill-32k-lr5e6_rejection-sample_think model is an 8 billion parameter language model, fine-tuned from Qwen/Qwen3-8B. It was trained on the prm_sft_train dataset with a context length of 32768 tokens. This model is a specialized iteration focusing on specific fine-tuning objectives, though further details on its unique capabilities are not explicitly provided in the available documentation. It is intended for applications requiring a Qwen3-8B base model with this particular fine-tuning configuration.
Loading preview...