LorenaYannnnn/general_reward-Qwen3-0.6B-baseline_all_tokens_w_kl-seed_0
TEXT GENERATIONConcurrency Cost:1Model Size:0.8BQuant:BF16Ctx Length:32kPublished:Mar 21, 2026Architecture:Transformer Warm
The LorenaYannnnn/general_reward-Qwen3-0.6B-baseline_all_tokens_w_kl-seed_0 model is a 0.8 billion parameter language model based on the Qwen3 architecture. This model is a baseline version, potentially serving as a foundation for further fine-tuning or research into reward modeling. Its primary use case is likely experimental or as a component in larger AI systems, given its baseline nature and smaller parameter count.
Loading preview...