LorenaYannnnn/20260216-Qwen3-0.6B_warmup_grpo_OURS_cl_0.6B_128000_episodes_seed_42
TEXT GENERATIONConcurrency Cost:1Model Size:0.8BQuant:BF16Ctx Length:32kPublished:Feb 16, 2026Architecture:Transformer Warm
The LorenaYannnnn/20260216-Qwen3-0.6B_warmup_grpo_OURS_cl_0.6B_128000_episodes_seed_42 model is a 0.8 billion parameter language model. This model is a Qwen3 variant, likely developed by LorenaYannnnn, and features a context length of 32768 tokens. Its specific training and optimization details are not provided, suggesting it may be a base or experimental model.
Loading preview...