LorenaYannnnn/20260217-Qwen3-0.6B_sycophancy_warmup_16000_ep_OURS_gdpo_192000_episodes_seed_42
TEXT GENERATIONConcurrency Cost:1Model Size:0.8BQuant:BF16Ctx Length:32kPublished:Feb 21, 2026Architecture:Transformer Warm

The LorenaYannnnn/20260217-Qwen3-0.6B_sycophancy_warmup_16000_ep_OURS_gdpo_192000_episodes_seed_42 model is a 0.8 billion parameter language model with a 32768 token context length. This model is part of the Qwen3 family, developed by LorenaYannnnn. Its specific training regimen, including "sycophancy warmup" and GDPO, suggests an optimization for nuanced conversational dynamics and alignment, potentially focusing on reducing undesirable model behaviors or enhancing specific interaction styles. Further details on its exact capabilities and intended use cases are not provided in the available documentation.

Loading preview...