choiqs/Qwen3-1.7B-tldr-bsz128-ts300-regular-qrm-skywork8b-seed42-lr1e-6-warmup10-checkpoint200
TEXT GENERATIONConcurrency Cost:1Model Size:2BQuant:BF16Ctx Length:32kPublished:Apr 9, 2026Architecture:Transformer Loading
The choiqs/Qwen3-1.7B-tldr-bsz128-ts300-regular-qrm-skywork8b-seed42-lr1e-6-warmup10-checkpoint200 is a 2 billion parameter language model with a 32768 token context length. This model is part of the Qwen family, developed by choiqs, and is designed for general language understanding and generation tasks. Its specific fine-tuning details suggest an optimization for particular training regimes, making it suitable for applications requiring efficient processing within its parameter class.
Loading preview...