choiqs/Qwen3-1.7B-tldr-bsz128-ts300-regular-qrm-skywork8b-seed42-lr1e-6-warmup10-checkpoint75
TEXT GENERATIONConcurrency Cost:1Model Size:2BQuant:BF16Ctx Length:32kPublished:Apr 9, 2026Architecture:Transformer Loading
The choiqs/Qwen3-1.7B-tldr-bsz128-ts300-regular-qrm-skywork8b-seed42-lr1e-6-warmup10-checkpoint75 is a 2 billion parameter language model based on the Qwen3 architecture, featuring a 32768 token context length. This model is specifically fine-tuned for TLDR (Too Long; Didn't Read) summarization tasks, indicating an optimization for concise information extraction. Its training configuration suggests a focus on efficient processing and performance for summarization, making it suitable for applications requiring quick content overviews.
Loading preview...