choiqs/Qwen3-1.7B-tldr-bsz128-ts300-regular-qrm-seed42-lr1e-6-warmup10-checkpoint75
TEXT GENERATIONConcurrency Cost:1Model Size:2BQuant:BF16Ctx Length:32kPublished:Apr 8, 2026Architecture:Transformer Loading

The choiqs/Qwen3-1.7B-tldr-bsz128-ts300-regular-qrm-seed42-lr1e-6-warmup10-checkpoint75 model is a 2 billion parameter language model based on the Qwen3 architecture, featuring a substantial 32768 token context length. This model is specifically fine-tuned for TLDR (Too Long; Didn't Read) summarization tasks, indicating an optimization for concise information extraction. Its design suggests a focus on processing lengthy inputs to generate brief, relevant summaries, making it suitable for applications requiring efficient content condensation.

Loading preview...