choiqs/Qwen3-1.7B-tldr-bsz128-ts300-regular-qrm-skywork8b-seed42-lr1e-6-warmup10-checkpoint175
TEXT GENERATIONConcurrency Cost:1Model Size:2BQuant:BF16Ctx Length:32kPublished:Apr 9, 2026Architecture:Transformer Loading

The choiqs/Qwen3-1.7B-tldr-bsz128-ts300-regular-qrm-skywork8b-seed42-lr1e-6-warmup10-checkpoint175 is a 1.7 billion parameter language model based on the Qwen3 architecture. This model is fine-tuned for specific tasks, indicated by its detailed naming convention, suggesting optimizations for batch size 128, a sequence length of 300, and regularized quantization-aware training (QRM). It is likely designed for efficient inference and deployment in resource-constrained environments, focusing on specific performance characteristics rather than broad general-purpose capabilities.

Loading preview...