OldKingMeister/Qwen2.5-1.5B-Instruct-YaRN
TEXT GENERATIONConcurrency Cost:1Model Size:1.5BQuant:BF16Ctx Length:32kPublished:Apr 4, 2025License:apache-2.0Architecture:Transformer0.0K Open Weights Cold

OldKingMeister/Qwen2.5-1.5B-Instruct-YaRN is a 1.54 billion parameter instruction-tuned causal language model based on the Qwen2.5 architecture, developed by Qwen. This model features an extended context length of 131,072 tokens, enabled by the YaRN technique, significantly enhancing its ability to process and generate long texts. It excels in coding, mathematics, instruction following, and structured data understanding, making it suitable for applications requiring extensive context and precise output.

Loading preview...