Qwen/Qwen2.5-1.5B-Instruct
Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:1.5BQuant:BF16Ctx Length:32kPublished:Sep 17, 2024License:apache-2.0Architecture:Transformer0.7K Open Weights Warm

Qwen2.5-1.5B-Instruct is a 1.54 billion parameter instruction-tuned causal language model developed by Qwen, part of the Qwen2.5 series. This model significantly improves upon Qwen2 in coding, mathematics, and instruction following, offering enhanced long-text generation and structured data understanding. It supports a 128K token context length and is optimized for generating structured outputs like JSON, making it suitable for diverse chatbot and data processing applications.

Loading preview...

Qwen2.5-1.5B-Instruct Overview

Qwen2.5-1.5B-Instruct is a 1.54 billion parameter instruction-tuned causal language model from the Qwen2.5 series, developed by Qwen. It builds upon the Qwen2 architecture with significant enhancements across several key areas, making it a versatile option for various NLP tasks.

Key Capabilities

  • Enhanced Knowledge & Reasoning: Demonstrates improved capabilities in coding and mathematics, benefiting from specialized expert models.
  • Instruction Following: Features significant improvements in adhering to instructions and generating structured outputs, particularly JSON.
  • Long-Context & Generation: Supports a context length of up to 128K tokens and can generate texts up to 8K tokens, ideal for complex and lengthy interactions.
  • Multilingual Support: Offers robust support for over 29 languages, including major global languages like Chinese, English, French, Spanish, and Japanese.
  • Structured Data Understanding: Excels at processing and understanding structured data, such as tables, and is more resilient to diverse system prompts for better role-play and condition-setting in chatbots.

Good For

  • Applications requiring strong coding and mathematical reasoning.
  • Use cases demanding precise instruction following and structured output generation (e.g., JSON).
  • Scenarios involving long-form text generation or processing extensive contexts.
  • Multilingual applications needing support for a wide array of languages.
  • Chatbots and agents that require robust role-play and condition-setting capabilities.
Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p