Qwen/Qwen1.5-4B-Chat
Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Jan 30, 2024License:tongyi-qianwen-researchArchitecture:Transformer0.0K Warm

Qwen1.5-4B-Chat is a 4 billion parameter, transformer-based decoder-only language model developed by Qwen, serving as a beta version of Qwen2. It features stable support for a 32K context length and offers significant performance improvements in human preference for chat models. This model provides robust multilingual support and is designed for general-purpose conversational AI applications.

Loading preview...

Qwen1.5-4B-Chat Overview

Qwen1.5-4B-Chat is a 4 billion parameter model from the Qwen1.5 series, representing a beta iteration of the Qwen2 transformer-based decoder-only language models. This series introduces several enhancements over previous Qwen versions, focusing on improved performance and broader applicability.

Key Capabilities & Features

  • Model Architecture: Based on the Transformer architecture, incorporating SwiGLU activation, attention QKV bias, and group query attention. It also features an improved tokenizer designed for multiple natural languages and code.
  • Context Length: Provides stable support for a 32K token context length across all model sizes, including this 4B variant.
  • Multilingual Support: Both base and chat models within the Qwen1.5 series offer comprehensive multilingual capabilities.
  • Chat Performance: Demonstrates significant improvements in human preference scores for chat-oriented tasks.
  • Training Methodology: Models are pretrained on extensive datasets, followed by post-training using supervised finetuning and direct preference optimization.

When to Use This Model

Qwen1.5-4B-Chat is suitable for developers seeking a moderately sized, multilingual chat model with a substantial context window. Its improved chat performance and stable 32K context make it a strong candidate for conversational AI, content generation, and applications requiring understanding and generation across various languages. The model's design also simplifies integration, as it does not require trust_remote_code.