Qwen/Qwen1.5-1.8B-Chat
Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:1.8BQuant:BF16Ctx Length:32kPublished:Jan 30, 2024License:otherArchitecture:Transformer0.1K Warm

Qwen1.5-1.8B-Chat is a 1.8 billion parameter, transformer-based decoder-only language model developed by Qwen. This chat-optimized model is part of the Qwen1.5 series, offering improved human preference performance and stable support for a 32K context length. It features an improved tokenizer for multilingual support and is designed for general-purpose conversational AI applications.

Loading preview...

Qwen1.5-1.8B-Chat Overview

Qwen1.5-1.8B-Chat is a 1.8 billion parameter model from the Qwen1.5 series, serving as a beta version of Qwen2. This transformer-based, decoder-only language model is pretrained on extensive data and further post-trained using supervised finetuning and direct preference optimization.

Key Capabilities & Improvements

  • Enhanced Chat Performance: Demonstrates significant improvements in human preference scores for chat models compared to previous Qwen versions.
  • Multilingual Support: Both base and chat models offer robust multilingual capabilities.
  • Extended Context Length: Provides stable support for a 32K token context length across all model sizes in the series.
  • Simplified Integration: No longer requires trust_remote_code for easier deployment.
  • Architectural Features: Utilizes a Transformer architecture with SwiGLU activation, attention QKV bias, and an improved tokenizer adaptive to multiple natural languages and code.

When to Use This Model

This model is well-suited for applications requiring a compact yet capable conversational AI. Its improved chat performance and multilingual support make it ideal for:

  • General-purpose chatbots and virtual assistants.
  • Multilingual text generation and understanding tasks.
  • Applications where a balance between performance and computational efficiency is crucial, leveraging its 1.8B parameter size.
Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p