vilm/Quyen-v0.1

Loading
Public
4B
BF16
32768
Feb 5, 2024
License: other
Hugging Face
Overview

Quyen-v0.1: An Instruction-Tuned Qwen1.5 Model

Quyen-v0.1 is a 4 billion parameter model from the Quyen series, developed by vilm. This series is built upon the Qwen1.5 architecture and includes various sizes, with Quyen-v0.1 representing the mid-range offering. The model has undergone extensive instruction tuning using both Supervised Fine-Tuning (SFT) and Direct Preference Optimization (DPO).

Key Training Details

  • Base Architecture: Qwen1.5 family.
  • Fine-tuning Methods: SFT and DPO.
  • Training Datasets: A combination of publicly available and private datasets, including:
    • OpenHermes-2.5 by Teknium
    • Capyabara by LDJ
    • argilla/distilabel-capybara-dpo-7k-binarized by argilla
    • orca_dpo_pairs by Intel
    • Private data from Ontocord and BEE-spoke-data.

Prompt Template

All Quyen models, including Quyen-v0.1, utilize the ChatML format for prompting. This standardized template ensures consistent interaction and optimal performance with the model. Developers can easily integrate this by applying the chat template via the tokenizer.

Use Cases

Quyen-v0.1 is suitable for a wide range of conversational AI applications, instruction following, and general text generation tasks, benefiting from its robust fine-tuning on diverse preference datasets. Its 4 billion parameters offer a balance between performance and computational efficiency.

Note: Benchmark results for the Quyen series are currently pending and will be released at a later date.