vilm/Quyen-Plus-v0.1

Cold
Public
7.7B
FP8
32768
Feb 6, 2024
License: other
Hugging Face
Overview

Quyen-Plus-v0.1 Overview

Quyen-Plus-v0.1 is a 7.7 billion parameter instruction-tuned large language model, part of the Quyen series developed by vilm. It is built upon the Qwen1.5 architecture and has been fine-tuned using Supervised Fine-Tuning (SFT) and Direct Preference Optimization (DPO).

Key Capabilities & Training

The model was trained on a comprehensive dataset, including well-known resources like OpenHermes-2.5 by Teknium, Capyabara by LDJ, argilla/distilabel-capybara-dpo-7k-binarized by argilla, and orca_dpo_pairs by Intel, alongside private data from Ontocord and BEE-spoke-data. This diverse training regimen aims to enhance its conversational and reasoning abilities.

Performance Highlights

Evaluated on the Open LLM Leaderboard, Quyen-Plus-v0.1 achieved an average score of 63.27. Specific benchmark results include:

  • AI2 Reasoning Challenge (25-Shot): 55.72
  • HellaSwag (10-Shot): 78.52
  • MMLU (5-Shot): 60.45
  • TruthfulQA (0-shot): 53.60
  • Winogrande (5-shot): 71.27
  • GSM8k (5-shot): 60.05

These scores indicate its proficiency in various tasks, from common sense reasoning to mathematical problem-solving. The model utilizes the ChatML prompt template for interaction, supporting a clear and structured conversational format.

When to Use This Model

Quyen-Plus-v0.1 is suitable for developers seeking a capable 7B-class model for general conversational AI applications, instruction following, and tasks requiring reasoning and language understanding. Its balanced performance across multiple benchmarks makes it a versatile choice for a range of use cases.