RoversX/StableBeluga-7B-Qlora-Samantha-Zh-V1

TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kArchitecture:Transformer Cold

RoversX/StableBeluga-7B-Qlora-Samantha-Zh-V1 is a 7 billion parameter conversational AI assistant based on Stability AI's StableBeluga-7B. It is fine-tuned using QLoRA on the Samantha-1.1-zh dataset, specializing in Chinese conversational interactions. This model is designed for generating assistant-like responses in Chinese, leveraging its 4096-token context length.

Loading preview...

Model Overview

RoversX/StableBeluga-7B-Qlora-Samantha-Zh-V1 is a 7 billion parameter conversational AI assistant. It is built upon the robust foundation of Stability AI's StableBeluga-7B and has been further fine-tuned using the QLoRA method. The training specifically utilized the Samantha-1.1-zh dataset, which focuses on Chinese language interactions.

Key Capabilities

  • Chinese Conversational AI: Optimized for generating assistant-like responses in Chinese.
  • Base Model: Leverages the capabilities of StableBeluga-7B, a strong base for instruction-following tasks.
  • Context Length: Supports a context window of 4096 tokens, allowing for more extensive conversations.

Usage

This model requires a specific prompt format for optimal performance:

### System:
This is a system prompt, please behave and help the user.

### User:
Your prompt here

### Assistant:
The output of Stable Beluga 7B

Good For

  • Developing Chinese-language chatbots or virtual assistants.
  • Applications requiring conversational AI with a focus on the Chinese language.
  • Research and development in multilingual LLMs, particularly for Chinese dialogues.