ICTNLP/bayling-13b-diff

TEXT GENERATIONConcurrency Cost:1Model Size:13BQuant:FP8Ctx Length:4kPublished:Jun 14, 2023License:gpl-3.0Architecture:Transformer0.0K Open Weights Cold

ICTNLP/bayling-13b-diff is a 13 billion parameter instruction-following large language model developed by the NLP Group of the Institute of Computing Technology, Chinese Academy of Sciences (ICT/CAS). This model is a weight-diff version of BayLing-13B-v1.0, designed for advanced language alignment and superior English/Chinese generation, instruction following, and multi-turn interaction. It is optimized for deployment on consumer-grade GPUs with 16GB of memory, assisting with tasks like translation, writing, and creation.

Loading preview...

BayLing-13B-diff: Cross-lingual Instruction Following LLM

BayLing-13B-diff is a 13 billion parameter instruction-following large language model developed by ICT/CAS, focusing on bridging cross-lingual alignment and instruction following. This model is a weight-diff version of BayLing-13B-v1.0, meaning it provides the differences from a base model, allowing for efficient deployment. It is designed to run on consumer-grade GPUs with 16GB of memory.

Key Capabilities

  • Advanced Language Alignment: Demonstrates superior capability in English and Chinese generation.
  • Instruction Following: Excels in understanding and executing instructions.
  • Multi-turn Interaction: Capable of engaging in extended conversations.
  • Resource-Efficient Deployment: Can be deployed on consumer-grade GPUs.
  • Multilingual Evaluation: The project includes a human-annotated evaluation set, BayLing-80 Test Set, for assessing multilingual and multi-turn interaction capabilities.

Limitations

BayLing-13B-diff, like other LLMs, has limitations. It may generate inaccurate factual information and lacks proficiency in reasoning, mathematics, and coding tasks. There is also a risk of generating harmful or biased content. For enhanced Chinese knowledge, users are directed to BayLing-13B-v1.1.