psyche/kollama2-7b

TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Jul 23, 2023License:apache-2.0Architecture:Transformer0.0K Open Weights Cold

psyche/kollama2-7b is a 7 billion parameter Llama-2 based language model further trained on Korean datasets. This model is specifically optimized for Korean language understanding and generation tasks. It offers a specialized alternative for applications requiring strong performance in Korean, leveraging the Llama-2 architecture.

Loading preview...

psyche/kollama2-7b: Korean-Optimized Llama-2 Model

psyche/kollama2-7b is a 7 billion parameter language model built upon the Llama-2 architecture, distinguished by its further training on extensive Korean datasets. This specialized training aims to enhance its proficiency and performance in processing and generating content in the Korean language.

Key Capabilities & Performance

The model's performance has been evaluated on the Open LLM Leaderboard, showcasing its capabilities across various benchmarks. Notable scores include:

  • HellaSwag (10-shot): 78.78
  • Winogrande (5-shot): 73.95
  • ARC (25-shot): 53.24
  • MMLU (5-shot): 42.31

These metrics indicate its general language understanding and reasoning abilities, with a specific focus on its performance within a Korean-centric context due to its training regimen.

Good For

  • Korean Language Applications: Ideal for tasks requiring robust understanding and generation in Korean.
  • Specialized NLP: Suitable for developers building applications where Korean language proficiency is a primary requirement, such as chatbots, content generation, or translation support for Korean speakers.
  • Llama-2 Ecosystem Users: Offers a Korean-focused variant for those already familiar with or utilizing the Llama-2 model family.