JY623/KoSOLAR-v2.0

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:10.7BQuant:FP8Ctx Length:4kLicense:apache-2.0Architecture:Transformer Open Weights Warm

JY623/KoSOLAR-v2.0 is a merged language model created using the TIES method, based on chihoonlee10/T3Q-ko-solar-dpo-v3.0. It integrates davidkim205/nox-solar-10.7b-v4 and Deepnoid/deep-solar-Rev-v3.0.4 to combine their respective strengths. This model is designed to leverage the collective capabilities of its constituent models for enhanced performance in Korean language tasks. Its architecture is optimized for general-purpose language generation and understanding within the Korean context.

Loading preview...

Model Overview

JY623/KoSOLAR-v2.0 is a merged language model developed by JY623, utilizing the TIES (Trimming, Iterative Retraining, and Sparsity) merge method. This approach combines the strengths of multiple pre-trained models into a single, more capable model.

Key Components and Merge Details

This model is built upon chihoonlee10/T3Q-ko-solar-dpo-v3.0 as its base. It integrates two additional models:

  • davidkim205/nox-solar-10.7b-v4
  • Deepnoid/deep-solar-Rev-v3.0.4

The TIES method was applied with specific density and weight parameters (0.5 for both density and weight) for the merged models, and normalization was enabled during the merge process. The model was produced with float16 data type.

Intended Use Cases

KoSOLAR-v2.0 is designed for applications requiring robust Korean language understanding and generation. By merging models known for their performance in Korean, it aims to provide a versatile tool for various NLP tasks, including but not limited to, text generation, summarization, and conversational AI in Korean.