vitus9988/Llama-3.2-1B-Instruct-Ko-SFT
TEXT GENERATIONConcurrency Cost:1Model Size:1BQuant:BF16Ctx Length:32kLicense:apache-2.0Architecture:Transformer Open Weights Warm
vitus9988/Llama-3.2-1B-Instruct-Ko-SFT is a 1 billion parameter instruction-tuned causal language model based on the Llama 3.2 architecture. This model is specifically fine-tuned for Korean language tasks, making it suitable for applications requiring instruction-following capabilities in Korean. It features a substantial 32,768 token context length, enhancing its ability to process and generate longer Korean texts.
Loading preview...
vitus9988/Llama-3.2-1B-Instruct-Ko-SFT Overview
This model is an instruction-tuned variant of the Llama 3.2 architecture, featuring 1 billion parameters. It has been specifically fine-tuned for Korean language processing, indicating an optimization for tasks and interactions in Korean.
Key Capabilities
- Korean Language Instruction Following: Designed to understand and execute instructions provided in Korean.
- Llama 3.2 Base: Built upon the Llama 3.2 foundational model, inheriting its core capabilities.
- Extended Context Window: Supports a context length of 32,768 tokens, allowing for the processing of longer and more complex Korean inputs.
Good For
- Applications requiring a compact yet capable model for Korean natural language understanding and generation.
- Instruction-based tasks where the primary language of interaction is Korean.
- Scenarios benefiting from a larger context window for handling extensive Korean text.