kopoYH/llm0308
VISIONConcurrency Cost:1Model Size:4.3BQuant:BF16Ctx Length:32kPublished:Mar 28, 2026License:apache-2.0Architecture:Transformer Open Weights Cold

The kopoYH/llm0308 is a 4.3 billion parameter instruction-tuned causal language model developed by kopoYH, based on Google's Gemma-3-4b-it architecture. This model is specifically fine-tuned for Korean language tasks, leveraging the kopoYH/llm0328 dataset. With a context length of 32768 tokens, it is optimized for text generation and demonstrates strong performance in finance-related applications.

Loading preview...

kopoYH/llm0308: A Korean-Optimized Gemma-Based LLM

The kopoYH/llm0308 is a 4.3 billion parameter instruction-tuned language model, building upon the robust google/gemma-3-4b-it architecture. Developed by kopoYH, this model has been specifically adapted and fine-tuned for the Korean language, utilizing the kopoYH/llm0328 dataset to enhance its understanding and generation capabilities in Korean contexts.

Key Capabilities

  • Korean Language Proficiency: Optimized for generating and understanding text in Korean, making it suitable for applications requiring native-level Korean language processing.
  • Instruction Following: As an instruction-tuned model, it is designed to accurately follow user prompts and instructions for various text generation tasks.
  • Extended Context Window: Features a substantial context length of 32768 tokens, allowing it to process and generate longer, more coherent texts while maintaining contextual relevance.
  • Finance Domain Relevance: Tagged with 'finance', indicating potential strengths or fine-tuning for financial text generation and analysis within the Korean language.

Good For

  • Korean Text Generation: Ideal for applications requiring the creation of Korean content, such as articles, summaries, or creative writing.
  • Instruction-Based Tasks: Effective for chatbots, virtual assistants, or any system where the model needs to respond to specific instructions in Korean.
  • Financial Text Processing: Potentially useful for tasks involving financial documents, reports, or queries in Korean, given its domain tagging.
  • Research and Development: Provides a strong base for further fine-tuning or research into Korean language models, especially those derived from the Gemma family.