hyokwan/kopo_gemma3_4b_fintech
VISIONConcurrency Cost:1Model Size:4.3BQuant:BF16Ctx Length:32kPublished:Apr 9, 2026License:apache-2.0Architecture:Transformer Open Weights Cold

The hyokwan/kopo_gemma3_4b_fintech model is a 4.3 billion parameter instruction-tuned language model developed by hyokwan, based on Google's Gemma-3-4b-it architecture. Fine-tuned on the hyokwan/fintech_data_2026 dataset, this model specializes in financial domain tasks with a focus on Korean language processing. It is designed for text generation applications requiring accuracy in fintech contexts, supporting a 32768 token context length.

Loading preview...

hyokwan/kopo_gemma3_4b_fintech: Specialized Financial LLM

The hyokwan/kopo_gemma3_4b_fintech model is a 4.3 billion parameter language model, building upon Google's gemma-3-4b-it architecture. Developed by hyokwan, this model has undergone specialized fine-tuning using the hyokwan/fintech_data_2026 dataset, making it particularly adept at understanding and generating content within the financial technology (fintech) domain.

Key Capabilities

  • Fintech Specialization: Optimized for tasks and queries related to finance and technology, leveraging its dedicated training data.
  • Korean Language Support: Primarily focused on processing and generating text in the Korean language, indicated by the ko language tag.
  • Instruction-Tuned: Designed to follow instructions effectively for various text generation tasks.
  • Extended Context Window: Supports a substantial context length of 32768 tokens, allowing for processing longer financial documents or conversations.

Good For

  • Financial Text Generation: Ideal for applications requiring the generation of financial reports, summaries, or responses.
  • Korean Fintech Applications: Suitable for chatbots, virtual assistants, or content creation tools operating in the Korean fintech sector.
  • Domain-Specific NLP: Useful for research and development in natural language processing within specialized financial contexts.