dh82/123456
VISIONConcurrency Cost:1Model Size:4.3BQuant:BF16Ctx Length:32kPublished:Mar 27, 2026License:apache-2.0Architecture:Transformer Open Weights Cold

dh82/123456 is a 4.3 billion parameter instruction-tuned causal language model, based on the Google Gemma-3-4b-it architecture. This model is primarily focused on Korean language processing, leveraging a specialized dataset for its training. With a context length of 32768 tokens, it is designed for text generation tasks within the Korean linguistic domain. Its development aims to provide a foundational model for further specialized applications in Korean.

Loading preview...