jaeyong2/Qwen2.5-3B-Instruct-Hi-SFT
TEXT GENERATIONConcurrency Cost:1Model Size:3.1BQuant:BF16Ctx Length:32kLicense:otherArchitecture:Transformer0.0K Warm

jaeyong2/Qwen2.5-3B-Instruct-Hi-SFT is a 3.1 billion parameter instruction-tuned causal language model built upon the Qwen architecture. This model is specifically fine-tuned on a Vietnamese dataset, making it particularly adept at processing and generating content in Vietnamese. It offers a substantial 32,768 token context length, suitable for applications requiring extensive textual understanding and generation in the Vietnamese language.

Loading preview...