josang1204/Qweb2.5-FT-CSY
TEXT GENERATIONConcurrency Cost:1Model Size:0.5BQuant:BF16Ctx Length:32kArchitecture:Transformer Warm

josang1204/Qweb2.5-FT-CSY is a 0.5 billion parameter language model, fine-tuned from Qwen/Qwen2.5-0.5B using the TRL library. This model has a notable context length of 131072 tokens, making it suitable for processing extensive inputs. It is specifically optimized through supervised fine-tuning (SFT) for general text generation tasks, building upon the capabilities of its base Qwen2.5 architecture.

Loading preview...