Thrillcrazyer/Qwen-7B_SFT
TEXT GENERATIONConcurrency Cost:1Model Size:7.6BQuant:FP8Ctx Length:32kPublished:Nov 30, 2025Architecture:Transformer Cold

Thrillcrazyer/Qwen-7B_SFT is a 7.6 billion parameter language model, fine-tuned from deepseek-ai/DeepSeek-R1-Distill-Qwen-7B using Supervised Fine-Tuning (SFT) with a context length of 32768 tokens. This model is designed for general text generation tasks, leveraging its SFT training to produce coherent and contextually relevant responses. It is suitable for applications requiring robust language understanding and generation capabilities based on its Qwen architecture.

Loading preview...