edbeeching/Qwen3-4B-Thinking-2507-SFT-tr5
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Mar 22, 2026Architecture:Transformer Warm

edbeeching/Qwen3-4B-Thinking-2507-SFT-tr5 is a 4 billion parameter language model, fine-tuned from Qwen/Qwen3-4B-Thinking-2507 using Supervised Fine-Tuning (SFT) with the TRL framework. This model is designed for text generation tasks, particularly for responding to complex, open-ended questions. It leverages a 32768 token context length, making it suitable for applications requiring extensive contextual understanding and coherent, detailed outputs.

Loading preview...