edbeeching/Qwen3-4B-Instruct-2507-SFT-tr5
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Mar 22, 2026Architecture:Transformer Warm

The edbeeching/Qwen3-4B-Instruct-2507-SFT-tr5 is a 4 billion parameter instruction-tuned causal language model, fine-tuned by edbeeching from the Qwen/Qwen3-4B-Instruct-2507 base model. This model leverages a 32768 token context length and was trained using SFT with the TRL framework. It is designed for general instruction-following tasks, building upon the capabilities of its Qwen3 base.

Loading preview...