hamishivi/sft_qwen3_4b_tmax_4node2203
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Mar 22, 2026Architecture:Transformer Warm

The hamishivi/sft_qwen3_4b_tmax_4node2203 is a 4 billion parameter language model, likely based on the Qwen3 architecture, developed by hamishivi. This model is instruction-tuned, indicated by 'sft' (supervised fine-tuning), and features a substantial context length of 32768 tokens. It is designed for general language understanding and generation tasks, leveraging its large context window for processing extensive inputs.

Loading preview...