W-61/hh-helpful-base-qwen3-8b-sft
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Mar 3, 2026Architecture:Transformer Cold

W-61/hh-helpful-base-qwen3-8b-sft is an 8 billion parameter language model fine-tuned from the Qwen/Qwen3-8B architecture. It has been specifically trained using Supervised Fine-Tuning (SFT) with TRL to enhance its helpfulness and instruction-following capabilities. This model is designed for general text generation tasks where helpful and coherent responses are required, leveraging its 32768 token context length for detailed interactions.

Loading preview...