Mushari440/Qwen3-8B-SFT-chatml
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Apr 5, 2026Architecture:Transformer Cold

Mushari440/Qwen3-8B-SFT-chatml is an 8 billion parameter causal language model developed by Mushari Alothman, fine-tuned from Qwen3-8B-Base. This supervised fine-tuned (SFT) model is optimized for accurate, clean supervision across both Arabic and English tasks. It excels in use cases such as MCQ answering, context-based QA/RAG, and general instruction following in both languages. The model supports a context length of 32768 tokens.

Loading preview...