huseyinatahaninan/appworld_distillation_sft-SFT-Qwen3-8B
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Jan 20, 2026License:apache-2.0Architecture:Transformer Open Weights Cold

The huseyinatahaninan/appworld_distillation_sft-SFT-Qwen3-8B is an 8 billion parameter language model, fine-tuned from the Qwen/Qwen3-8B architecture. It was trained on the appworld_distillation_sft dataset, achieving a validation loss of 0.2667. This model is specialized for tasks related to its specific fine-tuning dataset, making it suitable for applications requiring knowledge or generation aligned with 'appworld_distillation_sft' content.

Loading preview...