huseyinatahaninan/appworld_distillation_sft_v2-SFT-Qwen3-14B
TEXT GENERATIONConcurrency Cost:1Model Size:14BQuant:FP8Ctx Length:32kPublished:Jan 19, 2026License:apache-2.0Architecture:Transformer Open Weights Cold
The huseyinatahaninan/appworld_distillation_sft_v2-SFT-Qwen3-14B is a 14 billion parameter language model, fine-tuned from the Qwen3-14B architecture. This model has been specifically fine-tuned on the appworld_distillation_sft_v2 dataset, indicating a specialization in tasks related to that dataset's domain. It demonstrates a validation loss of 0.6408, suggesting its performance within the scope of its training data. This model is suitable for applications requiring capabilities aligned with the appworld_distillation_sft_v2 dataset.
Loading preview...