FlyPig23/Qwen3-4B_Paper_Impact_SFT_1ep
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Apr 7, 2026License:otherArchitecture:Transformer Cold

FlyPig23/Qwen3-4B_Paper_Impact_SFT_1ep is a 4 billion parameter language model, fine-tuned from the Qwen/Qwen3-4B-Instruct-2507 base model. This model was specifically trained for one epoch on the paper_impact_sft_train dataset, achieving a validation loss of 0.0623. It is optimized for tasks related to paper impact analysis, leveraging its Qwen3 architecture and a 32768 token context length.

Loading preview...