weizhepei/Qwen2.5-3B-WebArena-Lite-SFT-CoT-QwQ-32B-epoch-10
TEXT GENERATIONConcurrency Cost:1Model Size:3.1BQuant:BF16Ctx Length:32kPublished:Apr 2, 2025Architecture:Transformer Warm
The weizhepei/Qwen2.5-3B-WebArena-Lite-SFT-CoT-QwQ-32B-epoch-10 model is a 3.1 billion parameter language model developed by weizhepei, fine-tuned from Qwen/Qwen2.5-3B-Instruct. It is specifically trained on the WebArena-Lite-SFT-CoT-QwQ-32B dataset using Supervised Fine-Tuning (SFT) with the TRL framework. This model is optimized for tasks related to web environments and complex reasoning, leveraging its specialized training data to enhance performance in these domains.
Loading preview...