Rakancorle1/qwen2.5-3b_Instruct_policy_traj_30k_full
TEXT GENERATIONConcurrency Cost:1Model Size:3.1BQuant:BF16Ctx Length:32kPublished:Sep 5, 2025License:otherArchitecture:Transformer Warm

Rakancorle1/qwen2.5-3b_Instruct_policy_traj_30k_full is a 3.1 billion parameter instruction-tuned causal language model, fine-tuned from Qwen/Qwen2.5-3B-Instruct. This model has been specialized through further training on the Policy_Traj_0826_30k_train dataset. It is designed for tasks requiring adherence to specific policy trajectories, leveraging its Qwen2.5 architecture and a 32K context length.

Loading preview...