uc-rl/Qwen2.5-3B-UCRL
TEXT GENERATIONConcurrency Cost:1Model Size:3.1BQuant:BF16Ctx Length:32kPublished:Nov 7, 2025Architecture:Transformer Warm

uc-rl/Qwen2.5-3B-UCRL is a 3.1 billion parameter causal language model, fine-tuned from Qwen/Qwen2.5-3B-Instruct. Developed by uc-rl, this model specializes in mathematical reasoning and problem-solving, leveraging the GRPO training method. It is optimized for verifiable coding problems and tasks requiring robust logical deduction, offering a 32768-token context length.

Loading preview...

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p