Tiiny/SmallThinker-3B-Preview
TEXT GENERATIONConcurrency Cost:1Model Size:3.1BQuant:BF16Ctx Length:32kPublished:Dec 12, 2024Architecture:Transformer0.4K Warm

SmallThinker-3B-Preview is a 3.1 billion parameter language model fine-tuned from Qwen2.5-3B-Instruct by PowerInfer, specifically optimized for enhanced mathematical and reasoning capabilities. It features a 32768-token context length and demonstrates significant performance improvements on benchmarks like AIME24, GAOKAO2024, and MMLU_STEM. This model is primarily designed for efficient edge deployment on resource-constrained devices and can serve as a fast draft model for larger LLMs.

Loading preview...

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p