yusufcelebi/qwen3-8B-Base-orca_math-sparse-LoRA-step180-merged
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Jan 11, 2026Architecture:Transformer Warm

The yusufcelebi/qwen3-8B-Base-orca_math-sparse-LoRA-step180-merged model is an 8 billion parameter language model based on the Qwen3 architecture. This model has been fine-tuned using a sparse LoRA method, specifically optimized for mathematical reasoning tasks, as indicated by its 'orca_math' designation. It is designed to excel in scenarios requiring strong numerical and logical problem-solving capabilities, offering a context length of 32768 tokens.

Loading preview...

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p