wulli/Qwen2.5-0.5B-sft-capybara
TEXT GENERATIONConcurrency Cost:1Model Size:0.5BQuant:BF16Ctx Length:32kPublished:Nov 9, 2025Architecture:Transformer Warm
The wulli/Qwen2.5-0.5B-sft-capybara model is a 0.5 billion parameter language model, fine-tuned from the Qwen/Qwen2.5-0.5B architecture using the TRL framework. This model is specifically optimized for instruction following, leveraging supervised fine-tuning (SFT) to enhance its conversational capabilities. With a substantial context length of 131,072 tokens, it is suitable for applications requiring processing of extensive input. Its primary use case is generating coherent and contextually relevant text based on user prompts.
Loading preview...
Popular Sampler Settings
Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.
temperature
–
top_p
–
top_k
–
frequency_penalty
–
presence_penalty
–
repetition_penalty
–
min_p
–