Taichi11/sft_v7_dpo_v2_merged
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Feb 22, 2026License:apache-2.0Architecture:Transformer Open Weights Warm
Taichi11/sft_v7_dpo_v2_merged is a 4 billion parameter language model fine-tuned by Taichi11 using Direct Preference Optimization (DPO) on the Taichi11/LLM_main_v7_merged base model. Optimized for improved reasoning through Chain-of-Thought and enhanced structured response quality, this model is designed for applications requiring precise and well-organized outputs. It offers a 32768 token context length and is provided with full-merged 16-bit weights for direct use without adapter loading.
Loading preview...
Popular Sampler Settings
Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.
temperature
–
top_p
–
top_k
–
frequency_penalty
–
presence_penalty
–
repetition_penalty
–
min_p
–