Taichi11/sft_v7_dpo_v2_merged
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Feb 22, 2026License:apache-2.0Architecture:Transformer Open Weights Warm

Taichi11/sft_v7_dpo_v2_merged is a 4 billion parameter language model fine-tuned by Taichi11 using Direct Preference Optimization (DPO) on the Taichi11/LLM_main_v7_merged base model. Optimized for improved reasoning through Chain-of-Thought and enhanced structured response quality, this model is designed for applications requiring precise and well-organized outputs. It offers a 32768 token context length and is provided with full-merged 16-bit weights for direct use without adapter loading.

Loading preview...

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p