anirvankrishna/model_sft_dare
TEXT GENERATIONConcurrency Cost:1Model Size:1.5BQuant:BF16Ctx Length:32kPublished:Mar 27, 2026Architecture:Transformer Warm

The anirvankrishna/model_sft_dare is a 1.5 billion parameter language model with a 32768 token context length, created by anirvankrishna. This model was developed using the DARE TIES merge method, based on Qwen/Qwen2.5-1.5B-Instruct, and incorporates anirvankrishna/model_sft_lora_fused. It is specifically designed to leverage the DARE TIES merging technique for improved performance characteristics derived from its constituent models.

Loading preview...