anirvankrishna/model_sft_resta_dare
TEXT GENERATIONConcurrency Cost:1Model Size:1.5BQuant:BF16Ctx Length:32kPublished:Mar 29, 2026Architecture:Transformer Loading

The anirvankrishna/model_sft_resta_dare is a 1.5 billion parameter language model, based on the Qwen2.5-1.5B-Instruct architecture, developed by anirvankrishna. This model was created using a Task Arithmetic merge method, combining the base Qwen model with anirvankrishna/model_harmful_lora_fused. It is designed to leverage the strengths of its merged components, offering a 32K context length. The primary application of this model is to provide a refined language generation capability derived from its specific merging strategy.

Loading preview...