thrnn/qwen2.5-1.5b-sft-dare-resta
TEXT GENERATIONConcurrency Cost:1Model Size:1.5BQuant:BF16Ctx Length:32kPublished:Apr 2, 2026Architecture:Transformer Cold

thrnn/qwen2.5-1.5b-sft-dare-resta is a 1.5 billion parameter language model based on the Qwen2.5 architecture, created by thrnn. This model is a merge using the Task Arithmetic method, combining Qwen/Qwen2.5-1.5B-Instruct with a 'harmful_lora' component, using thrnn/qwen2.5-1.5b-medical-sft-dare as its base. Its unique merging strategy suggests a focus on modifying or exploring specific behavioral characteristics of the base model, potentially for safety research or content moderation. The model has a context length of 32768 tokens.

Loading preview...