anirvankrishna/model_sft_resta
TEXT GENERATIONConcurrency Cost:1Model Size:1.5BQuant:BF16Ctx Length:32kPublished:Mar 29, 2026Architecture:Transformer Loading
The anirvankrishna/model_sft_resta is a 1.5 billion parameter language model based on the Qwen2.5-1.5B-Instruct architecture, featuring a 32768-token context length. This model is a merge of Qwen/Qwen2.5-1.5B-Instruct and anirvankrishna/model_harmful_lora_fused, created using the Task Arithmetic method. Its primary characteristic is the specific merging approach, which can be used to explore model behavior when combining different pre-trained components.
Loading preview...