Sandeep0079/model_sft_dare_resta
TEXT GENERATIONConcurrency Cost:1Model Size:1.5BQuant:BF16Ctx Length:32kPublished:Apr 5, 2026Architecture:Transformer Loading
Sandeep0079/model_sft_dare_resta is a 1.5 billion parameter language model merged from Qwen/Qwen2.5-1.5B-Instruct and two specialized models using the linear merge method. This model is designed to integrate specific characteristics from its constituent models, offering a unique blend of capabilities derived from its base and specialized components. With a 32768 token context length, it aims to provide nuanced responses based on its merged training. Its primary differentiator lies in its unique merging strategy, which combines a base instruction-tuned model with custom 'dare' and 'harmful' models.
Loading preview...