Athkal/model-sft-resta
TEXT GENERATIONConcurrency Cost:1Model Size:1.5BQuant:BF16Ctx Length:32kPublished:Mar 22, 2026Architecture:Transformer Warm

Athkal/model-sft-resta is a merged language model created using the Task Arithmetic method, combining Athkal/model-sft-lora as a base with Qwen/Qwen2.5-1.5B-Instruct and a local model. This model leverages the Qwen2.5-1.5B-Instruct architecture, making it suitable for general instruction-following tasks. Its unique merging approach suggests a focus on refining specific behaviors or capabilities derived from its constituent models.

Loading preview...