Chaotically/model_sft_dare_resta
TEXT GENERATIONConcurrency Cost:1Model Size:1.5BQuant:BF16Ctx Length:32kPublished:Apr 3, 2026Architecture:Transformer Cold

Chaotically/model_sft_dare_resta is a 1.5 billion parameter language model created by Chaotically using a Task Arithmetic merge. This model combines 'dare_model_temp' and 'harmful_model_temp' with 'base_safe_temp' as its foundation, leveraging a 32768 token context length. Its unique merging strategy suggests a specialized focus, potentially for safety or specific content generation, by differentially weighting model components.

Loading preview...

Model Overview

Chaotically/model_sft_dare_resta is a 1.5 billion parameter language model developed by Chaotically. It was created using the Task Arithmetic merge method, which combines the strengths of multiple pre-trained models into a single, specialized model. The base model for this merge was base_safe_temp.

Merge Details

This model is a composite of three distinct components:

  • base_safe_temp: Served as the foundational model.
  • dare_model_temp: Integrated with a positive weight (1.0).
  • harmful_model_temp: Integrated with a negative weight (-1.0).

This specific configuration, particularly the negative weighting of harmful_model_temp, indicates an intentional design choice to influence the model's behavior, likely towards enhancing safety or mitigating undesirable outputs. The model supports a substantial context length of 32768 tokens.

Potential Use Cases

Given its unique merge strategy, this model could be particularly well-suited for applications requiring:

  • Content moderation: By potentially reducing the generation of harmful content.
  • Safe text generation: Where adherence to safety guidelines is paramount.
  • Exploration of model merging techniques: For researchers interested in the effects of Task Arithmetic with differential weighting.