kmseong/llama2_7b-chat-Safety-FT-lr5e-5

TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Apr 20, 2026License:llama3.2Architecture:Transformer Cold

The kmseong/llama2_7b-chat-Safety-FT-lr5e-5 is a 7 billion parameter Llama 2-based chat model, fine-tuned for safety alignment. This model incorporates attention mechanisms with q,k,v MLP application and per-layer adjustments, followed by non-freeze training. It is designed to enhance safety in conversational AI applications, offering a context length of 4096 tokens.

Loading preview...

Model Overview

The kmseong/llama2_7b-chat-Safety-FT-lr5e-5 is a 7 billion parameter model built upon the Llama 2 architecture, specifically fine-tuned for safety alignment in chat-based applications. This model differentiates itself through its training methodology, which involves applying MLP to the attention mechanism's q, k, and v components, alongside per-layer adjustments. The training process then proceeds with a non-freeze approach, allowing for comprehensive adaptation.

Key Capabilities

  • Safety Alignment: Primary focus on enhancing safety in conversational AI.
  • Llama 2 Base: Leverages the robust Llama 2 7B architecture.
  • Custom Training: Utilizes specific modifications to attention and MLP layers during fine-tuning.
  • Context Length: Supports a context window of 4096 tokens.

Good For

  • Developing safer conversational agents.
  • Research into safety alignment techniques for large language models.
  • Applications requiring a Llama 2-based model with enhanced safety characteristics.

Citation

This model is associated with the following citation:

@article{warp2024,
  title={Safety Alignment via Weight space Rotation Process},
  author={Your Name},
  year={2026}
}