kmseong/llama3.2_3b_instruct-WaRP-safety-basis-MATH-FT-lr1e-7

TEXT GENERATIONConcurrency Cost:1Model Size:3.2BQuant:BF16Ctx Length:32kPublished:Apr 11, 2026License:llama3.2Architecture:Transformer Cold

The kmseong/llama3.2_3b_instruct-WaRP-safety-basis-MATH-FT-lr1e-7 is a 3.2 billion parameter instruction-tuned language model, based on the Llama 3.2 architecture, developed by kmseong. This model incorporates a Weight space Rotation Process (WaRP) for safety alignment and is further fine-tuned for mathematical tasks. It is designed to provide enhanced safety characteristics while maintaining performance in mathematical reasoning, making it suitable for applications requiring both robust safety and numerical accuracy.

Loading preview...

Model Overview

The kmseong/llama3.2_3b_instruct-WaRP-safety-basis-MATH-FT-lr1e-7 is an instruction-tuned language model built upon the Llama 3.2 architecture, featuring 3.2 billion parameters. Developed by kmseong, this model integrates a specialized Weight space Rotation Process (WaRP) for safety alignment, aiming to enhance its robustness against undesirable outputs. Following this safety alignment, the model undergoes further fine-tuning specifically for mathematical tasks, indicated by "MATH-FT".

Key Capabilities

  • Safety Alignment: Utilizes the WaRP method to improve model safety characteristics.
  • Mathematical Fine-Tuning: Optimized for performance in mathematical reasoning and problem-solving.
  • Instruction Following: Designed to respond effectively to instructions, typical of instruct-tuned models.

Good For

  • Applications requiring a balance of safety and mathematical proficiency.
  • Tasks involving numerical reasoning where safety is a critical concern.
  • Instruction-based interactions in domains that benefit from mathematical understanding.