kmseong/llama3.2_3b_SSFT_epoch5_adam_lr4
TEXT GENERATIONConcurrency Cost:1Model Size:3.2BQuant:BF16Ctx Length:32kPublished:Apr 3, 2026License:llama3.2Architecture:Transformer Loading

The kmseong/llama3.2_3b_SSFT_epoch5_adam_lr4 is a 3.2 billion parameter Llama 3.2 architecture model, developed by Min-Seong Kim, that has undergone Phase 0 of Safety-WaRP (Weight space Rotation Process) training. This model is specifically fine-tuned on the Circuit Breakers dataset to establish base safety mechanisms, making it suitable for applications requiring initial safety filtering. It features a 32768 token context length and is designed as a foundational model for further safety and utility enhancements.

Loading preview...