danielkty22/TARS-SFT-1.5B
TEXT GENERATIONConcurrency Cost:1Model Size:1.5BQuant:BF16Ctx Length:32kPublished:Jul 16, 2025License:apache-2.0Architecture:Transformer Open Weights Warm

danielkty22/TARS-SFT-1.5B is a 1.5 billion parameter SFT-tuned reasoning model, developed for safety within the TARS (Training Adaptive Reasoners for Safety) framework. Based on Qwen2.5-1.5B-Instruct, this model serves as the base for reinforcement learning training. It is specifically designed to enhance adaptive reasoning capabilities for safety-critical applications, as detailed in the paper "Reasoning as an Adaptive Defense for Safety". Its primary use case is as a foundational component for developing more robust and safer AI systems through advanced reasoning. This model has a context length of 131072 tokens.

Loading preview...