SakanaAI/RLT-32B
TEXT GENERATIONConcurrency Cost:2Model Size:32.8BQuant:FP8Ctx Length:32kLicense:apache-2.0Architecture:Transformer0.0K Open Weights Cold

SakanaAI/RLT-32B is a 32.8 billion parameter autoregressive language model developed by Sakana AI. This student model was trained using the Reinforcement-Learned Teachers (RLT) pipeline, distilled from a 7B Reinforcement-Learned Teacher optimized for producing high-quality reasoning traces. It is specifically fine-tuned with a focus on reasoning, making it suitable for research and development in advanced AI reasoning tasks.

Loading preview...