araziziml/sft_trainer
TEXT GENERATIONConcurrency Cost:2Model Size:32.8BQuant:FP8Ctx Length:32kArchitecture:Transformer Cold

The araziziml/sft_trainer is a 32.8 billion parameter instruction-tuned causal language model, fine-tuned from Qwen/Qwen2.5-32B-Instruct using the TRL library. This model is optimized for following instructions and generating text based on user prompts, leveraging its large parameter count and extensive context length of 131,072 tokens for complex tasks. It is suitable for applications requiring advanced natural language understanding and generation capabilities.

Loading preview...