gshasiri/SmolLM3-SFT
TEXT GENERATIONConcurrency Cost:1Model Size:1BQuant:BF16Ctx Length:32kPublished:Nov 17, 2025Architecture:Transformer Warm
SmolLM3-SFT by gshasiri is a 1 billion parameter instruction-tuned causal language model, fine-tuned from gshasiri/SmolLM3-Mid using the TRL framework. This model is optimized for conversational AI and instruction following, leveraging its compact size for efficient deployment. With a 32768 token context length, it is suitable for applications requiring processing of longer prompts and generating coherent, extended responses.
Loading preview...