Rateesh12/qwen-story-model
TEXT GENERATIONConcurrency Cost:1Model Size:2BQuant:BF16Ctx Length:32kPublished:Apr 13, 2026Architecture:Transformer Cold
The Rateesh12/qwen-story-model is a 2 billion parameter language model fine-tuned from Qwen/Qwen3-1.7B. Developed by Rateesh12, this model specializes in generating narrative content and creative text. It was trained using the TRL framework, making it suitable for story generation and conversational applications requiring imaginative responses.
Loading preview...
Model Overview
The Rateesh12/qwen-story-model is a 2 billion parameter language model, fine-tuned from the Qwen/Qwen3-1.7B base model. This model has been specifically trained using the TRL framework to enhance its capabilities in generating creative and narrative text.
Key Capabilities
- Story Generation: Optimized for producing coherent and imaginative stories.
- Creative Text Generation: Excels at generating diverse forms of creative content.
- Conversational AI: Suitable for applications requiring engaging and imaginative responses in dialogue.
Training Details
The model underwent Supervised Fine-Tuning (SFT) using the TRL framework. The training environment utilized:
- TRL: 1.2.0
- Transformers: 5.0.0
- Pytorch: 2.10.0+cu128
- Datasets: 4.8.4
- Tokenizers: 0.22.2
Good For
- Developers looking for a compact model for creative writing tasks.
- Applications requiring imaginative story prompts or narrative completion.
- Integrating into chatbots or virtual assistants that need to generate engaging, story-like responses.