souradeepmukhopadhyay99/qwen3-4b-apigenmt-5k-trl-fullft
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Jan 19, 2026Architecture:Transformer Warm

The souradeepmukhopadhyay99/qwen3-4b-apigenmt-5k-trl-fullft model is a 4 billion parameter causal language model, fine-tuned from Qwen/Qwen3-4B-Instruct-2507. It was trained using the TRL framework with Supervised Fine-Tuning (SFT) methods. This model is designed for general text generation tasks, leveraging its Qwen3 architecture and fine-tuning for improved instruction following. It offers a context length of 40960 tokens, making it suitable for applications requiring extensive input processing.

Loading preview...