rohan2810/BASELINE_SFT_lastfm_Qwen3-4B-Instruct-2507
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Mar 10, 2026Architecture:Transformer Cold

The rohan2810/BASELINE_SFT_lastfm_Qwen3-4B-Instruct-2507 is a 4 billion parameter instruction-tuned model based on the Qwen3 architecture, developed by rohan2810. This model is designed for general language understanding and generation tasks, leveraging a 32768 token context length. Its primary application is in conversational AI and instruction following, providing a robust foundation for various NLP applications.

Loading preview...