rohan2810/NEW_BASELINE_SFT_hotpotqa_Qwen3-4B-Instruct
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Mar 23, 2026Architecture:Transformer Warm

rohan2810/NEW_BASELINE_SFT_hotpotqa_Qwen3-4B-Instruct is a 4 billion parameter instruction-tuned language model based on the Qwen3 architecture, developed by rohan2810. This model is fine-tuned for specific tasks, indicated by 'hotpotqa' in its name, suggesting optimization for question answering over complex, multi-hop information. With a context length of 32768 tokens, it is designed for processing and generating responses based on extensive input.

Loading preview...