xw1234gan/Merging_Qwen2.5-1.5B-Instruct_MedQA_lr1e-05_mb2_ga128_n2048_seed42
TEXT GENERATIONConcurrency Cost:1Model Size:1.5BQuant:BF16Ctx Length:32kPublished:Mar 16, 2026Architecture:Transformer Warm

This is a 1.5 billion parameter instruction-tuned model developed by xw1234gan, based on the Qwen2.5 architecture. It is fine-tuned for medical question answering (MedQA) tasks, leveraging a 32768 token context length. This model is specifically optimized for performance in medical domain inquiries.

Loading preview...