deep-div/MediLlama-3.2
TEXT GENERATIONConcurrency Cost:1Model Size:3.2BQuant:BF16Ctx Length:32kPublished:May 16, 2025License:apache-2.0Architecture:Transformer0.0K Open Weights Warm

deep-div/MediLlama-3.2 is a 3.2 billion parameter, 32768-token context length causal language model developed by InferenceLab. Fine-tuned from Meta's LLaMA 3.2 3B Instruct, it is specifically optimized for English-language medical and healthcare applications. This model excels at tasks like medical Q&A, symptom checking, and patient education, serving as a specialized medical chatbot.

Loading preview...