erdemyavuz/llama-3-8b-chat-doctor
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:8kPublished:Dec 9, 2024License:llama3Architecture:Transformer0.0K Cold

The erdemyavuz/llama-3-8b-chat-doctor is an 8 billion parameter Llama-3-8B-Instruct model, fine-tuned by Erdem Yavuz for medical and healthcare-related conversations. Quantized to GGUF (Q4_K_M) format, it offers efficient, lightweight inference on CPUs, reducing memory footprint from ~16GB to ~4.9GB. This model excels as an AI medical assistant, optimized for simulating patient-doctor dialogues and building healthcare-assistive UI applications.

Loading preview...