TUSGPT-TR-Medical-9B: Turkish Medical Language Model
TUSGPT-TR-Medical-9B, developed by turkerberkdonmez, is Turkey's first open-source Turkish medical language model. Built upon the Gemma-2 architecture with 9 billion parameters, it specializes in medical AI research and applications within the Turkish language.
Key Capabilities & Training
- Specialized Fine-tuning: The model is fine-tuned on the
ytu-ce-cosmos/Turkish-Gemma-9b-T1 base model. - Extensive Medical Dataset: It leverages 55,465 high-quality Turkish medical question-answer pairs from the
turkerberkdonmez/TUSGPT-TR-Medical-Dataset-v1 dataset. - Broad Medical Coverage: The training data spans numerous medical disciplines, including Basic Sciences (Pharmacology, Pathology, Anatomy, Physiology) and Clinical Sciences (Internal Medicine, Surgery, Pediatrics, Obstetrics & Gynecology), as well as Emergency Medicine, Neurology, Oncology, and Radiology.
- Advanced Training Methodology: Utilizes a 2-Stage DoRA (Weight-Decomposed Low-Rank Adaptation) method, trained on Apple Silicon hardware with bfloat16 precision.
Usage & Recommendations
- Python (Transformers): Provides clear examples for integration, recommending
temperature=0.6, top_p=0.95, top_k=20, and min_p=0 for generation. - GGUF Quantized Versions: Available in
Q8_0 (9.8 GB) and Q4_K_M (5.8 GB) for local deployment with Ollama and LM Studio, suitable for various hardware configurations.
Important Disclaimer
This model is intended for educational and research purposes only and should not be used as the sole basis for clinical decision-making. Users are advised to consult qualified physicians for medical decisions.