Henrychur/MMedS-Llama-3-8B

TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:8kPublished:Aug 30, 2024License:llama3Architecture:Transformer0.0K Cold

Henrychur/MMedS-Llama-3-8B is an 8 billion parameter multilingual medical language model, continuously pretrained on MMedC and fine-tuned with 13.5 million samples across 122 medical tasks. This model specializes in medical applications, building upon the Llama 3 architecture with an 8192-token context length. It is designed for versatile performance in various medical language processing scenarios.

Loading preview...

MMedS-Llama-3-8B Overview

MMedS-Llama-3-8B is an 8 billion parameter multilingual medical language model developed by Henrychur. It is built upon the Llama 3 foundation and has undergone specialized training to excel in medical domains. The model's development involved two key stages:

Key Capabilities & Training

  • Foundation Model: Based on MMed-Llama-3-8B, which is a multilingual medical language model.
  • Continuous Pretraining: Enhanced through continuous pretraining on the MMedC dataset.
  • Supervised Fine-Tuning (SFT): Further fine-tuned using MedS-Ins, a comprehensive dataset comprising 13.5 million samples across 122 distinct medical tasks. This extensive SFT process aims to imbue the model with versatile capabilities for various medical applications.
  • Context Length: Supports an 8192-token context length, consistent with its Llama 3 base.

Use Cases

This model is particularly well-suited for applications requiring advanced language understanding and generation within the medical field. Its specialized training on a vast array of medical tasks makes it a strong candidate for:

  • Medical text analysis
  • Clinical documentation assistance
  • Medical question answering
  • Research in medical language processing

For more in-depth information, refer to the official GitHub repository and the arXiv paper.