jdh-algo/Citrus1.0-llama-70B

Warm
Public
70B
FP8
32768
4
Feb 25, 2025
Hugging Face
Overview

Citrus1.0-llama-70B: Advanced Medical Decision Support

Citrus1.0-llama-70B is a 70 billion parameter medical language model developed by jdh-algo, built upon the Llama3.1-70B architecture. This model is uniquely designed to bridge the gap between clinical expertise and AI reasoning by emulating the cognitive pathways of medical professionals, offering a 32768 token context length.

Key Capabilities & Innovations

  • Expert Cognitive Pathway Emulation: Utilizes a novel training-free reasoning approach that simulates medical expert decision-making processes to enhance clinical diagnosis and treatment capabilities.
  • Specialized Training Data: Trained on a large corpus of simulated expert disease reasoning data, synthesized to accurately capture clinician decision pathways.
  • Multi-Stage Post-Training: Incorporates a multi-stage post-training methodology to further refine and improve medical performance.
  • Open-Source Resources: jdh-algo has made the Citrus model, its training data (Citrus_S3), and a large-scale, updatable clinical practice evaluation dataset (JMED) publicly available to foster research in AI-driven medical decision-making.

Ideal Use Cases

  • Clinical Diagnosis: Assisting in the diagnostic process by leveraging expert-like reasoning.
  • Treatment Planning: Supporting the development of treatment strategies based on emulated medical expertise.
  • Medical Research: Providing a foundation for further research in AI applications for healthcare, particularly in understanding and replicating human cognitive processes in medicine.