klyang/MentaLLaMA-chat-13B

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:13BQuant:FP8Ctx Length:4kPublished:Sep 27, 2023License:mitArchitecture:Transformer0.0K Open Weights Warm

MentaLLaMA-chat-13B, developed by klyang, is a 13 billion parameter instruction-following large language model fine-tuned from Meta's LLaMA2-chat-13B. It is specifically designed for interpretable mental health analysis, providing predictions and explanations for various mental health conditions. The model was trained on the IMHI dataset, comprising 75K high-quality natural language instructions, and is intended for non-clinical research applications.

Loading preview...

MentaLLaMA-chat-13B: Interpretable Mental Health Analysis

MentaLLaMA-chat-13B is a 13 billion parameter model from the MentaLLaMA project, developed by klyang. It is fine-tuned from Meta's LLaMA2-chat-13B using the comprehensive IMHI instruction tuning dataset, which includes 75,000 high-quality natural language instructions.

Key Capabilities

  • Interpretable Mental Health Analysis: Designed to perform complex mental health analyses and provide reliable explanations for its predictions across various mental health conditions.
  • Instruction-Following: Enhanced with instruction-following capabilities through fine-tuning on the IMHI dataset.
  • Performance: Approaches state-of-the-art discriminative methods in correctness on the IMHI benchmark (20K test samples) while generating high-quality explanations.

Ethical Considerations & Limitations

While showing promising performance, MentaLLaMA-chat-13B is strictly intended for non-clinical research purposes. Users seeking assistance should consult professional psychiatrists or clinical practitioners. The developers acknowledge potential biases (e.g., gender gaps), incorrect predictions, inappropriate explanations, and over-generalization as existing challenges, highlighting that significant work remains for real-world mental health monitoring system applications.

Good For

  • Researchers in AI and mental health exploring interpretable models.
  • Developing tools for non-clinical mental health analysis and explanation generation.
  • Investigating the application of LLMs in understanding and explaining mental health-related text.