klyang commited on
Commit
7a88fec
1 Parent(s): 267a852

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +9 -9
README.md CHANGED
@@ -11,7 +11,7 @@ tags:
11
  # MentaLLaMA-chat-13B
12
 
13
  MentaLLaMA-chat-13B is part of the [MentaLLaMA](https://github.com/SteveKGYang/MentalLLaMA) project, the first open-source large language model (LLM) series for
14
- interpretable mental health analysis with instruction-following capability.
15
  The model is expected to make complex mental health analyses for various mental health conditions and give reliable explanations for each of its predictions.
16
  It is fine-tuned on the IMHI dataset with 75K high-quality natural language instructions to boost its performance in downstream tasks.
17
  We perform a comprehensive evaluation on the IMHI benchmark with 20K test samples. The result shows that MentalLLaMA approaches state-of-the-art discriminative
@@ -21,11 +21,16 @@ methods in correctness and generates high-quality explanations.
21
 
22
  In addition to MentaLLaMA-chat-13B, the MentaLLaMA project includes another model: MentaLLaMA-chat-7B, MentalBART, MentalT5.
23
 
24
- - **MentaLLaMA-chat-7B**: This model
 
25
 
26
- - **MentalBART**: This model
 
 
27
 
28
- - **MentalT5**: This model
 
 
29
 
30
  ## Usage
31
 
@@ -44,11 +49,6 @@ use the GPU if it's available.
44
 
45
  MentaLLaMA-chat-13B is licensed under MIT. For more details, please see the MIT file.
46
 
47
- ## About
48
-
49
- This model is part of the MentaLLaMA project.
50
- For more information, you can visit the [MentaLLaMA](https://github.com/SteveKGYang/MentalLLaMA) project on GitHub.
51
-
52
  ## Citation
53
 
54
  If you use MentaLLaMA-chat-7B in your work, please cite our paper:
 
11
  # MentaLLaMA-chat-13B
12
 
13
  MentaLLaMA-chat-13B is part of the [MentaLLaMA](https://github.com/SteveKGYang/MentalLLaMA) project, the first open-source large language model (LLM) series for
14
+ interpretable mental health analysis with instruction-following capability. This model is finetuned based on the Meta LLaMA2-chat-13B foundation model and the full IMHI instruction tuning data.
15
  The model is expected to make complex mental health analyses for various mental health conditions and give reliable explanations for each of its predictions.
16
  It is fine-tuned on the IMHI dataset with 75K high-quality natural language instructions to boost its performance in downstream tasks.
17
  We perform a comprehensive evaluation on the IMHI benchmark with 20K test samples. The result shows that MentalLLaMA approaches state-of-the-art discriminative
 
21
 
22
  In addition to MentaLLaMA-chat-13B, the MentaLLaMA project includes another model: MentaLLaMA-chat-7B, MentalBART, MentalT5.
23
 
24
+ - **MentaLLaMA-chat-7B**: This model is finetuned based on the Meta LLaMA2-chat-7B foundation model and the full IMHI instruction tuning data. The training data covers
25
+ 10 mental health analysis tasks.
26
 
27
+ - **MentalBART**: This model is finetuned based on the BART-large foundation model and the full IMHI-completion data. The training data covers
28
+ 10 mental health analysis tasks. This model doesn't have instruction-following ability but is more lightweight and performs well in interpretable mental health analysis
29
+ in a completion-based manner.
30
 
31
+ - **MentalT5**: This model is finetuned based on the T5-large foundation model and the full IMHI-completion data. The training data covers
32
+ 10 mental health analysis tasks. This model doesn't have instruction-following ability but is more lightweight and performs well in interpretable mental health analysis
33
+ in a completion-based manner.
34
 
35
  ## Usage
36
 
 
49
 
50
  MentaLLaMA-chat-13B is licensed under MIT. For more details, please see the MIT file.
51
 
 
 
 
 
 
52
  ## Citation
53
 
54
  If you use MentaLLaMA-chat-7B in your work, please cite our paper: