Update README.md
Browse files
README.md
CHANGED
@@ -11,7 +11,7 @@ tags:
|
|
11 |
# MentaLLaMA-chat-13B
|
12 |
|
13 |
MentaLLaMA-chat-13B is part of the [MentaLLaMA](https://github.com/SteveKGYang/MentalLLaMA) project, the first open-source large language model (LLM) series for
|
14 |
-
interpretable mental health analysis with instruction-following capability.
|
15 |
The model is expected to make complex mental health analyses for various mental health conditions and give reliable explanations for each of its predictions.
|
16 |
It is fine-tuned on the IMHI dataset with 75K high-quality natural language instructions to boost its performance in downstream tasks.
|
17 |
We perform a comprehensive evaluation on the IMHI benchmark with 20K test samples. The result shows that MentalLLaMA approaches state-of-the-art discriminative
|
@@ -21,11 +21,16 @@ methods in correctness and generates high-quality explanations.
|
|
21 |
|
22 |
In addition to MentaLLaMA-chat-13B, the MentaLLaMA project includes another model: MentaLLaMA-chat-7B, MentalBART, MentalT5.
|
23 |
|
24 |
-
- **MentaLLaMA-chat-7B**: This model
|
|
|
25 |
|
26 |
-
- **MentalBART**: This model
|
|
|
|
|
27 |
|
28 |
-
- **MentalT5**: This model
|
|
|
|
|
29 |
|
30 |
## Usage
|
31 |
|
@@ -44,11 +49,6 @@ use the GPU if it's available.
|
|
44 |
|
45 |
MentaLLaMA-chat-13B is licensed under MIT. For more details, please see the MIT file.
|
46 |
|
47 |
-
## About
|
48 |
-
|
49 |
-
This model is part of the MentaLLaMA project.
|
50 |
-
For more information, you can visit the [MentaLLaMA](https://github.com/SteveKGYang/MentalLLaMA) project on GitHub.
|
51 |
-
|
52 |
## Citation
|
53 |
|
54 |
If you use MentaLLaMA-chat-7B in your work, please cite our paper:
|
|
|
11 |
# MentaLLaMA-chat-13B
|
12 |
|
13 |
MentaLLaMA-chat-13B is part of the [MentaLLaMA](https://github.com/SteveKGYang/MentalLLaMA) project, the first open-source large language model (LLM) series for
|
14 |
+
interpretable mental health analysis with instruction-following capability. This model is finetuned based on the Meta LLaMA2-chat-13B foundation model and the full IMHI instruction tuning data.
|
15 |
The model is expected to make complex mental health analyses for various mental health conditions and give reliable explanations for each of its predictions.
|
16 |
It is fine-tuned on the IMHI dataset with 75K high-quality natural language instructions to boost its performance in downstream tasks.
|
17 |
We perform a comprehensive evaluation on the IMHI benchmark with 20K test samples. The result shows that MentalLLaMA approaches state-of-the-art discriminative
|
|
|
21 |
|
22 |
In addition to MentaLLaMA-chat-13B, the MentaLLaMA project includes another model: MentaLLaMA-chat-7B, MentalBART, MentalT5.
|
23 |
|
24 |
+
- **MentaLLaMA-chat-7B**: This model is finetuned based on the Meta LLaMA2-chat-7B foundation model and the full IMHI instruction tuning data. The training data covers
|
25 |
+
10 mental health analysis tasks.
|
26 |
|
27 |
+
- **MentalBART**: This model is finetuned based on the BART-large foundation model and the full IMHI-completion data. The training data covers
|
28 |
+
10 mental health analysis tasks. This model doesn't have instruction-following ability but is more lightweight and performs well in interpretable mental health analysis
|
29 |
+
in a completion-based manner.
|
30 |
|
31 |
+
- **MentalT5**: This model is finetuned based on the T5-large foundation model and the full IMHI-completion data. The training data covers
|
32 |
+
10 mental health analysis tasks. This model doesn't have instruction-following ability but is more lightweight and performs well in interpretable mental health analysis
|
33 |
+
in a completion-based manner.
|
34 |
|
35 |
## Usage
|
36 |
|
|
|
49 |
|
50 |
MentaLLaMA-chat-13B is licensed under MIT. For more details, please see the MIT file.
|
51 |
|
|
|
|
|
|
|
|
|
|
|
52 |
## Citation
|
53 |
|
54 |
If you use MentaLLaMA-chat-7B in your work, please cite our paper:
|