Update README.md
Browse files
README.md
CHANGED
@@ -21,15 +21,16 @@ datasets:
|
|
21 |
# Model description
|
22 |
Heidrun-Mistral-7B-chat is a chat-model based on [Heidrun-Mistral-7B-base](https://huggingface.co/Mabeck/Heidrun-Mistral-7B-base), finetuned on [danish-OpenHermes](https://huggingface.co/datasets/Mabeck/danish-OpenHermes) and [skoleGPT](https://huggingface.co/datasets/kobprof/skolegpt-instruct) for a instruction/chat format.
|
23 |
|
|
|
24 |
|
25 |
# Datasets
|
26 |
-
This model is trained on Danish instruction datasets, which have not been safeguarded or alligned.
|
27 |
|
28 |
Most of the data has been machine-translated and may contain incorrect responses.
|
29 |
|
30 |
# Samples
|
31 |
|
32 |
-
This model uses the ChatML format. Using other formats will severely degrade the
|
33 |
ChatML format:
|
34 |
```
|
35 |
<|im_start|>system
|
|
|
21 |
# Model description
|
22 |
Heidrun-Mistral-7B-chat is a chat-model based on [Heidrun-Mistral-7B-base](https://huggingface.co/Mabeck/Heidrun-Mistral-7B-base), finetuned on [danish-OpenHermes](https://huggingface.co/datasets/Mabeck/danish-OpenHermes) and [skoleGPT](https://huggingface.co/datasets/kobprof/skolegpt-instruct) for a instruction/chat format.
|
23 |
|
24 |
+
It achieved 35.66%+-0.85%/51.68+-0.63% on MMLU-Da in [ScandEval](https://github.com/ScandEval/ScandEval) and achieves the best non-merged opensource model score to date (13/02/24).
|
25 |
|
26 |
# Datasets
|
27 |
+
This model is trained on Danish instruction datasets [danish-OpenHermes](Mabeck/danish-OpenHermes) and [skoleGPT](https://huggingface.co/datasets/kobprof/skolegpt-instruct), which have not been safeguarded or alligned.
|
28 |
|
29 |
Most of the data has been machine-translated and may contain incorrect responses.
|
30 |
|
31 |
# Samples
|
32 |
|
33 |
+
This model uses the ChatML format. Using other formats will severely degrade the model's performance.
|
34 |
ChatML format:
|
35 |
```
|
36 |
<|im_start|>system
|