Update README.md
Browse files
README.md
CHANGED
@@ -8,7 +8,9 @@ license: apache-2.0
|
|
8 |
|
9 |
# Model Card for Capybara-Finnish-V1-8B
|
10 |
|
11 |
-
This is a merge of [mpasila/Capybara-Finnish-V1-8B-LoRA](https://huggingface.co/mpasila/Capybara-Finnish-V1-8B-LoRA/).
|
|
|
|
|
12 |
|
13 |
It uses Alpaca format but with a translated instruction at the start:
|
14 |
```
|
@@ -19,7 +21,7 @@ It uses Alpaca format but with a translated instruction at the start:
|
|
19 |
|
20 |
```
|
21 |
|
22 |
-
Merged using this [Colab notebook](https://colab.research.google.com/drive/1a76Y21GfPtmVs71Uztlgk2xzPA4_vVjs?usp=sharing).
|
23 |
|
24 |
### Framework versions
|
25 |
|
|
|
8 |
|
9 |
# Model Card for Capybara-Finnish-V1-8B
|
10 |
|
11 |
+
This is a merge of [mpasila/Capybara-Finnish-V1-8B-LoRA](https://huggingface.co/mpasila/Capybara-Finnish-V1-8B-LoRA/).
|
12 |
+
|
13 |
+
Base model used: [mpasila/gpt3-finnish-8B-gptq-4bit](https://huggingface.co/mpasila/gpt3-finnish-8B-gptq-4bit/) and the original unquantized model: [TurkuNLP/gpt3-finnish-8B](https://huggingface.co/TurkuNLP/gpt3-finnish-8B/). Dataset used with the LoRA is [Finnish-NLP/Capybara-fi-deepl-translated-sft](https://huggingface.co/datasets/Finnish-NLP/Capybara-fi-deepl-translated-sft/) with some modifications so it uses Alpaca formatting [modified dataset](https://huggingface.co/datasets/mpasila/Capybara-fi-deepl-translated-sft-alpaca/).
|
14 |
|
15 |
It uses Alpaca format but with a translated instruction at the start:
|
16 |
```
|
|
|
21 |
|
22 |
```
|
23 |
|
24 |
+
Merged using this [Colab notebook](https://colab.research.google.com/drive/1a76Y21GfPtmVs71Uztlgk2xzPA4_vVjs?usp=sharing). It might not be the best way to merge a quantized LoRA on to a float16 model but I just wanted to quickly do something. You can try merging it better if you want.
|
25 |
|
26 |
### Framework versions
|
27 |
|