Text Generation
Transformers
Safetensors
Finnish
bloom
text-generation-inference
Inference Endpoints
mpasila commited on
Commit
1e72606
1 Parent(s): 8640f52

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -2
README.md CHANGED
@@ -8,7 +8,9 @@ license: apache-2.0
8
 
9
  # Model Card for Capybara-Finnish-V1-8B
10
 
11
- This is a merge of [mpasila/Capybara-Finnish-V1-8B-LoRA](https://huggingface.co/mpasila/Capybara-Finnish-V1-8B-LoRA/). Base model used: [mpasila/gpt3-finnish-8B-gptq-4bit](https://huggingface.co/mpasila/gpt3-finnish-8B-gptq-4bit/) and the original unquantized model: [TurkuNLP/gpt3-finnish-8B](https://huggingface.co/TurkuNLP/gpt3-finnish-8B/). Dataset used with the LoRA is [Finnish-NLP/Capybara-fi-deepl-translated-sft](https://huggingface.co/datasets/Finnish-NLP/Capybara-fi-deepl-translated-sft/) with some modifications so it uses Alpaca formatting [modified dataset](https://huggingface.co/datasets/mpasila/Capybara-fi-deepl-translated-sft-alpaca/).
 
 
12
 
13
  It uses Alpaca format but with a translated instruction at the start:
14
  ```
@@ -19,7 +21,7 @@ It uses Alpaca format but with a translated instruction at the start:
19
 
20
  ```
21
 
22
- Merged using this [Colab notebook](https://colab.research.google.com/drive/1a76Y21GfPtmVs71Uztlgk2xzPA4_vVjs?usp=sharing).
23
 
24
  ### Framework versions
25
 
 
8
 
9
  # Model Card for Capybara-Finnish-V1-8B
10
 
11
+ This is a merge of [mpasila/Capybara-Finnish-V1-8B-LoRA](https://huggingface.co/mpasila/Capybara-Finnish-V1-8B-LoRA/).
12
+
13
+ Base model used: [mpasila/gpt3-finnish-8B-gptq-4bit](https://huggingface.co/mpasila/gpt3-finnish-8B-gptq-4bit/) and the original unquantized model: [TurkuNLP/gpt3-finnish-8B](https://huggingface.co/TurkuNLP/gpt3-finnish-8B/). Dataset used with the LoRA is [Finnish-NLP/Capybara-fi-deepl-translated-sft](https://huggingface.co/datasets/Finnish-NLP/Capybara-fi-deepl-translated-sft/) with some modifications so it uses Alpaca formatting [modified dataset](https://huggingface.co/datasets/mpasila/Capybara-fi-deepl-translated-sft-alpaca/).
14
 
15
  It uses Alpaca format but with a translated instruction at the start:
16
  ```
 
21
 
22
  ```
23
 
24
+ Merged using this [Colab notebook](https://colab.research.google.com/drive/1a76Y21GfPtmVs71Uztlgk2xzPA4_vVjs?usp=sharing). It might not be the best way to merge a quantized LoRA on to a float16 model but I just wanted to quickly do something. You can try merging it better if you want.
25
 
26
  ### Framework versions
27