Update README.md
Browse files
README.md
CHANGED
@@ -1,3 +1,25 @@
|
|
1 |
---
|
2 |
license: mit
|
3 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
---
|
2 |
license: mit
|
3 |
---
|
4 |
+
### acsr-y34b-4bpw-hb6-exl2
|
5 |
+
|
6 |
+
- base model: [Yi-34B-Chat](https://huggingface.co/01-ai/Yi-34B-Chat)
|
7 |
+
- LoRA: [Yi-34b-alpaca-cot-lora](https://huggingface.co/zzlgreat/Yi-34b-alpaca-cot-lora) 支持Alpaca格式对话,效果很差不推荐使用Alpaca指令格式
|
8 |
+
- LoRA: [Yi-34B-Spicyboros-3.1-LoRA](https://huggingface.co/LoneStriker/Yi-34B-Spicyboros-3.1-LoRA) 非官方对话数据集
|
9 |
+
- LoRA: [limarpv3-yi-llama-34b-lora](https://huggingface.co/Doctor-Shotgun/limarpv3-yi-llama-34b-lora) 扮演类长回复
|
10 |
+
- Instruction template: ChatML
|
11 |
+
- Origin max token size: 4096
|
12 |
+
- 在使用text-generation-webui作后台,SillyTavern做前端时,webui上MaxToken调到8K时SillyTavern的MaxToken要调整到18K,WebUI才不会提前截断
|
13 |
+
- 在SillyTavern上要得到Roleplay样式的长回复
|
14 |
+
- 使用ChatML作为指令格式
|
15 |
+
- 最后输出序列调整为 `### Response (2 paragraphs, engaging, natural, authentic, descriptive, creative):`
|
16 |
+
- 自定义停止符调整为 `["<|im_end|>用户\n", "\n### Input"]`
|
17 |
+
- 系统提示调整为 `Avoid repetition, don't loop. Develop the plot slowly, always stay in character. Describe all actions in full, elaborate, explicit, graphic, and vivid detail. Mention all relevant sensory perceptions.`
|
18 |
+
|
19 |
+
### description
|
20 |
+
|
21 |
+
- This is test for [exllamav2](https://github.com/turboderp/exllamav2) model.
|
22 |
+
- 4bpw `python convert.py -i acsr-v2-y34b -c exl2/0000.parquet -o acsr-v2-y34b-4bpw-hb6-exl2 -hb 6 -l 4096 -b 4.15`
|
23 |
+
- [convert doc](https://github.com/turboderp/exllamav2/blob/master/doc/convert.md)
|
24 |
+
- calibration dataset: [WikiText-2-v1](https://huggingface.co/datasets/wikitext/blob/refs%2Fconvert%2Fparquet/wikitext-2-v1/test/0000.parquet)
|
25 |
+
- oobabooga/text-generation-webui must add `--trust-remote-code` into CMD_FLAGS.txt and use ExLlamav2 to load model
|