IHaveNoClueAndIMustPost
commited on
Commit
•
c00e3ae
1
Parent(s):
9849f66
Update README.md
Browse files
README.md
CHANGED
@@ -7,16 +7,14 @@ This is [llama2-13b-orca-8k-3319](https://huggingface.co/OpenAssistant/llama2-13
|
|
7 |
I had to apply this [workaround](https://huggingface.co/OpenAssistant/oasst-sft-6-llama-30b-xor/discussions/2) to pad the vocab and quantize the models, this may or may not affect performance.<br>
|
8 |
I have no idea what I'm doing so if something doesn't work as it should or at all that's likely on me, not the models themselves.
|
9 |
|
10 |
-
Below is the suggested prompt format from the original repo
|
11 |
|
12 |
For the initial response use (e.g. the [llama2 default system prompt](https://github.com/facebookresearch/llama/blob/6c7fe276574e78057f917549435a2554000a876d/llama/generation.py#L46) works well):
|
13 |
|
14 |
```
|
15 |
<|system|>system message</s><|prompter|>user prompt</s><|assistant|>
|
16 |
```
|
17 |
-
|
18 |
For multi-turn conversations use:
|
19 |
-
|
20 |
```
|
21 |
<|system|>system message</s><|prompter|>Q1</s><|assistant|>A1</s><|prompter|>Q2</s><|assistant|>
|
22 |
```
|
|
|
7 |
I had to apply this [workaround](https://huggingface.co/OpenAssistant/oasst-sft-6-llama-30b-xor/discussions/2) to pad the vocab and quantize the models, this may or may not affect performance.<br>
|
8 |
I have no idea what I'm doing so if something doesn't work as it should or at all that's likely on me, not the models themselves.
|
9 |
|
10 |
+
Below is the suggested prompt format from the original repo:
|
11 |
|
12 |
For the initial response use (e.g. the [llama2 default system prompt](https://github.com/facebookresearch/llama/blob/6c7fe276574e78057f917549435a2554000a876d/llama/generation.py#L46) works well):
|
13 |
|
14 |
```
|
15 |
<|system|>system message</s><|prompter|>user prompt</s><|assistant|>
|
16 |
```
|
|
|
17 |
For multi-turn conversations use:
|
|
|
18 |
```
|
19 |
<|system|>system message</s><|prompter|>Q1</s><|assistant|>A1</s><|prompter|>Q2</s><|assistant|>
|
20 |
```
|