IHaveNoClueAndIMustPost commited on
Commit
c00e3ae
1 Parent(s): 9849f66

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -3
README.md CHANGED
@@ -7,16 +7,14 @@ This is [llama2-13b-orca-8k-3319](https://huggingface.co/OpenAssistant/llama2-13
7
  I had to apply this [workaround](https://huggingface.co/OpenAssistant/oasst-sft-6-llama-30b-xor/discussions/2) to pad the vocab and quantize the models, this may or may not affect performance.<br>
8
  I have no idea what I'm doing so if something doesn't work as it should or at all that's likely on me, not the models themselves.
9
 
10
- Below is the suggested prompt format from the original repo.
11
 
12
  For the initial response use (e.g. the [llama2 default system prompt](https://github.com/facebookresearch/llama/blob/6c7fe276574e78057f917549435a2554000a876d/llama/generation.py#L46) works well):
13
 
14
  ```
15
  <|system|>system message</s><|prompter|>user prompt</s><|assistant|>
16
  ```
17
-
18
  For multi-turn conversations use:
19
-
20
  ```
21
  <|system|>system message</s><|prompter|>Q1</s><|assistant|>A1</s><|prompter|>Q2</s><|assistant|>
22
  ```
 
7
  I had to apply this [workaround](https://huggingface.co/OpenAssistant/oasst-sft-6-llama-30b-xor/discussions/2) to pad the vocab and quantize the models, this may or may not affect performance.<br>
8
  I have no idea what I'm doing so if something doesn't work as it should or at all that's likely on me, not the models themselves.
9
 
10
+ Below is the suggested prompt format from the original repo:
11
 
12
  For the initial response use (e.g. the [llama2 default system prompt](https://github.com/facebookresearch/llama/blob/6c7fe276574e78057f917549435a2554000a876d/llama/generation.py#L46) works well):
13
 
14
  ```
15
  <|system|>system message</s><|prompter|>user prompt</s><|assistant|>
16
  ```
 
17
  For multi-turn conversations use:
 
18
  ```
19
  <|system|>system message</s><|prompter|>Q1</s><|assistant|>A1</s><|prompter|>Q2</s><|assistant|>
20
  ```