Update README.md
Browse files
README.md
CHANGED
@@ -9,7 +9,7 @@ model-index:
|
|
9 |
---
|
10 |
|
11 |
- Original model is [spow12/Ko-Qwen2-7B-Instruct](https://huggingface.co/spow12/Ko-Qwen2-7B-Instruct)
|
12 |
-
- quantized using [llama.cpp](https://github.com/ggerganov/llama.cpp)
|
13 |
|
14 |
```prompt
|
15 |
<|im_start|>system
|
@@ -19,4 +19,4 @@ model-index:
|
|
19 |
<|im_start|>assistant
|
20 |
{Assistant}
|
21 |
```
|
22 |
-
"Flash Attention" function must be activated. [why?](https://www.reddit.com/r/LocalLLaMA/comments/1da19nu/if_your_qwen2_gguf_is_spitting_nonsense_enable/)
|
|
|
9 |
---
|
10 |
|
11 |
- Original model is [spow12/Ko-Qwen2-7B-Instruct](https://huggingface.co/spow12/Ko-Qwen2-7B-Instruct)
|
12 |
+
- quantized using [llama.cpp](https://github.com/ggerganov/llama.cpp) - [3510](https://github.com/ggerganov/llama.cpp/releases/tag/b3510)
|
13 |
|
14 |
```prompt
|
15 |
<|im_start|>system
|
|
|
19 |
<|im_start|>assistant
|
20 |
{Assistant}
|
21 |
```
|
22 |
+
~~"Flash Attention" function must be activated. [why?](https://www.reddit.com/r/LocalLLaMA/comments/1da19nu/if_your_qwen2_gguf_is_spitting_nonsense_enable/)~~
|