--- license: cc-by-nc-4.0 base_model: spow12/Ko-Qwen2-7B-Instruct tags: - gguf model-index: - name: joongi007/Ko-Qwen2-7B-Instruct-GGUF results: [] --- - Original model is [spow12/Ko-Qwen2-7B-Instruct](https://huggingface.co/spow12/Ko-Qwen2-7B-Instruct) - quantized using [llama.cpp](https://github.com/ggerganov/llama.cpp) ```prompt <|im_start|>system {System}<|im_end|> <|im_start|>user {User}<|im_end|> <|im_start|>assistant {Assistant} ``` "Flash Attention" function must be activated. [why?](https://www.reddit.com/r/LocalLLaMA/comments/1da19nu/if_your_qwen2_gguf_is_spitting_nonsense_enable/)