Upload README.md
Browse files
README.md
CHANGED
@@ -3,7 +3,10 @@ base_model: NousResearch/Yarn-Mistral-7b-128k
|
|
3 |
datasets:
|
4 |
- emozilla/yarn-train-tokenized-16k-mistral
|
5 |
inference: false
|
|
|
|
|
6 |
library_name: transformers
|
|
|
7 |
metrics:
|
8 |
- perplexity
|
9 |
model_creator: NousResearch
|
@@ -50,7 +53,7 @@ These files were quantised using hardware kindly provided by [Massed Compute](ht
|
|
50 |
|
51 |
GGUF is a new format introduced by the llama.cpp team on August 21st 2023. It is a replacement for GGML, which is no longer supported by llama.cpp.
|
52 |
|
53 |
-
Here is an
|
54 |
|
55 |
* [llama.cpp](https://github.com/ggerganov/llama.cpp). The source project for GGUF. Offers a CLI and a server option.
|
56 |
* [text-generation-webui](https://github.com/oobabooga/text-generation-webui), the most widely used web UI, with many features and powerful extensions. Supports GPU acceleration.
|
@@ -310,6 +313,11 @@ model = AutoModelForCausalLM.from_pretrained("NousResearch/Yarn-Mistral-7b-128k"
|
|
310 |
trust_remote_code=True)
|
311 |
```
|
312 |
|
|
|
|
|
|
|
|
|
|
|
313 |
## Benchmarks
|
314 |
|
315 |
Long context benchmarks:
|
|
|
3 |
datasets:
|
4 |
- emozilla/yarn-train-tokenized-16k-mistral
|
5 |
inference: false
|
6 |
+
language:
|
7 |
+
- en
|
8 |
library_name: transformers
|
9 |
+
license: apache-2.0
|
10 |
metrics:
|
11 |
- perplexity
|
12 |
model_creator: NousResearch
|
|
|
53 |
|
54 |
GGUF is a new format introduced by the llama.cpp team on August 21st 2023. It is a replacement for GGML, which is no longer supported by llama.cpp.
|
55 |
|
56 |
+
Here is an incomplete list of clients and libraries that are known to support GGUF:
|
57 |
|
58 |
* [llama.cpp](https://github.com/ggerganov/llama.cpp). The source project for GGUF. Offers a CLI and a server option.
|
59 |
* [text-generation-webui](https://github.com/oobabooga/text-generation-webui), the most widely used web UI, with many features and powerful extensions. Supports GPU acceleration.
|
|
|
313 |
trust_remote_code=True)
|
314 |
```
|
315 |
|
316 |
+
In addition you will need to use the latest version of `transformers` (until 4.35 comes out)
|
317 |
+
```sh
|
318 |
+
pip install git+https://github.com/huggingface/transformers
|
319 |
+
```
|
320 |
+
|
321 |
## Benchmarks
|
322 |
|
323 |
Long context benchmarks:
|