Transformers
GGUF
English
mistral
TheBloke commited on
Commit
96d9f24
1 Parent(s): 5fc2fe1

Upload README.md

Browse files
Files changed (1) hide show
  1. README.md +9 -1
README.md CHANGED
@@ -3,7 +3,10 @@ base_model: NousResearch/Yarn-Mistral-7b-128k
3
  datasets:
4
  - emozilla/yarn-train-tokenized-16k-mistral
5
  inference: false
 
 
6
  library_name: transformers
 
7
  metrics:
8
  - perplexity
9
  model_creator: NousResearch
@@ -50,7 +53,7 @@ These files were quantised using hardware kindly provided by [Massed Compute](ht
50
 
51
  GGUF is a new format introduced by the llama.cpp team on August 21st 2023. It is a replacement for GGML, which is no longer supported by llama.cpp.
52
 
53
- Here is an incomplate list of clients and libraries that are known to support GGUF:
54
 
55
  * [llama.cpp](https://github.com/ggerganov/llama.cpp). The source project for GGUF. Offers a CLI and a server option.
56
  * [text-generation-webui](https://github.com/oobabooga/text-generation-webui), the most widely used web UI, with many features and powerful extensions. Supports GPU acceleration.
@@ -310,6 +313,11 @@ model = AutoModelForCausalLM.from_pretrained("NousResearch/Yarn-Mistral-7b-128k"
310
  trust_remote_code=True)
311
  ```
312
 
 
 
 
 
 
313
  ## Benchmarks
314
 
315
  Long context benchmarks:
 
3
  datasets:
4
  - emozilla/yarn-train-tokenized-16k-mistral
5
  inference: false
6
+ language:
7
+ - en
8
  library_name: transformers
9
+ license: apache-2.0
10
  metrics:
11
  - perplexity
12
  model_creator: NousResearch
 
53
 
54
  GGUF is a new format introduced by the llama.cpp team on August 21st 2023. It is a replacement for GGML, which is no longer supported by llama.cpp.
55
 
56
+ Here is an incomplete list of clients and libraries that are known to support GGUF:
57
 
58
  * [llama.cpp](https://github.com/ggerganov/llama.cpp). The source project for GGUF. Offers a CLI and a server option.
59
  * [text-generation-webui](https://github.com/oobabooga/text-generation-webui), the most widely used web UI, with many features and powerful extensions. Supports GPU acceleration.
 
313
  trust_remote_code=True)
314
  ```
315
 
316
+ In addition you will need to use the latest version of `transformers` (until 4.35 comes out)
317
+ ```sh
318
+ pip install git+https://github.com/huggingface/transformers
319
+ ```
320
+
321
  ## Benchmarks
322
 
323
  Long context benchmarks: