TheBloke commited on
Commit
d39d17b
1 Parent(s): ddc8190

Initial GPTQ model commit

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -21,7 +21,7 @@ model_type: llama
21
  <!-- header end -->
22
 
23
  # MythoBoros 13B - GPTQ
24
- - Model creator: Gryphe
25
  - Original model: [MythoBoros 13B](https://huggingface.co/Gryphe/MythoBoros-13b)
26
 
27
  ## Description
@@ -35,7 +35,7 @@ Multiple GPTQ parameter permutations are provided; see Provided Files below for
35
 
36
  * [GPTQ models for GPU inference, with multiple quantisation parameter options.](https://huggingface.co/TheBloke/MythoBoros-13B-GPTQ)
37
  * [2, 3, 4, 5, 6 and 8-bit GGML models for CPU+GPU inference](https://huggingface.co/TheBloke/MythoBoros-13B-GGML)
38
- * [Gryphe's original unquantised fp16 model in pytorch format, for GPU inference and for further conversions](https://huggingface.co/Gryphe/MythoBoros-13b)
39
 
40
  ## Prompt template: Alpaca
41
 
 
21
  <!-- header end -->
22
 
23
  # MythoBoros 13B - GPTQ
24
+ - Model creator: [Gryphe](https://huggingface.co/Gryphe)
25
  - Original model: [MythoBoros 13B](https://huggingface.co/Gryphe/MythoBoros-13b)
26
 
27
  ## Description
 
35
 
36
  * [GPTQ models for GPU inference, with multiple quantisation parameter options.](https://huggingface.co/TheBloke/MythoBoros-13B-GPTQ)
37
  * [2, 3, 4, 5, 6 and 8-bit GGML models for CPU+GPU inference](https://huggingface.co/TheBloke/MythoBoros-13B-GGML)
38
+ * [[Gryphe](https://huggingface.co/Gryphe)'s original unquantised fp16 model in pytorch format, for GPU inference and for further conversions](https://huggingface.co/Gryphe/MythoBoros-13b)
39
 
40
  ## Prompt template: Alpaca
41