Lewdiculous commited on
Commit
1cd4d2a
1 Parent(s): 726dea9

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +39 -2
README.md CHANGED
@@ -12,7 +12,7 @@ tags:
12
  pipeline_tag: text-generation
13
  inference: false
14
  ---
15
- # **GGUF-Imatrix quantizations for [Pasta-Lake-7b](https://huggingface.co/Test157t/Pasta-Lake-7b).**
16
 
17
  ## *This has been my personal favourite and daily-driver role-play model for a while, so I decided to make new quantizations for it using the full F16-Imatrix data.*
18
 
@@ -30,4 +30,41 @@ Using [llama.cpp](https://github.com/ggerganov/llama.cpp/)-[b2254](https://githu
30
 
31
  For --imatrix data, `imatrix-Kunocchini-7b-128k-test-F16.dat` was used.
32
 
33
- # Original model information:
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
12
  pipeline_tag: text-generation
13
  inference: false
14
  ---
15
+ # **GGUF-Imatrix quantizations for [Kunocchini-7b-128k-test](https://huggingface.co/Test157t/Kunocchini-7b-128k-test/).**
16
 
17
  ## *This has been my personal favourite and daily-driver role-play model for a while, so I decided to make new quantizations for it using the full F16-Imatrix data.*
18
 
 
30
 
31
  For --imatrix data, `imatrix-Kunocchini-7b-128k-test-F16.dat` was used.
32
 
33
+ # Original model information:
34
+
35
+ Thanks to @Epiculous for the dope model/ help with llm backends and support overall.
36
+
37
+ Id like to also thank @kalomaze for the dope sampler additions to ST.
38
+
39
+ @SanjiWatsuki Thank you very much for the help, and the model!
40
+
41
+ ST users can find the TextGenPreset in the folder labeled so.
42
+
43
+ ![image/jpeg](https://cdn-uploads.huggingface.co/production/uploads/642265bc01c62c1e4102dc36/9obNSalcJqCilQwr_4ssM.jpeg)
44
+
45
+ The following models were included in the merge:
46
+ * [SanjiWatsuki/Kunoichi-DPO-v2-7B](https://huggingface.co/SanjiWatsuki/Kunoichi-DPO-v2-7B)
47
+ * [Epiculous/Fett-uccine-Long-Noodle-7B-120k-Context](https://huggingface.co/Epiculous/Fett-uccine-Long-Noodle-7B-120k-Context)
48
+
49
+ ### Configuration
50
+
51
+ The following YAML configuration was used to produce this model:
52
+
53
+ ```yaml
54
+ slices:
55
+ - sources:
56
+ - model: SanjiWatsuki/Kunoichi-DPO-v2-7B
57
+ layer_range: [0, 32]
58
+ - model: Epiculous/Fett-uccine-Long-Noodle-7B-120k-Context
59
+ layer_range: [0, 32]
60
+ merge_method: slerp
61
+ base_model: SanjiWatsuki/Kunoichi-DPO-v2-7B
62
+ parameters:
63
+ t:
64
+ - filter: self_attn
65
+ value: [0, 0.5, 0.3, 0.7, 1]
66
+ - filter: mlp
67
+ value: [1, 0.5, 0.7, 0.3, 0]
68
+ - value: 0.5
69
+ dtype: bfloat16
70
+ ```