Lewdiculous commited on
Commit
c4e4931
1 Parent(s): c590194

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +38 -1
README.md CHANGED
@@ -47,4 +47,41 @@ If you want anything that's not here or another model, feel free to request.
47
 
48
  ![image/jpeg](https://cdn-uploads.huggingface.co/production/uploads/65d4cf2693a0a3744a27536c/sO69MKgyqTF9s0xY2lg-0.jpeg)
49
 
50
- **Original model information:**
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
47
 
48
  ![image/jpeg](https://cdn-uploads.huggingface.co/production/uploads/65d4cf2693a0a3744a27536c/sO69MKgyqTF9s0xY2lg-0.jpeg)
49
 
50
+ **Original model information:**
51
+
52
+ # kuno-kunoichi-v1-DPO-v2-SLERP-7B
53
+
54
+ kuno-kunoichi-v1-DPO-v2-SLERP-7B is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
55
+ I'm hoping that the result is more robust against errors or when merging due to "denseness", as the two models likely implement comparable reasoning at least somewhat differently.
56
+
57
+ I've performed some testing with ChatML format prompting using temperature=1.1 and minP=0.03. The model also supports Alpaca format promtps.
58
+ ## Merge Details
59
+ ### Merge Method
60
+
61
+ This model was merged using the SLERP merge method.
62
+
63
+ ### Models Merged
64
+
65
+ The following models were included in the merge:
66
+ * [SanjiWatsuki/Kunoichi-7B](https://huggingface.co/SanjiWatsuki/Kunoichi-7B)
67
+ * [SanjiWatsuki/Kunoichi-DPO-v2-7B](https://huggingface.co/SanjiWatsuki/Kunoichi-DPO-v2-7B)
68
+
69
+ ### Configuration
70
+
71
+ The following YAML configuration was used to produce this model:
72
+
73
+ ```yaml
74
+ slices:
75
+ - sources:
76
+ - model: SanjiWatsuki/Kunoichi-7B
77
+ layer_range: [0,32]
78
+ - model: SanjiWatsuki/Kunoichi-DPO-v2-7B
79
+ layer_range: [0,32]
80
+ merge_method: slerp
81
+ base_model: SanjiWatsuki/Kunoichi-7B
82
+ parameters:
83
+ t:
84
+ - value: 0.5
85
+ dtype: float16
86
+
87
+ ```