Lewdiculous
commited on
Commit
•
c4e4931
1
Parent(s):
c590194
Update README.md
Browse files
README.md
CHANGED
@@ -47,4 +47,41 @@ If you want anything that's not here or another model, feel free to request.
|
|
47 |
|
48 |
![image/jpeg](https://cdn-uploads.huggingface.co/production/uploads/65d4cf2693a0a3744a27536c/sO69MKgyqTF9s0xY2lg-0.jpeg)
|
49 |
|
50 |
-
**Original model information:**
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
47 |
|
48 |
![image/jpeg](https://cdn-uploads.huggingface.co/production/uploads/65d4cf2693a0a3744a27536c/sO69MKgyqTF9s0xY2lg-0.jpeg)
|
49 |
|
50 |
+
**Original model information:**
|
51 |
+
|
52 |
+
# kuno-kunoichi-v1-DPO-v2-SLERP-7B
|
53 |
+
|
54 |
+
kuno-kunoichi-v1-DPO-v2-SLERP-7B is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
|
55 |
+
I'm hoping that the result is more robust against errors or when merging due to "denseness", as the two models likely implement comparable reasoning at least somewhat differently.
|
56 |
+
|
57 |
+
I've performed some testing with ChatML format prompting using temperature=1.1 and minP=0.03. The model also supports Alpaca format promtps.
|
58 |
+
## Merge Details
|
59 |
+
### Merge Method
|
60 |
+
|
61 |
+
This model was merged using the SLERP merge method.
|
62 |
+
|
63 |
+
### Models Merged
|
64 |
+
|
65 |
+
The following models were included in the merge:
|
66 |
+
* [SanjiWatsuki/Kunoichi-7B](https://huggingface.co/SanjiWatsuki/Kunoichi-7B)
|
67 |
+
* [SanjiWatsuki/Kunoichi-DPO-v2-7B](https://huggingface.co/SanjiWatsuki/Kunoichi-DPO-v2-7B)
|
68 |
+
|
69 |
+
### Configuration
|
70 |
+
|
71 |
+
The following YAML configuration was used to produce this model:
|
72 |
+
|
73 |
+
```yaml
|
74 |
+
slices:
|
75 |
+
- sources:
|
76 |
+
- model: SanjiWatsuki/Kunoichi-7B
|
77 |
+
layer_range: [0,32]
|
78 |
+
- model: SanjiWatsuki/Kunoichi-DPO-v2-7B
|
79 |
+
layer_range: [0,32]
|
80 |
+
merge_method: slerp
|
81 |
+
base_model: SanjiWatsuki/Kunoichi-7B
|
82 |
+
parameters:
|
83 |
+
t:
|
84 |
+
- value: 0.5
|
85 |
+
dtype: float16
|
86 |
+
|
87 |
+
```
|