Update README.md
Browse filesAdded licence, clarification of testing experience
README.md
CHANGED
@@ -6,7 +6,7 @@ library_name: transformers
|
|
6 |
tags:
|
7 |
- mergekit
|
8 |
- merge
|
9 |
-
|
10 |
---
|
11 |
# franken-kunoichi-IDUS-11B
|
12 |
|
@@ -14,7 +14,7 @@ This is a merge of pre-trained language models created using [mergekit](https://
|
|
14 |
|
15 |
The Interwoven Depth Up-Scaling merge formula was adapted from [Sanji Watsuki's longcat-10.7B](https://huggingface.co/SanjiWatsuki/longcat-10.7B).
|
16 |
|
17 |
-
I consider this to be a negative result, but perhaps an interesting one. I've tested casually with temperature 0.7-1.2 and minP 0.01-0.03, with both Alpaca and ChatML prompts. The resulting generated text is interesting for RP, mostly sticking to grammatically correct output, but
|
18 |
|
19 |
## Merge Details
|
20 |
### Merge Method
|
@@ -138,4 +138,4 @@ slices:
|
|
138 |
merge_method: passthrough
|
139 |
dtype: float16
|
140 |
|
141 |
-
```
|
|
|
6 |
tags:
|
7 |
- mergekit
|
8 |
- merge
|
9 |
+
license: cc-by-nc-4.0
|
10 |
---
|
11 |
# franken-kunoichi-IDUS-11B
|
12 |
|
|
|
14 |
|
15 |
The Interwoven Depth Up-Scaling merge formula was adapted from [Sanji Watsuki's longcat-10.7B](https://huggingface.co/SanjiWatsuki/longcat-10.7B).
|
16 |
|
17 |
+
I consider this to be a negative result, but perhaps an interesting one. I've tested casually with temperature 0.7-1.2 and minP 0.01-0.03, with both Alpaca and ChatML prompts. The resulting generated text is interesting for RP due to chaotic variation, mostly sticking to grammatically correct output, but easily veers too far into chaos (e.g., abruptly switching language) while having difficulty tracking things. Given that, the inherited 8K context length is of dubious benefit.
|
18 |
|
19 |
## Merge Details
|
20 |
### Merge Method
|
|
|
138 |
merge_method: passthrough
|
139 |
dtype: float16
|
140 |
|
141 |
+
```
|