Update README.md
Browse files
README.md
CHANGED
@@ -3,19 +3,19 @@ license: other
|
|
3 |
language:
|
4 |
- en
|
5 |
---
|
6 |
-
The Llama 2 sequel to my original experiment with gradient merges using [the following script](https://github.com/Gryphe/BlockMerge_Gradient). Its three models ([Hermes](https://huggingface.co/NousResearch/Nous-Hermes-Llama2-13b), [Chronos](https://huggingface.co/elinas/chronos-13b-v2) and [Airoboros](https://huggingface.co/jondurbin/airoboros-l2-13b-gpt4-2.0)) are almost evenly divided over the layer structure this time. Airoboros was the "wildcard model" due to its superior ability to understand complex instructions.
|
7 |
|
8 |
## Model details
|
9 |
|
10 |
-
As
|
11 |
|
12 |
-
Illustrated below are the gradients used for this specific L2 recipe
|
13 |
|
14 |
-
[](MythoLogic-L2.png)
|
15 |
|
16 |
## Prompt Format
|
17 |
|
18 |
-
This model primarily uses (and
|
19 |
```
|
20 |
### Instruction:
|
21 |
Your instruction or question here.
|
|
|
3 |
language:
|
4 |
- en
|
5 |
---
|
6 |
+
The Llama 2 sequel to my [original experiment](https://huggingface.co/Gryphe/MythoLogic-13b) with gradient merges using [the following script](https://github.com/Gryphe/BlockMerge_Gradient). Its three models ([Hermes](https://huggingface.co/NousResearch/Nous-Hermes-Llama2-13b), [Chronos](https://huggingface.co/elinas/chronos-13b-v2) and [Airoboros](https://huggingface.co/jondurbin/airoboros-l2-13b-gpt4-2.0)) are almost evenly divided over the layer structure this time. Airoboros was the "wildcard model" due to its superior ability to understand complex instructions.
|
7 |
|
8 |
## Model details
|
9 |
|
10 |
+
As before, the main objective was to create an all-round model with improved roleplaying capabilities. MythoLogic-L2 differs from its predecessor in that it focuses primarily on the understanding of instructions and personalities of complex character cards.
|
11 |
|
12 |
+
Illustrated below are the gradients used for this specific L2 recipe, with the top of the image representing layer 0 and the bottom layer 40.
|
13 |
|
14 |
+
![](MythoLogic-L2.png)
|
15 |
|
16 |
## Prompt Format
|
17 |
|
18 |
+
This model primarily uses (and was heavily tested with) Alpaca formatting, so for optimal model performance, use:
|
19 |
```
|
20 |
### Instruction:
|
21 |
Your instruction or question here.
|