Update README.md
Browse files
README.md
CHANGED
@@ -1,37 +1,77 @@
|
|
1 |
---
|
2 |
base_model:
|
|
|
|
|
3 |
- v000000/L3.1-Niitorm-8B-t0.0001
|
4 |
-
- v000000/NII_8B_DPO
|
5 |
library_name: transformers
|
6 |
tags:
|
7 |
-
- mergekit
|
8 |
- merge
|
9 |
-
|
|
|
|
|
|
|
10 |
---
|
11 |
-
# merge
|
12 |
|
13 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
14 |
|
15 |
## Merge Details
|
16 |
### Merge Method
|
17 |
|
18 |
-
This model was merged using the
|
19 |
|
20 |
### Models Merged
|
21 |
|
22 |
The following models were included in the merge:
|
23 |
-
* [
|
|
|
24 |
|
25 |
### Configuration
|
26 |
|
27 |
The following YAML configuration was used to produce this model:
|
28 |
|
29 |
```yaml
|
30 |
-
|
31 |
-
-
|
32 |
-
|
33 |
-
|
34 |
-
|
35 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
36 |
|
37 |
```
|
|
|
|
|
|
|
|
1 |
---
|
2 |
base_model:
|
3 |
+
- akjindal53244/Llama-3.1-Storm-8B
|
4 |
+
- Sao10K/L3.1-8B-Niitama-v1.1
|
5 |
- v000000/L3.1-Niitorm-8B-t0.0001
|
|
|
6 |
library_name: transformers
|
7 |
tags:
|
|
|
8 |
- merge
|
9 |
+
- llama
|
10 |
+
- dpo
|
11 |
+
datasets:
|
12 |
+
- jondurbin/gutenberg-dpo-v0.1
|
13 |
---
|
|
|
14 |
|
15 |
+
# Llama-3.1-Niitorm-8B-DPO
|
16 |
+
|
17 |
+
* *DPO Trained, Llama3.1-8B.*
|
18 |
+
|
19 |
+
![image/png](https://cdn-uploads.huggingface.co/production/uploads/64f74b6e6389380c77562762/J7sCzFV3_wHtXaU8jpep5.png)
|
20 |
+
|
21 |
+
DPO Gutenberg Finetune version. RP model, Niitama 1.1 as a base, nearswapped with one of the smartest 3.1 models "Storm", mostly abliterated.
|
22 |
+
|
23 |
+
-------------------------------------------------------------------------------
|
24 |
+
|
25 |
+
## finetune and merge
|
26 |
+
|
27 |
+
This is a merge and finetune of pre-trained language models.
|
28 |
+
|
29 |
+
*Resultant merge finetuned* on [jondurbin/gutenberg-dpo-v0.1](https://huggingface.co/datasets/jondurbin/gutenberg-dpo-v0.1) for 1 epoch 1.5e-5 learning rate.
|
30 |
|
31 |
## Merge Details
|
32 |
### Merge Method
|
33 |
|
34 |
+
This model was merged using the <b>NEARSWAP t0.0001</b> merge algorithm.
|
35 |
|
36 |
### Models Merged
|
37 |
|
38 |
The following models were included in the merge:
|
39 |
+
* Base Model: [Sao10K/L3.1-8B-Niitama-v1.1](https://huggingface.co/Sao10K/L3.1-8B-Niitama-v1.1) + [grimjim/Llama-3-Instruct-abliteration-LoRA-8B](https://huggingface.co/grimjim/Llama-3-Instruct-abliteration-LoRA-8B)
|
40 |
+
* [akjindal53244/Llama-3.1-Storm-8B](https://huggingface.co/akjindal53244/Llama-3.1-Storm-8B)
|
41 |
|
42 |
### Configuration
|
43 |
|
44 |
The following YAML configuration was used to produce this model:
|
45 |
|
46 |
```yaml
|
47 |
+
slices:
|
48 |
+
- sources:
|
49 |
+
- model: Sao10K/L3.1-8B-Niitama-v1.1+grimjim/Llama-3-Instruct-abliteration-LoRA-8B
|
50 |
+
layer_range: [0, 32]
|
51 |
+
- model: akjindal53244/Llama-3.1-Storm-8B
|
52 |
+
layer_range: [0, 32]
|
53 |
+
merge_method: nearswap
|
54 |
+
base_model: Sao10K/L3.1-8B-Niitama-v1.1+grimjim/Llama-3-Instruct-abliteration-LoRA-8B
|
55 |
+
parameters:
|
56 |
+
t:
|
57 |
+
- value: 0.0001
|
58 |
+
dtype: bfloat16
|
59 |
+
out_type: float16 #oops
|
60 |
+
```
|
61 |
+
|
62 |
+
*Resultant merge finetuned* on [jondurbin/gutenberg-dpo-v0.1](https://huggingface.co/datasets/jondurbin/gutenberg-dpo-v0.1) for 1 epoch 1.5e-5 learning rate.
|
63 |
+
|
64 |
+
# Prompt Template:
|
65 |
+
```bash
|
66 |
+
<|begin_of_text|><|start_header_id|>system<|end_header_id|>
|
67 |
+
|
68 |
+
{system_prompt}<|eot_id|><|start_header_id|>user<|end_header_id|>
|
69 |
+
|
70 |
+
{input}<|eot_id|><|start_header_id|>assistant<|end_header_id|>
|
71 |
+
|
72 |
+
{output}<|eot_id|>
|
73 |
|
74 |
```
|
75 |
+
Credit to Alchemonaut.
|
76 |
+
|
77 |
+
Credit to woofwolfy for the idea.
|