v000000 commited on
Commit
b4d596a
1 Parent(s): 21cf3e2

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +53 -13
README.md CHANGED
@@ -1,37 +1,77 @@
1
  ---
2
  base_model:
 
 
3
  - v000000/L3.1-Niitorm-8B-t0.0001
4
- - v000000/NII_8B_DPO
5
  library_name: transformers
6
  tags:
7
- - mergekit
8
  - merge
9
-
 
 
 
10
  ---
11
- # merge
12
 
13
- This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
 
 
 
 
 
 
 
 
 
 
 
 
 
 
14
 
15
  ## Merge Details
16
  ### Merge Method
17
 
18
- This model was merged using the [linear](https://arxiv.org/abs/2203.05482) merge method.
19
 
20
  ### Models Merged
21
 
22
  The following models were included in the merge:
23
- * [v000000/L3.1-Niitorm-8B-t0.0001](https://huggingface.co/v000000/L3.1-Niitorm-8B-t0.0001) + [v000000/NII_8B_DPO](https://huggingface.co/v000000/NII_8B_DPO)
 
24
 
25
  ### Configuration
26
 
27
  The following YAML configuration was used to produce this model:
28
 
29
  ```yaml
30
- models:
31
- - model: v000000/L3.1-Niitorm-8B-t0.0001+v000000/NII_8B_DPO
32
- parameters:
33
- weight: 1.0
34
- merge_method: linear
35
- dtype: float16
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
36
 
37
  ```
 
 
 
 
1
  ---
2
  base_model:
3
+ - akjindal53244/Llama-3.1-Storm-8B
4
+ - Sao10K/L3.1-8B-Niitama-v1.1
5
  - v000000/L3.1-Niitorm-8B-t0.0001
 
6
  library_name: transformers
7
  tags:
 
8
  - merge
9
+ - llama
10
+ - dpo
11
+ datasets:
12
+ - jondurbin/gutenberg-dpo-v0.1
13
  ---
 
14
 
15
+ # Llama-3.1-Niitorm-8B-DPO
16
+
17
+ * *DPO Trained, Llama3.1-8B.*
18
+
19
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/64f74b6e6389380c77562762/J7sCzFV3_wHtXaU8jpep5.png)
20
+
21
+ DPO Gutenberg Finetune version. RP model, Niitama 1.1 as a base, nearswapped with one of the smartest 3.1 models "Storm", mostly abliterated.
22
+
23
+ -------------------------------------------------------------------------------
24
+
25
+ ## finetune and merge
26
+
27
+ This is a merge and finetune of pre-trained language models.
28
+
29
+ *Resultant merge finetuned* on [jondurbin/gutenberg-dpo-v0.1](https://huggingface.co/datasets/jondurbin/gutenberg-dpo-v0.1) for 1 epoch 1.5e-5 learning rate.
30
 
31
  ## Merge Details
32
  ### Merge Method
33
 
34
+ This model was merged using the <b>NEARSWAP t0.0001</b> merge algorithm.
35
 
36
  ### Models Merged
37
 
38
  The following models were included in the merge:
39
+ * Base Model: [Sao10K/L3.1-8B-Niitama-v1.1](https://huggingface.co/Sao10K/L3.1-8B-Niitama-v1.1) + [grimjim/Llama-3-Instruct-abliteration-LoRA-8B](https://huggingface.co/grimjim/Llama-3-Instruct-abliteration-LoRA-8B)
40
+ * [akjindal53244/Llama-3.1-Storm-8B](https://huggingface.co/akjindal53244/Llama-3.1-Storm-8B)
41
 
42
  ### Configuration
43
 
44
  The following YAML configuration was used to produce this model:
45
 
46
  ```yaml
47
+ slices:
48
+ - sources:
49
+ - model: Sao10K/L3.1-8B-Niitama-v1.1+grimjim/Llama-3-Instruct-abliteration-LoRA-8B
50
+ layer_range: [0, 32]
51
+ - model: akjindal53244/Llama-3.1-Storm-8B
52
+ layer_range: [0, 32]
53
+ merge_method: nearswap
54
+ base_model: Sao10K/L3.1-8B-Niitama-v1.1+grimjim/Llama-3-Instruct-abliteration-LoRA-8B
55
+ parameters:
56
+ t:
57
+ - value: 0.0001
58
+ dtype: bfloat16
59
+ out_type: float16 #oops
60
+ ```
61
+
62
+ *Resultant merge finetuned* on [jondurbin/gutenberg-dpo-v0.1](https://huggingface.co/datasets/jondurbin/gutenberg-dpo-v0.1) for 1 epoch 1.5e-5 learning rate.
63
+
64
+ # Prompt Template:
65
+ ```bash
66
+ <|begin_of_text|><|start_header_id|>system<|end_header_id|>
67
+
68
+ {system_prompt}<|eot_id|><|start_header_id|>user<|end_header_id|>
69
+
70
+ {input}<|eot_id|><|start_header_id|>assistant<|end_header_id|>
71
+
72
+ {output}<|eot_id|>
73
 
74
  ```
75
+ Credit to Alchemonaut.
76
+
77
+ Credit to woofwolfy for the idea.