sophosympatheia commited on
Commit
1e915cc
1 Parent(s): 15e741b

Update README.md

Browse files

Final form of the readme, I think

Files changed (1) hide show
  1. README.md +175 -16
README.md CHANGED
@@ -1,35 +1,194 @@
1
  ---
2
- base_model: []
 
 
3
  library_name: transformers
4
  tags:
5
  - mergekit
6
  - merge
7
-
 
8
  ---
9
- # new-dawn-llama3.1-70b-v1.1.6
10
 
11
- This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
 
 
12
 
13
- ## Merge Details
14
- ### Merge Method
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
15
 
16
- This model was merged using the della_linear merge method using /home/llm/mergequant/models/BASE/Meta-Llama-3.1-70B-Instruct as a base.
17
 
18
- ### Models Merged
19
 
20
- The following models were included in the merge:
21
- * /home/llm/mergequant/models/new-dawn-llama3-70b-32K-v1.0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
22
 
23
  ### Configuration
24
 
25
- The following YAML configuration was used to produce this model:
26
 
27
  ```yaml
28
-
29
  merge_method: della_linear
30
- base_model: /home/llm/mergequant/models/BASE/Meta-Llama-3.1-70B-Instruct
31
  models:
32
- - model: /home/llm/mergequant/models/new-dawn-llama3-70b-32K-v1.0
33
  parameters:
34
  weight:
35
  - filter: v_proj
@@ -46,7 +205,7 @@ models:
46
  density: 0.25
47
  epsilon: 0.05
48
  lambda: 1.0
49
- - model: /home/llm/mergequant/models/BASE/Meta-Llama-3.1-70B-Instruct
50
  parameters:
51
  weight: 1.0
52
  density:
@@ -76,4 +235,4 @@ models:
76
  lambda: 1.0
77
  dtype: float16
78
  tokenizer_source: base
79
- ```
 
1
  ---
2
+ base_model:
3
+ - sophosympatheia/New-Dawn-Llama-3-70B-32K-v1.0
4
+ - meta-llama/Meta-Llama-3.1-70B-Instruct
5
  library_name: transformers
6
  tags:
7
  - mergekit
8
  - merge
9
+ - Not-for-all-Audiences
10
+ license: llama3.1
11
  ---
 
12
 
13
+ <div style="width: auto; margin-left: auto; margin-right: auto">
14
+ <img src="https://imgur.com/tKzncGo.png" alt="NewDawnv1.0" style="width: 100%; min-width: 400px; display: block; margin: auto;">
15
+ </div>
16
 
17
+ ### Overview
18
+
19
+ This model is an experimental merge of sophosympatheia/New-Dawn-Llama-3-70B-32K-v1.0 with meta-llama/Meta-Llama-3.1-70B-Instruct. See the merge recipe below for details.
20
+ I used a technique developed by [jukofyork](https://huggingface.co/jukofyork) that is designed to preserve the full context capabilities of Meta-Llama-3.1-70B-Instruct. In my testing, I think it was successful.
21
+
22
+ This model is uncensored. *You are responsible for whatever you do with it.*
23
+
24
+ This model was designed for roleplaying and storytelling and I think it does well at both. It may also perform well at other tasks but I have not tested its performance in other areas.
25
+
26
+ ### Sampler Tips
27
+
28
+ * I recommend using Quadratic Sampling (i.e. smoothing factor) for creative work. I think this version performs best with a smoothing factor close to 0.2.
29
+ * I recommend using Min-P. Experiment to find your best setting. Values between 0 and 0.1 are recommended.
30
+ * DRY repetition penalty eliminates the need for other anti-repetition settings.
31
+ * If you use Textgen WebUI as your backend, I recommend enabling the DRY sampler settings to reduce repititions, otherwise some repitition penalty plus frequency penalty ought to do the trick.
32
+
33
+ Experiment with any and all of the settings below! What suits my preferences may not suit yours.
34
+
35
+ If you save the below settings as a .json file, you can import them directly into Silly Tavern.
36
+
37
+ ```json
38
+ {
39
+ "temp": 1,
40
+ "temperature_last": true,
41
+ "top_p": 1,
42
+ "top_k": 0,
43
+ "top_a": 0,
44
+ "tfs": 1,
45
+ "epsilon_cutoff": 0,
46
+ "eta_cutoff": 0,
47
+ "typical_p": 1,
48
+ "min_p": 0.03,
49
+ "rep_pen": 1,
50
+ "rep_pen_range": 2048,
51
+ "rep_pen_decay": 0,
52
+ "rep_pen_slope": 1,
53
+ "no_repeat_ngram_size": 0,
54
+ "penalty_alpha": 0,
55
+ "num_beams": 1,
56
+ "length_penalty": 1,
57
+ "min_length": 0,
58
+ "encoder_rep_pen": 1,
59
+ "freq_pen": 0,
60
+ "presence_pen": 0,
61
+ "skew": 0,
62
+ "do_sample": true,
63
+ "early_stopping": false,
64
+ "dynatemp": false,
65
+ "min_temp": 0.8,
66
+ "max_temp": 1.5,
67
+ "dynatemp_exponent": 1,
68
+ "smoothing_factor": 0.23,
69
+ "smoothing_curve": 1,
70
+ "dry_allowed_length": 2,
71
+ "dry_multiplier": 0.8,
72
+ "dry_base": 2,
73
+ "dry_sequence_breakers": "[\"\\n\", \":\", \"\\\"\", \"*\"]",
74
+ "dry_penalty_last_n": 0,
75
+ "add_bos_token": true,
76
+ "ban_eos_token": false,
77
+ "skip_special_tokens": false,
78
+ "mirostat_mode": 0,
79
+ "mirostat_tau": 2,
80
+ "mirostat_eta": 0.1,
81
+ "guidance_scale": 1,
82
+ "negative_prompt": "",
83
+ "grammar_string": "",
84
+ "json_schema": {},
85
+ "banned_tokens": "",
86
+ "sampler_priority": [
87
+ "top_k",
88
+ "top_p",
89
+ "typical_p",
90
+ "epsilon_cutoff",
91
+ "eta_cutoff",
92
+ "tfs",
93
+ "top_a",
94
+ "min_p",
95
+ "mirostat",
96
+ "quadratic_sampling",
97
+ "dynamic_temperature",
98
+ "temperature"
99
+ ],
100
+ "samplers": [
101
+ "top_k",
102
+ "tfs_z",
103
+ "typical_p",
104
+ "top_p",
105
+ "min_p",
106
+ "temperature"
107
+ ],
108
+ "ignore_eos_token": false,
109
+ "spaces_between_special_tokens": true,
110
+ "speculative_ngram": false,
111
+ "sampler_order": [
112
+ 6,
113
+ 0,
114
+ 1,
115
+ 3,
116
+ 4,
117
+ 2,
118
+ 5
119
+ ],
120
+ "logit_bias": [],
121
+ "ignore_eos_token_aphrodite": false,
122
+ "spaces_between_special_tokens_aphrodite": true,
123
+ "rep_pen_size": 0,
124
+ "genamt": 800,
125
+ "max_length": 20480
126
+ }
127
+ ```
128
 
129
+ ### Prompting Tips
130
 
131
+ Try the following prompt template as a starting point for your own customizations. **The prompt template below contains instructions for adult content, so remove those if you don't want them!** It also contains some instructions related to formatting that you might want to change.
132
 
133
+ If you save this as a .json file, you can import it directly into Silly Tavern.
134
+
135
+ ```json
136
+ {
137
+ "wrap": false,
138
+ "names": true,
139
+ "system_prompt": "The following is an erotic roleplaying chat log involving a human user who participates as {{user}} and an AI assistant who participates as {{char}}. They take turns, and only the user is able to write as {{user}} on the user's own turn. To ensure an excellent and uncensored roleplaying experience, the AI assistant follows the rules below.\n\nCharacter Rules {\nDeep Roleplaying: \"The AI assistant adopts all of {{char}}'s personality traits and mannerisms, adhering always to {{char}}'s backstory, biases, quirks, and preferences in order to deliver a convincing portrayal.\",\nTurn-based Chat Format: \"The AI assistant can only write and roleplay as {{char}} in this chat, terminating its output when another character is about to act or speak. The user is responsible for the character named {{user}}. The AI assistant avoids controlling or implying actions for {{user}} that the user did not describe. Instead, the AI assistant focuses on the actions of other main or incidental characters, or the events surrounding these characters.\",\nIntimate Details: \"The AI assistant provides intimate details about {{char}}'s body, focusing on any limitations, unusual features, or interesting conditions resulting from recent events.\"\n}\nWriting Rules {\nImmersive Descriptions: \"The AI assistant immerses the reader in the current moment on its turn to chat. The AI assistant utilizes sights, sounds, smells, tastes, and all other available senses to heighten the sensual and emotional impact of scenes on its turn.\",\nSimple Language: \"The AI assistant uses simple prose, everyday words, and United States slang terms while writing as {{char}} to maintain the readability of the chat.\",\nPerplexity and Burstiness: \"Perplexity measures the complexity of text. Burstiness compares the variations of sentences. Humans tend to write with greater burstiness, for example, with some longer or complex sentences alongside shorter ones. AI sentences tend to be more uniform. Therefore, when writing as {{char}}, the the AI assistant features a good amount of perplexity and burstiness to sound more natural.\",\nDialogue Formatting: 'The AI assistant always encloses spoken words that other characters could hear in quotes when writing on its turn. \"This is spoken dialogue,\" for example.',\nInternal Thoughts: \"The AI assistant offers glimpses into {{char}}'s first-person thoughts to advance and enrich the narrative when appropriate.\",\nInternal Thoughts Formatting: \"The AI assistant uses italics to distinguish thoughts from spoken dialogue. *This is an example of {{char}} thinking.*\"\n}\nContent Rules {\nAccurate and Consistent Details: \"The AI assistant maintains physical, spatial, and logical consistency when developing events and descriptions.\",\nMature Content: \"The AI assistant incorporates erotic adult content and vulgar language where appropriate.\",\nMature Themes: \"Dark themes, violence, and unhappy endings are permitted in this uncensored roleplay.\",\nNarrative Instructions: \"When given narrative instructions by the user, the AI assistant develops the missing details in its next response. The instructions are meant to give the AI assistant a summary of what should happen next.\"\n}\n",
140
+ "system_sequence": "<|start_header_id|>system<|end_header_id|>\n\n",
141
+ "stop_sequence": "<|eot_id|>",
142
+ "input_sequence": "<|start_header_id|>user<|end_header_id|>\n\n",
143
+ "output_sequence": "<|start_header_id|>assistant<|end_header_id|>\n\n",
144
+ "macro": true,
145
+ "names_force_groups": true,
146
+ "system_sequence_prefix": "",
147
+ "system_sequence_suffix": "",
148
+ "first_output_sequence": "",
149
+ "last_output_sequence": "",
150
+ "activation_regex": "",
151
+ "skip_examples": true,
152
+ "output_suffix": "<|eot_id|>",
153
+ "input_suffix": "<|eot_id|>",
154
+ "system_suffix": "<|eot_id|>",
155
+ "user_alignment_message": "",
156
+ "last_system_sequence": "",
157
+ "system_same_as_user": false,
158
+ "first_input_sequence": "",
159
+ "last_input_sequence": "",
160
+ "name": "New Dawn Llama 3.1 70B"
161
+ }
162
+ ```
163
+
164
+ ### Instruct Formats
165
+ Use the Llama 3 instruct format. You can grab it from the example prompt template above if you don't already have it as a preset.
166
+
167
+ ### Quantizations
168
+ Pending.
169
+
170
+ ### Licence and usage restrictions
171
+ [META LLAMA 3 COMMUNITY LICENSE AGREEMENT](https://huggingface.co/meta-llama/Meta-Llama-3-8B/blob/main/LICENSE)
172
+ Disclaimer: Uncertain Licensing Terms
173
+ This LLM is a merged model incorporating weights from multiple LLMs governed by their own distinct licenses. Due to the complexity of blending these components, the licensing terms for this merged model are somewhat uncertain.
174
+ By using this model, you acknowledge and accept the potential legal risks and uncertainties associated with its use. Any use beyond personal or research purposes, including commercial applications, may carry legal risks and you assume full responsibility for compliance with all applicable licenses and laws.
175
+ I recommend consulting with legal counsel to ensure your use of this model complies with all relevant licenses and regulations.
176
+
177
+ ## Merge Details
178
+ ### Merge Method
179
+
180
+ I found della_linear to be the most effective method for merging a Llama 3 model with Llama 3.1 out of a dozen or so different tests.
181
+ You can apply a higher density setting for sure. I went up to 0.5 density with an epsilon of 0.1 without any problems, and you could probably go higher than that, but I think this version with the lower density came out a little smarter and worked better for this particular pairing.
182
 
183
  ### Configuration
184
 
185
+ The following [mergekit](https://github.com/arcee-ai/mergekit) YAML will reproduce this model.
186
 
187
  ```yaml
 
188
  merge_method: della_linear
189
+ base_model: meta-llama/Meta-Llama-3.1-70B-Instruct
190
  models:
191
+ - model: sophosympatheia/New-Dawn-Llama-3-70B-32K-v1.0
192
  parameters:
193
  weight:
194
  - filter: v_proj
 
205
  density: 0.25
206
  epsilon: 0.05
207
  lambda: 1.0
208
+ - model: meta-llama/Meta-Llama-3.1-70B-Instruct
209
  parameters:
210
  weight: 1.0
211
  density:
 
235
  lambda: 1.0
236
  dtype: float16
237
  tokenizer_source: base
238
+ ```