ryzen88 commited on
Commit
a25e394
1 Parent(s): a1b8b75

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +46 -46
README.md CHANGED
@@ -1,46 +1,46 @@
1
- ---
2
- base_model: []
3
- library_name: transformers
4
- tags:
5
- - mergekit
6
- - merge
7
-
8
- ---
9
- # Merge_XL_model_Stock
10
- This model switches to the Smaug instruct 32K for the base bodel.
11
- Expanded with Giraffe and Gradient to keep a robuust long context window.
12
- Higgs and cat for most of the story and RP aspects.
13
- Hermes and Chinese chat are for overall intelligence and understanding.
14
-
15
- ## Merge Details
16
- ### Merge Method
17
-
18
- This model was merged using the [Model Stock](https://arxiv.org/abs/2403.19522) merge method using \Smaug-Llama-3-70B-Instruct-32K as a base.
19
-
20
- ### Models Merged
21
-
22
- The following models were included in the merge:
23
- * \Llama-3-Giraffe-70B-Instruct
24
- * \Llama-3-70B-Instruct-Gradient-262k
25
- * \Hermes-2-Theta-Llama-3-70B
26
- * \Higgs-Llama-3-70B
27
- * \Llama3-70B-Chinese-Chat
28
- * \Meta-LLama-3-Cat-A-LLama-70b
29
-
30
- ### Configuration
31
-
32
- The following YAML configuration was used to produce this model:
33
-
34
- ```yaml
35
- models:
36
- - model: \Smaug-Llama-3-70B-Instruct-32K
37
- - model: \Llama-3-70B-Instruct-Gradient-262k
38
- - model: \Llama-3-Giraffe-70B-Instruct
39
- - model: \Higgs-Llama-3-70B
40
- - model: \Llama3-70B-Chinese-Chat
41
- - model: \Meta-LLama-3-Cat-A-LLama-70b
42
- - model: \Hermes-2-Theta-Llama-3-70B
43
- merge_method: model_stock
44
- base_model: \Smaug-Llama-3-70B-Instruct-32K
45
- dtype: bfloat16
46
- ```
 
1
+ ---
2
+ base_model: []
3
+ library_name: transformers
4
+ tags:
5
+ - mergekit
6
+ - merge
7
+
8
+ ---
9
+ # Merge_XL_model_Stock
10
+ This model switches to the Smaug instruct 32K for the base bodel.
11
+ Expanded with Giraffe and Gradient to keep a robuust long context window.
12
+ Higgs and cat for most of the story and RP aspects.
13
+ Hermes and Chinese chat are for overall intelligence and understanding.
14
+
15
+ ## Merge Details
16
+ ### Merge Method
17
+
18
+ This model was merged using the [Model Stock](https://arxiv.org/abs/2403.19522) merge method using \Smaug-Llama-3-70B-Instruct-32K as a base.
19
+
20
+ ### Models Merged
21
+
22
+ The following models were included in the merge:
23
+ * \Llama-3-Giraffe-70B-Instruct
24
+ * \Llama-3-70B-Instruct-Gradient-262k
25
+ * \Hermes-2-Theta-Llama-3-70B
26
+ * \Higgs-Llama-3-70B
27
+ * \Llama3-70B-Chinese-Chat
28
+ * \Meta-LLama-3-Cat-A-LLama-70b
29
+
30
+ ### Configuration
31
+
32
+ The following YAML configuration was used to produce this model:
33
+
34
+ ```yaml
35
+ models:
36
+ - model: \Smaug-Llama-3-70B-Instruct-32K
37
+ - model: \Llama-3-70B-Instruct-Gradient-262k
38
+ - model: \Llama-3-Giraffe-70B-Instruct
39
+ - model: \Higgs-Llama-3-70B
40
+ - model: \Llama3-70B-Chinese-Chat
41
+ - model: \Meta-LLama-3-Cat-A-LLama-70b
42
+ - model: \Hermes-2-Theta-Llama-3-70B
43
+ merge_method: model_stock
44
+ base_model: \Smaug-Llama-3-70B-Instruct-32K
45
+ dtype: bfloat16
46
+ ```