Azazelle commited on
Commit
61212c3
1 Parent(s): f699e49

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +56 -0
README.md ADDED
@@ -0,0 +1,56 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model:
3
+ - Open-Orca/Mistral-7B-OpenOrca
4
+ - akjindal53244/Mistral-7B-v0.1-Open-Platypus
5
+ - WizardLM/WizardMath-7B-V1.1
6
+ - mistralai/Mistral-7B-v0.1
7
+ library_name: transformers
8
+ tags:
9
+ - mergekit
10
+ - merge
11
+
12
+ ---
13
+ # models
14
+
15
+ This is a merge of pre-trained language models created using a fork of [mergekit](https://github.com/cg123/mergekit).
16
+
17
+ ## Merge Details
18
+ ### Merge Method
19
+
20
+ This model was merged using an experimental merge method using [mistralai/Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) as a base.
21
+
22
+ ### Models Merged
23
+
24
+ The following models were included in the merge:
25
+ * [Open-Orca/Mistral-7B-OpenOrca](https://huggingface.co/Open-Orca/Mistral-7B-OpenOrca)
26
+ * [akjindal53244/Mistral-7B-v0.1-Open-Platypus](https://huggingface.co/akjindal53244/Mistral-7B-v0.1-Open-Platypus)
27
+ * [WizardLM/WizardMath-7B-V1.1](https://huggingface.co/WizardLM/WizardMath-7B-V1.1)
28
+
29
+ ### Configuration
30
+
31
+ The following YAML configuration was used to produce this model:
32
+
33
+ ```yaml
34
+ models:
35
+ - model: Open-Orca/Mistral-7B-OpenOrca
36
+ parameters:
37
+ density: [1, 0.7, 0.1] # density gradient
38
+ weight: 1.0
39
+ - model: akjindal53244/Mistral-7B-v0.1-Open-Platypus
40
+ parameters:
41
+ density: 0.5
42
+ weight: [0, 0.3, 0.7, 1] # weight gradient
43
+ - model: WizardLM/WizardMath-7B-V1.1
44
+ parameters:
45
+ density: 0.33
46
+ weight:
47
+ - filter: mlp
48
+ value: 0.5
49
+ - value: 0
50
+ merge_method: sample_ties
51
+ base_model: mistralai/Mistral-7B-v0.1
52
+ parameters:
53
+ normalize: true
54
+ int8_mask: true
55
+ dtype: float16
56
+ ```