SamaOkasha commited on
Commit
c8811f4
1 Parent(s): a7100c2

Upload folder using huggingface_hub

Browse files
Files changed (3) hide show
  1. README.json +37 -0
  2. README.md +36 -0
  3. config.json +53 -0
README.json ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ {}
3
+ ---
4
+ {
5
+ "license": "apache-2.0",
6
+ "tags": [
7
+ "merge",
8
+ "mergekit",
9
+ "lazymergekit",
10
+ "allenai/led-base-16384",
11
+ "allenai/led-base-16384",
12
+ ],
13
+ "model_name": "LaMa-Merged-slerp",
14
+ "model": "Your Model Information Here",
15
+ "description": "LaMa-Merged-slerp is a merge of the following models using mergekit.",
16
+ "external_datasets": {},
17
+ "git_repo": "https://github.com/cg123/mergekit",
18
+ "website": "",
19
+ "tasks": [],
20
+ "languages": [],
21
+ "image": "",
22
+ "files": [
23
+ {
24
+ "filename": "config.json",
25
+ "type": "yaml",
26
+ "title": "Configuration"
27
+ }
28
+ ],
29
+ "framework": "",
30
+ "format": "json",
31
+ "references": [],
32
+ "contact": "",
33
+ "creation_date": "2022-03-21",
34
+ "authors": [
35
+ "SamaOkasha"
36
+ ]
37
+ }
README.md ADDED
@@ -0,0 +1,36 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ tags:
4
+ - merge
5
+ - mergekit
6
+ - lazymergekit
7
+ - allenai/led-base-16384
8
+ - allenai/led-base-16384
9
+ ---
10
+
11
+ # LaMa-Merged-slerp
12
+
13
+ LaMa-Merged-slerp is a merge of the following models using [mergekit](https://github.com/cg123/mergekit):
14
+ * [allenai/led-base-16384](https://huggingface.co/allenai/led-base-16384)
15
+ * [allenai/led-base-16384](https://huggingface.co/allenai/led-base-16384)
16
+
17
+ ## 🧩 Configuration
18
+
19
+ ```yaml
20
+ slices:
21
+ - sources:
22
+ - model: allenai/led-base-16384
23
+ layer_range: [0, 12] # Taking the initial layers from LED Base model
24
+ - model: allenai/led-base-16384
25
+ layer_range: [12, 24] # Taking the later layers from LED Base model
26
+ merge_method: slerp
27
+ base_model: allenai/led-base-16384 # Using LED Base model as the base model
28
+ parameters:
29
+ t:
30
+ - filter: self_attn
31
+ value: [0, 0.5, 0.3, 0.7, 1] # Interpolation values for self-attention layers
32
+ - filter: mlp
33
+ value: [1, 0.5, 0.7, 0.3, 0] # Interpolation values for MLP layers
34
+ - value: 0.5 # Default interpolation value
35
+ dtype: bfloat16
36
+ ```
config.json ADDED
@@ -0,0 +1,53 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "slices": [
3
+ {
4
+ "sources": [
5
+ {
6
+ "model": "allenai/led-base-16384",
7
+ "layer_range": [
8
+ 0,
9
+ 12
10
+ ]
11
+ },
12
+ {
13
+ "model": "allenai/led-base-16384",
14
+ "layer_range": [
15
+ 12,
16
+ 24
17
+ ]
18
+ }
19
+ ]
20
+ }
21
+ ],
22
+ "merge_method": "slerp",
23
+ "base_model": "allenai/led-base-16384",
24
+ "parameters": {
25
+ "t": [
26
+ {
27
+ "filter": "self_attn",
28
+ "value": [
29
+ 0,
30
+ 0.5,
31
+ 0.3,
32
+ 0.7,
33
+ 1
34
+ ]
35
+ },
36
+ {
37
+ "filter": "mlp",
38
+ "value": [
39
+ 1,
40
+ 0.5,
41
+ 0.7,
42
+ 0.3,
43
+ 0
44
+ ]
45
+ },
46
+ {
47
+ "value": 0.5
48
+ }
49
+ ]
50
+ },
51
+ "dtype": "bfloat16",
52
+ "model_type": "led"
53
+ }