SamaOkasha
commited on
Commit
•
c8811f4
1
Parent(s):
a7100c2
Upload folder using huggingface_hub
Browse files- README.json +37 -0
- README.md +36 -0
- config.json +53 -0
README.json
ADDED
@@ -0,0 +1,37 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
{}
|
3 |
+
---
|
4 |
+
{
|
5 |
+
"license": "apache-2.0",
|
6 |
+
"tags": [
|
7 |
+
"merge",
|
8 |
+
"mergekit",
|
9 |
+
"lazymergekit",
|
10 |
+
"allenai/led-base-16384",
|
11 |
+
"allenai/led-base-16384",
|
12 |
+
],
|
13 |
+
"model_name": "LaMa-Merged-slerp",
|
14 |
+
"model": "Your Model Information Here",
|
15 |
+
"description": "LaMa-Merged-slerp is a merge of the following models using mergekit.",
|
16 |
+
"external_datasets": {},
|
17 |
+
"git_repo": "https://github.com/cg123/mergekit",
|
18 |
+
"website": "",
|
19 |
+
"tasks": [],
|
20 |
+
"languages": [],
|
21 |
+
"image": "",
|
22 |
+
"files": [
|
23 |
+
{
|
24 |
+
"filename": "config.json",
|
25 |
+
"type": "yaml",
|
26 |
+
"title": "Configuration"
|
27 |
+
}
|
28 |
+
],
|
29 |
+
"framework": "",
|
30 |
+
"format": "json",
|
31 |
+
"references": [],
|
32 |
+
"contact": "",
|
33 |
+
"creation_date": "2022-03-21",
|
34 |
+
"authors": [
|
35 |
+
"SamaOkasha"
|
36 |
+
]
|
37 |
+
}
|
README.md
ADDED
@@ -0,0 +1,36 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: apache-2.0
|
3 |
+
tags:
|
4 |
+
- merge
|
5 |
+
- mergekit
|
6 |
+
- lazymergekit
|
7 |
+
- allenai/led-base-16384
|
8 |
+
- allenai/led-base-16384
|
9 |
+
---
|
10 |
+
|
11 |
+
# LaMa-Merged-slerp
|
12 |
+
|
13 |
+
LaMa-Merged-slerp is a merge of the following models using [mergekit](https://github.com/cg123/mergekit):
|
14 |
+
* [allenai/led-base-16384](https://huggingface.co/allenai/led-base-16384)
|
15 |
+
* [allenai/led-base-16384](https://huggingface.co/allenai/led-base-16384)
|
16 |
+
|
17 |
+
## 🧩 Configuration
|
18 |
+
|
19 |
+
```yaml
|
20 |
+
slices:
|
21 |
+
- sources:
|
22 |
+
- model: allenai/led-base-16384
|
23 |
+
layer_range: [0, 12] # Taking the initial layers from LED Base model
|
24 |
+
- model: allenai/led-base-16384
|
25 |
+
layer_range: [12, 24] # Taking the later layers from LED Base model
|
26 |
+
merge_method: slerp
|
27 |
+
base_model: allenai/led-base-16384 # Using LED Base model as the base model
|
28 |
+
parameters:
|
29 |
+
t:
|
30 |
+
- filter: self_attn
|
31 |
+
value: [0, 0.5, 0.3, 0.7, 1] # Interpolation values for self-attention layers
|
32 |
+
- filter: mlp
|
33 |
+
value: [1, 0.5, 0.7, 0.3, 0] # Interpolation values for MLP layers
|
34 |
+
- value: 0.5 # Default interpolation value
|
35 |
+
dtype: bfloat16
|
36 |
+
```
|
config.json
ADDED
@@ -0,0 +1,53 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"slices": [
|
3 |
+
{
|
4 |
+
"sources": [
|
5 |
+
{
|
6 |
+
"model": "allenai/led-base-16384",
|
7 |
+
"layer_range": [
|
8 |
+
0,
|
9 |
+
12
|
10 |
+
]
|
11 |
+
},
|
12 |
+
{
|
13 |
+
"model": "allenai/led-base-16384",
|
14 |
+
"layer_range": [
|
15 |
+
12,
|
16 |
+
24
|
17 |
+
]
|
18 |
+
}
|
19 |
+
]
|
20 |
+
}
|
21 |
+
],
|
22 |
+
"merge_method": "slerp",
|
23 |
+
"base_model": "allenai/led-base-16384",
|
24 |
+
"parameters": {
|
25 |
+
"t": [
|
26 |
+
{
|
27 |
+
"filter": "self_attn",
|
28 |
+
"value": [
|
29 |
+
0,
|
30 |
+
0.5,
|
31 |
+
0.3,
|
32 |
+
0.7,
|
33 |
+
1
|
34 |
+
]
|
35 |
+
},
|
36 |
+
{
|
37 |
+
"filter": "mlp",
|
38 |
+
"value": [
|
39 |
+
1,
|
40 |
+
0.5,
|
41 |
+
0.7,
|
42 |
+
0.3,
|
43 |
+
0
|
44 |
+
]
|
45 |
+
},
|
46 |
+
{
|
47 |
+
"value": 0.5
|
48 |
+
}
|
49 |
+
]
|
50 |
+
},
|
51 |
+
"dtype": "bfloat16",
|
52 |
+
"model_type": "led"
|
53 |
+
}
|