jumtul commited on
Commit
76110eb
1 Parent(s): 85cba7a

model card

Browse files
Files changed (1) hide show
  1. README.md +73 -0
README.md ADDED
@@ -0,0 +1,73 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - ko
4
+ base_model:
5
+ - LDCC/LDCC-SOLAR-10.7B
6
+ - hyeogi/SOLAR-10.7B-dpo-v1
7
+ tags:
8
+ - mergekit
9
+ - merge
10
+ - LDCC/LDCC-SOLAR-10.7B
11
+ - hyeogi/SOLAR-10.7B-dpo-v1
12
+ license: apache-2.0
13
+ ---
14
+ # merge
15
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
16
+
17
+ ## Merge Details
18
+ ### Merge Method
19
+
20
+ This model was merged using the SLERP merge method.
21
+
22
+ ### Models Merged
23
+
24
+ The following models were included in the merge:
25
+ * [hyeogi/SOLAR-10.7B-dpo-v1](https://huggingface.co/hyeogi/SOLAR-10.7B-dpo-v1)
26
+ * [LDCC/LDCC-SOLAR-10.7B](https://huggingface.co/LDCC/LDCC-SOLAR-10.7B)
27
+
28
+ ### Configuration
29
+
30
+ The following YAML configuration was used to produce this model:
31
+
32
+ ```yaml
33
+
34
+
35
+ slices:
36
+
37
+ - sources:
38
+
39
+ - model: LDCC/LDCC-SOLAR-10.7B
40
+
41
+ layer_range: [0, 48]
42
+
43
+ - model: hyeogi/SOLAR-10.7B-dpo-v1
44
+
45
+ layer_range: [0, 48]
46
+
47
+ merge_method: slerp
48
+ tokenizer_source: base
49
+ base_model: LDCC/LDCC-SOLAR-10.7B
50
+
51
+ embed_slerp: true
52
+
53
+ parameters:
54
+
55
+ t:
56
+
57
+ - filter: self_attn
58
+
59
+ value: [0, 0.5, 0.3, 0.7, 1]
60
+
61
+ - filter: mlp
62
+
63
+ value: [1, 0.5, 0.7, 0.3, 0]
64
+
65
+ - value: 0.5
66
+
67
+ dtype: bfloat16
68
+
69
+
70
+ ```
71
+ ## Datasets
72
+
73
+ Finetuned using LoRA with [kyujinpy/OpenOrca-KO](https://huggingface.co/datasets/kyujinpy/OpenOrca-KO)