AptaArkana commited on
Commit
621c1fe
1 Parent(s): e4f7694

Training complete

Browse files
README.md CHANGED
@@ -1,6 +1,6 @@
1
  ---
2
- license: mit
3
- base_model: cahya/bert-base-indonesian-1.5G
4
  tags:
5
  - generated_from_trainer
6
  metrics:
@@ -18,13 +18,13 @@ should probably proofread and complete it, then remove this comment. -->
18
 
19
  # belajarner_bert_case
20
 
21
- This model is a fine-tuned version of [cahya/bert-base-indonesian-1.5G](https://huggingface.co/cahya/bert-base-indonesian-1.5G) on an unknown dataset.
22
  It achieves the following results on the evaluation set:
23
- - Loss: 0.3302
24
- - Precision: 0.8314
25
- - Recall: 0.8484
26
- - F1: 0.8398
27
- - Accuracy: 0.9471
28
 
29
  ## Model description
30
 
@@ -55,19 +55,19 @@ The following hyperparameters were used during training:
55
 
56
  | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
57
  |:-------------:|:-----:|:-----:|:---------------:|:---------:|:------:|:------:|:--------:|
58
- | 0.1904 | 1.0 | 1567 | 0.2090 | 0.7825 | 0.8298 | 0.8054 | 0.9385 |
59
- | 0.1195 | 2.0 | 3134 | 0.2141 | 0.8312 | 0.8237 | 0.8274 | 0.9446 |
60
- | 0.0772 | 3.0 | 4701 | 0.2200 | 0.8220 | 0.8427 | 0.8322 | 0.9457 |
61
- | 0.0463 | 4.0 | 6268 | 0.2636 | 0.8158 | 0.8498 | 0.8324 | 0.9446 |
62
- | 0.0342 | 5.0 | 7835 | 0.2878 | 0.8351 | 0.8392 | 0.8372 | 0.9456 |
63
- | 0.0223 | 6.0 | 9402 | 0.3066 | 0.8267 | 0.8414 | 0.8340 | 0.9454 |
64
- | 0.0139 | 7.0 | 10969 | 0.3250 | 0.8302 | 0.8476 | 0.8388 | 0.9466 |
65
- | 0.0096 | 8.0 | 12536 | 0.3302 | 0.8314 | 0.8484 | 0.8398 | 0.9471 |
66
 
67
 
68
  ### Framework versions
69
 
70
  - Transformers 4.35.2
71
  - Pytorch 2.1.0+cu121
72
- - Datasets 2.17.0
73
  - Tokenizers 0.15.2
 
1
  ---
2
+ license: afl-3.0
3
+ base_model: Davlan/bert-base-multilingual-cased-ner-hrl
4
  tags:
5
  - generated_from_trainer
6
  metrics:
 
18
 
19
  # belajarner_bert_case
20
 
21
+ This model is a fine-tuned version of [Davlan/bert-base-multilingual-cased-ner-hrl](https://huggingface.co/Davlan/bert-base-multilingual-cased-ner-hrl) on an unknown dataset.
22
  It achieves the following results on the evaluation set:
23
+ - Loss: 0.3367
24
+ - Precision: 0.8139
25
+ - Recall: 0.8422
26
+ - F1: 0.8278
27
+ - Accuracy: 0.9420
28
 
29
  ## Model description
30
 
 
55
 
56
  | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
57
  |:-------------:|:-----:|:-----:|:---------------:|:---------:|:------:|:------:|:--------:|
58
+ | 0.2503 | 1.0 | 1567 | 0.2331 | 0.7484 | 0.8148 | 0.7802 | 0.9294 |
59
+ | 0.1645 | 2.0 | 3134 | 0.2307 | 0.7987 | 0.8158 | 0.8072 | 0.9363 |
60
+ | 0.1097 | 3.0 | 4701 | 0.2588 | 0.7764 | 0.8334 | 0.8039 | 0.9360 |
61
+ | 0.0822 | 4.0 | 6268 | 0.2624 | 0.8056 | 0.8389 | 0.8219 | 0.9409 |
62
+ | 0.061 | 5.0 | 7835 | 0.2927 | 0.8183 | 0.8275 | 0.8229 | 0.9414 |
63
+ | 0.0407 | 6.0 | 9402 | 0.3156 | 0.8021 | 0.8350 | 0.8182 | 0.9399 |
64
+ | 0.0315 | 7.0 | 10969 | 0.3257 | 0.8102 | 0.8381 | 0.8239 | 0.9413 |
65
+ | 0.0238 | 8.0 | 12536 | 0.3367 | 0.8139 | 0.8422 | 0.8278 | 0.9420 |
66
 
67
 
68
  ### Framework versions
69
 
70
  - Transformers 4.35.2
71
  - Pytorch 2.1.0+cu121
72
+ - Datasets 2.17.1
73
  - Tokenizers 0.15.2
config.json CHANGED
@@ -1,97 +1,98 @@
1
  {
2
- "_name_or_path": "cahya/bert-base-indonesian-1.5G",
3
  "architectures": [
4
  "BertForTokenClassification"
5
  ],
6
  "attention_probs_dropout_prob": 0.1,
7
  "classifier_dropout": null,
 
8
  "gradient_checkpointing": false,
9
  "hidden_act": "gelu",
10
  "hidden_dropout_prob": 0.1,
11
  "hidden_size": 768,
12
  "id2label": {
13
- "0": "B-CRD",
14
- "1": "B-DAT",
15
- "2": "B-EVT",
16
- "3": "B-FAC",
17
- "4": "B-GPE",
18
- "5": "B-LAN",
19
- "6": "B-LAW",
20
- "7": "B-LOC",
21
- "8": "B-MON",
22
- "9": "B-NOR",
23
- "10": "B-ORD",
24
- "11": "B-ORG",
25
- "12": "B-PER",
26
- "13": "B-PRC",
27
- "14": "B-PRD",
28
- "15": "B-QTY",
29
- "16": "B-REG",
30
- "17": "B-TIM",
31
- "18": "B-WOA",
32
- "19": "I-CRD",
33
- "20": "I-DAT",
34
- "21": "I-EVT",
35
- "22": "I-FAC",
36
- "23": "I-GPE",
37
- "24": "I-LAN",
38
- "25": "I-LAW",
39
- "26": "I-LOC",
40
- "27": "I-MON",
41
- "28": "I-NOR",
42
- "29": "I-ORD",
43
- "30": "I-ORG",
44
- "31": "I-PER",
45
- "32": "I-PRC",
46
- "33": "I-PRD",
47
- "34": "I-QTY",
48
- "35": "I-REG",
49
  "36": "I-TIM",
50
- "37": "I-WOA",
51
- "38": "O"
52
  },
53
  "initializer_range": 0.02,
54
  "intermediate_size": 3072,
55
  "label2id": {
56
- "B-CRD": 0,
57
- "B-DAT": 1,
58
- "B-EVT": 2,
59
- "B-FAC": 3,
60
- "B-GPE": 4,
61
- "B-LAN": 5,
62
- "B-LAW": 6,
63
- "B-LOC": 7,
64
- "B-MON": 8,
65
- "B-NOR": 9,
66
- "B-ORD": 10,
67
- "B-ORG": 11,
68
- "B-PER": 12,
69
- "B-PRC": 13,
70
- "B-PRD": 14,
71
- "B-QTY": 15,
72
- "B-REG": 16,
73
- "B-TIM": 17,
74
- "B-WOA": 18,
75
- "I-CRD": 19,
76
- "I-DAT": 20,
77
- "I-EVT": 21,
78
- "I-FAC": 22,
79
- "I-GPE": 23,
80
- "I-LAN": 24,
81
- "I-LAW": 25,
82
- "I-LOC": 26,
83
- "I-MON": 27,
84
- "I-NOR": 28,
85
- "I-ORD": 29,
86
- "I-ORG": 30,
87
- "I-PER": 31,
88
- "I-PRC": 32,
89
- "I-PRD": 33,
90
- "I-QTY": 34,
91
- "I-REG": 35,
92
  "I-TIM": 36,
93
- "I-WOA": 37,
94
- "O": 38
95
  },
96
  "layer_norm_eps": 1e-12,
97
  "max_position_embeddings": 512,
@@ -99,10 +100,15 @@
99
  "num_attention_heads": 12,
100
  "num_hidden_layers": 12,
101
  "pad_token_id": 0,
 
 
 
 
 
102
  "position_embedding_type": "absolute",
103
  "torch_dtype": "float32",
104
  "transformers_version": "4.35.2",
105
  "type_vocab_size": 2,
106
  "use_cache": true,
107
- "vocab_size": 32000
108
  }
 
1
  {
2
+ "_name_or_path": "Davlan/bert-base-multilingual-cased-ner-hrl",
3
  "architectures": [
4
  "BertForTokenClassification"
5
  ],
6
  "attention_probs_dropout_prob": 0.1,
7
  "classifier_dropout": null,
8
+ "directionality": "bidi",
9
  "gradient_checkpointing": false,
10
  "hidden_act": "gelu",
11
  "hidden_dropout_prob": 0.1,
12
  "hidden_size": 768,
13
  "id2label": {
14
+ "0": "O",
15
+ "1": "B-CRD",
16
+ "2": "I-CRD",
17
+ "3": "B-DAT",
18
+ "4": "I-DAT",
19
+ "5": "B-EVT",
20
+ "6": "I-EVT",
21
+ "7": "B-FAC",
22
+ "8": "I-FAC",
23
+ "9": "B-GPE",
24
+ "10": "I-GPE",
25
+ "11": "B-LAN",
26
+ "12": "I-LAN",
27
+ "13": "B-LAW",
28
+ "14": "I-LAW",
29
+ "15": "B-LOC",
30
+ "16": "I-LOC",
31
+ "17": "B-MON",
32
+ "18": "I-MON",
33
+ "19": "B-NOR",
34
+ "20": "I-NOR",
35
+ "21": "B-ORD",
36
+ "22": "I-ORD",
37
+ "23": "B-ORG",
38
+ "24": "I-ORG",
39
+ "25": "B-PER",
40
+ "26": "I-PER",
41
+ "27": "B-PRC",
42
+ "28": "I-PRC",
43
+ "29": "B-PRD",
44
+ "30": "I-PRD",
45
+ "31": "B-QTY",
46
+ "32": "I-QTY",
47
+ "33": "B-REG",
48
+ "34": "I-REG",
49
+ "35": "B-TIM",
50
  "36": "I-TIM",
51
+ "37": "B-WOA",
52
+ "38": "I-WOA"
53
  },
54
  "initializer_range": 0.02,
55
  "intermediate_size": 3072,
56
  "label2id": {
57
+ "B-CRD": 1,
58
+ "B-DAT": 3,
59
+ "B-EVT": 5,
60
+ "B-FAC": 7,
61
+ "B-GPE": 9,
62
+ "B-LAN": 11,
63
+ "B-LAW": 13,
64
+ "B-LOC": 15,
65
+ "B-MON": 17,
66
+ "B-NOR": 19,
67
+ "B-ORD": 21,
68
+ "B-ORG": 23,
69
+ "B-PER": 25,
70
+ "B-PRC": 27,
71
+ "B-PRD": 29,
72
+ "B-QTY": 31,
73
+ "B-REG": 33,
74
+ "B-TIM": 35,
75
+ "B-WOA": 37,
76
+ "I-CRD": 2,
77
+ "I-DAT": 4,
78
+ "I-EVT": 6,
79
+ "I-FAC": 8,
80
+ "I-GPE": 10,
81
+ "I-LAN": 12,
82
+ "I-LAW": 14,
83
+ "I-LOC": 16,
84
+ "I-MON": 18,
85
+ "I-NOR": 20,
86
+ "I-ORD": 22,
87
+ "I-ORG": 24,
88
+ "I-PER": 26,
89
+ "I-PRC": 28,
90
+ "I-PRD": 30,
91
+ "I-QTY": 32,
92
+ "I-REG": 34,
93
  "I-TIM": 36,
94
+ "I-WOA": 38,
95
+ "O": 0
96
  },
97
  "layer_norm_eps": 1e-12,
98
  "max_position_embeddings": 512,
 
100
  "num_attention_heads": 12,
101
  "num_hidden_layers": 12,
102
  "pad_token_id": 0,
103
+ "pooler_fc_size": 768,
104
+ "pooler_num_attention_heads": 12,
105
+ "pooler_num_fc_layers": 3,
106
+ "pooler_size_per_head": 128,
107
+ "pooler_type": "first_token_transform",
108
  "position_embedding_type": "absolute",
109
  "torch_dtype": "float32",
110
  "transformers_version": "4.35.2",
111
  "type_vocab_size": 2,
112
  "use_cache": true,
113
+ "vocab_size": 119547
114
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f48a4068dfc6c563caa3269fa4354d48b78e054e08c560b9d28fa3c9b46343dc
3
- size 440250324
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3c5ff0211f38026f5243f61a377f5e7f5d8e834144928cff583aa6dfd93a67de
3
+ size 709194716
runs/Feb20_02-50-19_21f700017ff9/events.out.tfevents.1708397421.21f700017ff9.191.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b98dd95f9bd184fed07eca29f126ee7a893236f569fd66035cb7ce48927970d8
3
+ size 13914
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -1,38 +1,38 @@
1
  {
2
  "added_tokens_decoder": {
3
  "0": {
4
- "content": "[UNK]",
5
  "lstrip": false,
6
  "normalized": false,
7
  "rstrip": false,
8
  "single_word": false,
9
  "special": true
10
  },
11
- "1": {
12
- "content": "[SEP]",
13
  "lstrip": false,
14
  "normalized": false,
15
  "rstrip": false,
16
  "single_word": false,
17
  "special": true
18
  },
19
- "2": {
20
- "content": "[PAD]",
21
  "lstrip": false,
22
  "normalized": false,
23
  "rstrip": false,
24
  "single_word": false,
25
  "special": true
26
  },
27
- "3": {
28
- "content": "[CLS]",
29
  "lstrip": false,
30
  "normalized": false,
31
  "rstrip": false,
32
  "single_word": false,
33
  "special": true
34
  },
35
- "4": {
36
  "content": "[MASK]",
37
  "lstrip": false,
38
  "normalized": false,
@@ -44,10 +44,9 @@
44
  "clean_up_tokenization_spaces": true,
45
  "cls_token": "[CLS]",
46
  "do_basic_tokenize": true,
47
- "do_lower_case": true,
48
- "full_tokenizer_file": null,
49
  "mask_token": "[MASK]",
50
- "model_max_length": 1000000000000000019884624838656,
51
  "never_split": null,
52
  "pad_token": "[PAD]",
53
  "sep_token": "[SEP]",
 
1
  {
2
  "added_tokens_decoder": {
3
  "0": {
4
+ "content": "[PAD]",
5
  "lstrip": false,
6
  "normalized": false,
7
  "rstrip": false,
8
  "single_word": false,
9
  "special": true
10
  },
11
+ "100": {
12
+ "content": "[UNK]",
13
  "lstrip": false,
14
  "normalized": false,
15
  "rstrip": false,
16
  "single_word": false,
17
  "special": true
18
  },
19
+ "101": {
20
+ "content": "[CLS]",
21
  "lstrip": false,
22
  "normalized": false,
23
  "rstrip": false,
24
  "single_word": false,
25
  "special": true
26
  },
27
+ "102": {
28
+ "content": "[SEP]",
29
  "lstrip": false,
30
  "normalized": false,
31
  "rstrip": false,
32
  "single_word": false,
33
  "special": true
34
  },
35
+ "103": {
36
  "content": "[MASK]",
37
  "lstrip": false,
38
  "normalized": false,
 
44
  "clean_up_tokenization_spaces": true,
45
  "cls_token": "[CLS]",
46
  "do_basic_tokenize": true,
47
+ "do_lower_case": false,
 
48
  "mask_token": "[MASK]",
49
+ "model_max_length": 512,
50
  "never_split": null,
51
  "pad_token": "[PAD]",
52
  "sep_token": "[SEP]",
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:089227e81d92fd690978c5ba0e6b83d332833d004b1e9c55aab1b9fbaff3fc02
3
  size 4600
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aa0914b1da1ced7966de88c48803983604e5c535c3c6a4412d9e069e170a0add
3
  size 4600
vocab.txt CHANGED
The diff for this file is too large to render. See raw diff