AptaArkana commited on
Commit
6a16f22
1 Parent(s): 85c56f5

Training complete

Browse files
README.md CHANGED
@@ -1,6 +1,6 @@
1
  ---
2
  license: mit
3
- base_model: indolem/indobert-base-uncased
4
  tags:
5
  - generated_from_trainer
6
  datasets:
@@ -25,16 +25,16 @@ model-index:
25
  metrics:
26
  - name: Precision
27
  type: precision
28
- value: 0.7905982905982906
29
  - name: Recall
30
  type: recall
31
- value: 0.8383685800604229
32
  - name: F1
33
  type: f1
34
- value: 0.8137829912023461
35
  - name: Accuracy
36
  type: accuracy
37
- value: 0.9516761543327008
38
  ---
39
 
40
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -42,13 +42,13 @@ should probably proofread and complete it, then remove this comment. -->
42
 
43
  # belajarner
44
 
45
- This model is a fine-tuned version of [indolem/indobert-base-uncased](https://huggingface.co/indolem/indobert-base-uncased) on the indonlu_nergrit dataset.
46
  It achieves the following results on the evaluation set:
47
- - Loss: 0.2235
48
- - Precision: 0.7906
49
- - Recall: 0.8384
50
- - F1: 0.8138
51
- - Accuracy: 0.9517
52
 
53
  ## Model description
54
 
@@ -79,14 +79,14 @@ The following hyperparameters were used during training:
79
 
80
  | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
81
  |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:|
82
- | No log | 1.0 | 209 | 0.1750 | 0.7418 | 0.8157 | 0.7770 | 0.9469 |
83
- | No log | 2.0 | 418 | 0.1590 | 0.7677 | 0.8338 | 0.7994 | 0.9491 |
84
- | 0.2398 | 3.0 | 627 | 0.1720 | 0.7817 | 0.8112 | 0.7961 | 0.9476 |
85
- | 0.2398 | 4.0 | 836 | 0.1812 | 0.7948 | 0.8248 | 0.8095 | 0.9510 |
86
- | 0.0753 | 5.0 | 1045 | 0.1934 | 0.7872 | 0.8384 | 0.8120 | 0.9545 |
87
- | 0.0753 | 6.0 | 1254 | 0.2178 | 0.7805 | 0.8323 | 0.8056 | 0.9497 |
88
- | 0.0753 | 7.0 | 1463 | 0.2199 | 0.7943 | 0.8459 | 0.8193 | 0.9522 |
89
- | 0.0374 | 8.0 | 1672 | 0.2235 | 0.7906 | 0.8384 | 0.8138 | 0.9517 |
90
 
91
 
92
  ### Framework versions
 
1
  ---
2
  license: mit
3
+ base_model: cahya/bert-base-indonesian-NER
4
  tags:
5
  - generated_from_trainer
6
  datasets:
 
25
  metrics:
26
  - name: Precision
27
  type: precision
28
+ value: 0.7716312056737589
29
  - name: Recall
30
  type: recall
31
+ value: 0.8217522658610272
32
  - name: F1
33
  type: f1
34
+ value: 0.7959034381858083
35
  - name: Accuracy
36
  type: accuracy
37
+ value: 0.9477048970719857
38
  ---
39
 
40
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
42
 
43
  # belajarner
44
 
45
+ This model is a fine-tuned version of [cahya/bert-base-indonesian-NER](https://huggingface.co/cahya/bert-base-indonesian-NER) on the indonlu_nergrit dataset.
46
  It achieves the following results on the evaluation set:
47
+ - Loss: 0.2621
48
+ - Precision: 0.7716
49
+ - Recall: 0.8218
50
+ - F1: 0.7959
51
+ - Accuracy: 0.9477
52
 
53
  ## Model description
54
 
 
79
 
80
  | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
81
  |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:|
82
+ | No log | 1.0 | 209 | 0.1633 | 0.7678 | 0.8142 | 0.7903 | 0.9476 |
83
+ | No log | 2.0 | 418 | 0.1623 | 0.7631 | 0.8127 | 0.7871 | 0.9462 |
84
+ | 0.1633 | 3.0 | 627 | 0.1978 | 0.7535 | 0.8172 | 0.7841 | 0.9459 |
85
+ | 0.1633 | 4.0 | 836 | 0.2103 | 0.7573 | 0.8202 | 0.7875 | 0.9460 |
86
+ | 0.0423 | 5.0 | 1045 | 0.2236 | 0.7757 | 0.8097 | 0.7923 | 0.9487 |
87
+ | 0.0423 | 6.0 | 1254 | 0.2529 | 0.7843 | 0.8293 | 0.8062 | 0.9474 |
88
+ | 0.0423 | 7.0 | 1463 | 0.2559 | 0.77 | 0.8142 | 0.7915 | 0.9467 |
89
+ | 0.0136 | 8.0 | 1672 | 0.2621 | 0.7716 | 0.8218 | 0.7959 | 0.9477 |
90
 
91
 
92
  ### Framework versions
config.json CHANGED
@@ -1,12 +1,11 @@
1
  {
2
- "_name_or_path": "indolem/indobert-base-uncased",
3
  "architectures": [
4
  "BertForTokenClassification"
5
  ],
6
  "attention_probs_dropout_prob": 0.1,
7
- "bos_token_id": 0,
8
  "classifier_dropout": null,
9
- "eos_token_ids": 0,
10
  "hidden_act": "gelu",
11
  "hidden_dropout_prob": 0.1,
12
  "hidden_size": 768,
@@ -35,12 +34,11 @@
35
  "model_type": "bert",
36
  "num_attention_heads": 12,
37
  "num_hidden_layers": 12,
38
- "output_past": true,
39
  "pad_token_id": 0,
40
  "position_embedding_type": "absolute",
41
  "torch_dtype": "float32",
42
  "transformers_version": "4.35.2",
43
  "type_vocab_size": 2,
44
  "use_cache": true,
45
- "vocab_size": 31923
46
  }
 
1
  {
2
+ "_name_or_path": "cahya/bert-base-indonesian-NER",
3
  "architectures": [
4
  "BertForTokenClassification"
5
  ],
6
  "attention_probs_dropout_prob": 0.1,
 
7
  "classifier_dropout": null,
8
+ "gradient_checkpointing": false,
9
  "hidden_act": "gelu",
10
  "hidden_dropout_prob": 0.1,
11
  "hidden_size": 768,
 
34
  "model_type": "bert",
35
  "num_attention_heads": 12,
36
  "num_hidden_layers": 12,
 
37
  "pad_token_id": 0,
38
  "position_embedding_type": "absolute",
39
  "torch_dtype": "float32",
40
  "transformers_version": "4.35.2",
41
  "type_vocab_size": 2,
42
  "use_cache": true,
43
+ "vocab_size": 32000
44
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:dfff3039d4cddf8b238e162b9c970f03f8fedd66952cfad1842bc5790bdad899
3
- size 439915340
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c91e35db013f96ddb47572c0bfbad876fb17a171656a56d2961e52aebf1d61b1
3
+ size 440151884
runs/Feb19_02-09-05_51cb7f447e43/events.out.tfevents.1708308548.51cb7f447e43.919.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a3fa2a3aa31dba2bbf2782631a3eb5b1297551d349341d71845d8d63c260422b
3
+ size 9135
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "added_tokens_decoder": {
3
  "0": {
4
- "content": "[PAD]",
5
  "lstrip": false,
6
  "normalized": false,
7
  "rstrip": false,
@@ -9,7 +9,7 @@
9
  "special": true
10
  },
11
  "1": {
12
- "content": "[UNK]",
13
  "lstrip": false,
14
  "normalized": false,
15
  "rstrip": false,
@@ -17,7 +17,7 @@
17
  "special": true
18
  },
19
  "2": {
20
- "content": "[MASK]",
21
  "lstrip": false,
22
  "normalized": false,
23
  "rstrip": false,
@@ -33,7 +33,7 @@
33
  "special": true
34
  },
35
  "4": {
36
- "content": "[SEP]",
37
  "lstrip": false,
38
  "normalized": false,
39
  "rstrip": false,
@@ -45,6 +45,7 @@
45
  "cls_token": "[CLS]",
46
  "do_basic_tokenize": true,
47
  "do_lower_case": true,
 
48
  "mask_token": "[MASK]",
49
  "model_max_length": 1000000000000000019884624838656,
50
  "never_split": null,
 
1
  {
2
  "added_tokens_decoder": {
3
  "0": {
4
+ "content": "[UNK]",
5
  "lstrip": false,
6
  "normalized": false,
7
  "rstrip": false,
 
9
  "special": true
10
  },
11
  "1": {
12
+ "content": "[SEP]",
13
  "lstrip": false,
14
  "normalized": false,
15
  "rstrip": false,
 
17
  "special": true
18
  },
19
  "2": {
20
+ "content": "[PAD]",
21
  "lstrip": false,
22
  "normalized": false,
23
  "rstrip": false,
 
33
  "special": true
34
  },
35
  "4": {
36
+ "content": "[MASK]",
37
  "lstrip": false,
38
  "normalized": false,
39
  "rstrip": false,
 
45
  "cls_token": "[CLS]",
46
  "do_basic_tokenize": true,
47
  "do_lower_case": true,
48
+ "full_tokenizer_file": null,
49
  "mask_token": "[MASK]",
50
  "model_max_length": 1000000000000000019884624838656,
51
  "never_split": null,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e4ed29da338e3530fe523346f79bf0087cdb1f41c81618a56fb4712afb10311d
3
  size 4536
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1ea1fcab412f0bd86e96dcf5c086dbfbbdb78863f4bc80a9f8e9fd9187208469
3
  size 4536
vocab.txt CHANGED
The diff for this file is too large to render. See raw diff