ashhadahsan commited on
Commit
7064ce9
1 Parent(s): 7f3f810

Training in progress epoch 0

Browse files
README.md CHANGED
@@ -15,11 +15,11 @@ probably proofread and complete it, then remove this comment. -->
15
 
16
  This model is a fine-tuned version of [bert-base-uncased](https://huggingface.co/bert-base-uncased) on an unknown dataset.
17
  It achieves the following results on the evaluation set:
18
- - Train Loss: 0.0048
19
- - Train Accuracy: 1.0
20
- - Validation Loss: 3.5602
21
- - Validation Accuracy: 0.5128
22
- - Epoch: 7
23
 
24
  ## Model description
25
 
@@ -45,19 +45,12 @@ The following hyperparameters were used during training:
45
 
46
  | Train Loss | Train Accuracy | Validation Loss | Validation Accuracy | Epoch |
47
  |:----------:|:--------------:|:---------------:|:-------------------:|:-----:|
48
- | 0.0367 | 0.9990 | 3.0998 | 0.4872 | 0 |
49
- | 0.0184 | 1.0 | 3.1577 | 0.5128 | 1 |
50
- | 0.0136 | 1.0 | 3.3107 | 0.4615 | 2 |
51
- | 0.0099 | 1.0 | 3.3581 | 0.4872 | 3 |
52
- | 0.0078 | 1.0 | 3.4509 | 0.5128 | 4 |
53
- | 0.0063 | 1.0 | 3.4978 | 0.5128 | 5 |
54
- | 0.0054 | 1.0 | 3.6166 | 0.4872 | 6 |
55
- | 0.0048 | 1.0 | 3.5602 | 0.5128 | 7 |
56
 
57
 
58
  ### Framework versions
59
 
60
- - Transformers 4.34.0
61
- - TensorFlow 2.13.0
62
- - Datasets 2.14.5
63
- - Tokenizers 0.14.1
 
15
 
16
  This model is a fine-tuned version of [bert-base-uncased](https://huggingface.co/bert-base-uncased) on an unknown dataset.
17
  It achieves the following results on the evaluation set:
18
+ - Train Loss: 3.5410
19
+ - Train Accuracy: 0.2497
20
+ - Validation Loss: 3.0774
21
+ - Validation Accuracy: 0.3333
22
+ - Epoch: 0
23
 
24
  ## Model description
25
 
 
45
 
46
  | Train Loss | Train Accuracy | Validation Loss | Validation Accuracy | Epoch |
47
  |:----------:|:--------------:|:---------------:|:-------------------:|:-----:|
48
+ | 3.5410 | 0.2497 | 3.0774 | 0.3333 | 0 |
 
 
 
 
 
 
 
49
 
50
 
51
  ### Framework versions
52
 
53
+ - Transformers 4.33.0
54
+ - TensorFlow 2.12.0
55
+ - Datasets 2.1.0
56
+ - Tokenizers 0.13.3
config.json CHANGED
@@ -184,7 +184,7 @@
184
  "num_hidden_layers": 12,
185
  "pad_token_id": 0,
186
  "position_embedding_type": "absolute",
187
- "transformers_version": "4.34.0",
188
  "type_vocab_size": 2,
189
  "use_cache": true,
190
  "vocab_size": 30522
 
184
  "num_hidden_layers": 12,
185
  "pad_token_id": 0,
186
  "position_embedding_type": "absolute",
187
+ "transformers_version": "4.33.0",
188
  "type_vocab_size": 2,
189
  "use_cache": true,
190
  "vocab_size": 30522
logs/train/events.out.tfevents.1697308621.99df1aa9d054.28.0.v2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:02cf3c11cb839271e992283241f2ae2fa21f528dcbf1934bd477415b46d028d4
3
+ size 4190389
logs/validation/events.out.tfevents.1697308729.99df1aa9d054.28.1.v2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a30293994f49d12d22e33ad581540ac68dea365c46653dc2c8f71668e6a4f636
3
+ size 392
tf_model.h5 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5d428a21b8fb21d3917ef150feede837c504b28d39f208677fc2798427ab642d
3
  size 438466132
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:074fb72c9a0496e6a2c9ae5a23dde44b311ec9fb6fad41e5980d0badc72decad
3
  size 438466132
tokenizer_config.json CHANGED
@@ -1,47 +1,4 @@
1
  {
2
- "added_tokens_decoder": {
3
- "0": {
4
- "content": "[PAD]",
5
- "lstrip": false,
6
- "normalized": false,
7
- "rstrip": false,
8
- "single_word": false,
9
- "special": false
10
- },
11
- "100": {
12
- "content": "[UNK]",
13
- "lstrip": false,
14
- "normalized": false,
15
- "rstrip": false,
16
- "single_word": false,
17
- "special": false
18
- },
19
- "101": {
20
- "content": "[CLS]",
21
- "lstrip": false,
22
- "normalized": false,
23
- "rstrip": false,
24
- "single_word": false,
25
- "special": false
26
- },
27
- "102": {
28
- "content": "[SEP]",
29
- "lstrip": false,
30
- "normalized": false,
31
- "rstrip": false,
32
- "single_word": false,
33
- "special": false
34
- },
35
- "103": {
36
- "content": "[MASK]",
37
- "lstrip": false,
38
- "normalized": false,
39
- "rstrip": false,
40
- "single_word": false,
41
- "special": false
42
- }
43
- },
44
- "additional_special_tokens": [],
45
  "clean_up_tokenization_spaces": true,
46
  "cls_token": "[CLS]",
47
  "do_basic_tokenize": true,
@@ -54,6 +11,5 @@
54
  "strip_accents": null,
55
  "tokenize_chinese_chars": true,
56
  "tokenizer_class": "BertTokenizer",
57
- "tokenizer_file": "/root/.cache/huggingface/hub/models--bert-base-uncased/snapshots/1dbc166cf8765166998eff31ade2eb64c8a40076/tokenizer.json",
58
  "unk_token": "[UNK]"
59
  }
 
1
  {
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2
  "clean_up_tokenization_spaces": true,
3
  "cls_token": "[CLS]",
4
  "do_basic_tokenize": true,
 
11
  "strip_accents": null,
12
  "tokenize_chinese_chars": true,
13
  "tokenizer_class": "BertTokenizer",
 
14
  "unk_token": "[UNK]"
15
  }