Rami commited on
Commit
6794b02
1 Parent(s): f7bbfec

update model card README.md

Browse files
Files changed (1) hide show
  1. README.md +21 -21
README.md CHANGED
@@ -1,5 +1,4 @@
1
  ---
2
- license: apache-2.0
3
  tags:
4
  - generated_from_trainer
5
  model-index:
@@ -12,11 +11,11 @@ should probably proofread and complete it, then remove this comment. -->
12
 
13
  # multi-label-class-classification-on-github-issues
14
 
15
- This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the None dataset.
16
  It achieves the following results on the evaluation set:
17
- - Loss: 0.1391
18
- - Micro f1: 0.5005
19
- - Macro f1: 0.0340
20
 
21
  ## Model description
22
 
@@ -36,32 +35,33 @@ More information needed
36
 
37
  The following hyperparameters were used during training:
38
  - learning_rate: 3e-05
39
- - train_batch_size: 128
40
  - eval_batch_size: 8
41
  - seed: 42
42
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
43
  - lr_scheduler_type: linear
44
  - num_epochs: 15
 
45
 
46
  ### Training results
47
 
48
  | Training Loss | Epoch | Step | Validation Loss | Micro f1 | Macro f1 |
49
  |:-------------:|:-----:|:----:|:---------------:|:--------:|:--------:|
50
- | No log | 1.0 | 13 | 0.3988 | 0.3783 | 0.0172 |
51
- | No log | 2.0 | 26 | 0.3288 | 0.3791 | 0.0172 |
52
- | No log | 3.0 | 39 | 0.2753 | 0.3791 | 0.0172 |
53
- | No log | 4.0 | 52 | 0.2347 | 0.3791 | 0.0172 |
54
- | No log | 5.0 | 65 | 0.2051 | 0.3791 | 0.0172 |
55
- | No log | 6.0 | 78 | 0.1849 | 0.3791 | 0.0172 |
56
- | No log | 7.0 | 91 | 0.1710 | 0.3791 | 0.0172 |
57
- | No log | 8.0 | 104 | 0.1614 | 0.3791 | 0.0172 |
58
- | No log | 9.0 | 117 | 0.1546 | 0.3791 | 0.0172 |
59
- | No log | 10.0 | 130 | 0.1497 | 0.3796 | 0.0173 |
60
- | No log | 11.0 | 143 | 0.1458 | 0.4143 | 0.0238 |
61
- | No log | 12.0 | 156 | 0.1427 | 0.4567 | 0.0295 |
62
- | No log | 13.0 | 169 | 0.1406 | 0.4709 | 0.0310 |
63
- | No log | 14.0 | 182 | 0.1391 | 0.4886 | 0.0327 |
64
- | No log | 15.0 | 195 | 0.1386 | 0.4909 | 0.0330 |
65
 
66
 
67
  ### Framework versions
 
1
  ---
 
2
  tags:
3
  - generated_from_trainer
4
  model-index:
 
11
 
12
  # multi-label-class-classification-on-github-issues
13
 
14
+ This model is a fine-tuned version of [neuralmagic/oBERT-12-upstream-pruned-unstructured-97](https://huggingface.co/neuralmagic/oBERT-12-upstream-pruned-unstructured-97) on the None dataset.
15
  It achieves the following results on the evaluation set:
16
+ - Loss: 0.1301
17
+ - Micro f1: 0.5159
18
+ - Macro f1: 0.0352
19
 
20
  ## Model description
21
 
 
35
 
36
  The following hyperparameters were used during training:
37
  - learning_rate: 3e-05
38
+ - train_batch_size: 64
39
  - eval_batch_size: 8
40
  - seed: 42
41
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
42
  - lr_scheduler_type: linear
43
  - num_epochs: 15
44
+ - mixed_precision_training: Native AMP
45
 
46
  ### Training results
47
 
48
  | Training Loss | Epoch | Step | Validation Loss | Micro f1 | Macro f1 |
49
  |:-------------:|:-----:|:----:|:---------------:|:--------:|:--------:|
50
+ | No log | 1.0 | 25 | 0.3994 | 0.3783 | 0.0172 |
51
+ | No log | 2.0 | 50 | 0.2846 | 0.3791 | 0.0172 |
52
+ | No log | 3.0 | 75 | 0.2159 | 0.3791 | 0.0172 |
53
+ | No log | 4.0 | 100 | 0.1802 | 0.3791 | 0.0172 |
54
+ | No log | 5.0 | 125 | 0.1618 | 0.3791 | 0.0172 |
55
+ | No log | 6.0 | 150 | 0.1515 | 0.3791 | 0.0172 |
56
+ | No log | 7.0 | 175 | 0.1452 | 0.3791 | 0.0172 |
57
+ | No log | 8.0 | 200 | 0.1411 | 0.3931 | 0.0202 |
58
+ | No log | 9.0 | 225 | 0.1379 | 0.4413 | 0.0277 |
59
+ | No log | 10.0 | 250 | 0.1350 | 0.4694 | 0.0309 |
60
+ | No log | 11.0 | 275 | 0.1327 | 0.4993 | 0.0336 |
61
+ | No log | 12.0 | 300 | 0.1309 | 0.5084 | 0.0344 |
62
+ | No log | 13.0 | 325 | 0.1297 | 0.5147 | 0.0349 |
63
+ | No log | 14.0 | 350 | 0.1291 | 0.5060 | 0.0343 |
64
+ | No log | 15.0 | 375 | 0.1287 | 0.5107 | 0.0346 |
65
 
66
 
67
  ### Framework versions