LawItApps/roberta_echr_truncated_facts_all_labels
Browse files- README.md +4 -23
- config.json +1 -5
- model.safetensors +2 -2
- runs/Aug29_10-13-16_n8ix64biat/events.out.tfevents.1724926397.n8ix64biat.125.1 +3 -0
- runs/Aug31_21-02-06_n2js55wtpp/events.out.tfevents.1725138134.n2js55wtpp.292.0 +3 -0
- runs/Sep01_09-31-58_nnh10dp4rb/events.out.tfevents.1725183123.nnh10dp4rb.76.0 +3 -0
- runs/Sep01_09-31-58_nnh10dp4rb/events.out.tfevents.1725188101.nnh10dp4rb.76.1 +3 -0
- training_args.bin +2 -2
README.md
CHANGED
@@ -3,25 +3,17 @@ license: mit
|
|
3 |
base_model: roberta-base
|
4 |
tags:
|
5 |
- generated_from_trainer
|
6 |
-
metrics:
|
7 |
-
- f1
|
8 |
-
- accuracy
|
9 |
model-index:
|
10 |
-
- name:
|
11 |
results: []
|
12 |
---
|
13 |
|
14 |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
15 |
should probably proofread and complete it, then remove this comment. -->
|
16 |
|
17 |
-
#
|
18 |
|
19 |
This model is a fine-tuned version of [roberta-base](https://huggingface.co/roberta-base) on an unknown dataset.
|
20 |
-
It achieves the following results on the evaluation set:
|
21 |
-
- Loss: 0.0784
|
22 |
-
- F1: 0.7108
|
23 |
-
- Roc Auc: 0.8186
|
24 |
-
- Accuracy: 0.4818
|
25 |
|
26 |
## Model description
|
27 |
|
@@ -41,24 +33,13 @@ More information needed
|
|
41 |
|
42 |
The following hyperparameters were used during training:
|
43 |
- learning_rate: 2e-05
|
44 |
-
- train_batch_size:
|
45 |
-
- eval_batch_size:
|
46 |
- seed: 42
|
47 |
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
48 |
- lr_scheduler_type: linear
|
49 |
- num_epochs: 5
|
50 |
|
51 |
-
### Training results
|
52 |
-
|
53 |
-
| Training Loss | Epoch | Step | Validation Loss | F1 | Roc Auc | Accuracy |
|
54 |
-
|:-------------:|:-----:|:----:|:---------------:|:------:|:-------:|:--------:|
|
55 |
-
| 0.0924 | 1.0 | 1765 | 0.0886 | 0.6469 | 0.7600 | 0.3996 |
|
56 |
-
| 0.0763 | 2.0 | 3530 | 0.0804 | 0.6878 | 0.7951 | 0.4289 |
|
57 |
-
| 0.0678 | 3.0 | 5295 | 0.0780 | 0.7062 | 0.8103 | 0.4653 |
|
58 |
-
| 0.0569 | 4.0 | 7060 | 0.0769 | 0.7091 | 0.8143 | 0.4818 |
|
59 |
-
| 0.0518 | 5.0 | 8825 | 0.0784 | 0.7108 | 0.8186 | 0.4818 |
|
60 |
-
|
61 |
-
|
62 |
### Framework versions
|
63 |
|
64 |
- Transformers 4.35.2
|
|
|
3 |
base_model: roberta-base
|
4 |
tags:
|
5 |
- generated_from_trainer
|
|
|
|
|
|
|
6 |
model-index:
|
7 |
+
- name: roberta_echr_truncated_facts_all_labels
|
8 |
results: []
|
9 |
---
|
10 |
|
11 |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
12 |
should probably proofread and complete it, then remove this comment. -->
|
13 |
|
14 |
+
# roberta_echr_truncated_facts_all_labels
|
15 |
|
16 |
This model is a fine-tuned version of [roberta-base](https://huggingface.co/roberta-base) on an unknown dataset.
|
|
|
|
|
|
|
|
|
|
|
17 |
|
18 |
## Model description
|
19 |
|
|
|
33 |
|
34 |
The following hyperparameters were used during training:
|
35 |
- learning_rate: 2e-05
|
36 |
+
- train_batch_size: 4
|
37 |
+
- eval_batch_size: 4
|
38 |
- seed: 42
|
39 |
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
40 |
- lr_scheduler_type: linear
|
41 |
- num_epochs: 5
|
42 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
43 |
### Framework versions
|
44 |
|
45 |
- Transformers 4.35.2
|
config.json
CHANGED
@@ -38,9 +38,7 @@
|
|
38 |
"24": "LABEL_24",
|
39 |
"25": "LABEL_25",
|
40 |
"26": "LABEL_26",
|
41 |
-
"27": "LABEL_27"
|
42 |
-
"28": "LABEL_28",
|
43 |
-
"29": "LABEL_29"
|
44 |
},
|
45 |
"initializer_range": 0.02,
|
46 |
"intermediate_size": 3072,
|
@@ -66,8 +64,6 @@
|
|
66 |
"LABEL_25": 25,
|
67 |
"LABEL_26": 26,
|
68 |
"LABEL_27": 27,
|
69 |
-
"LABEL_28": 28,
|
70 |
-
"LABEL_29": 29,
|
71 |
"LABEL_3": 3,
|
72 |
"LABEL_4": 4,
|
73 |
"LABEL_5": 5,
|
|
|
38 |
"24": "LABEL_24",
|
39 |
"25": "LABEL_25",
|
40 |
"26": "LABEL_26",
|
41 |
+
"27": "LABEL_27"
|
|
|
|
|
42 |
},
|
43 |
"initializer_range": 0.02,
|
44 |
"intermediate_size": 3072,
|
|
|
64 |
"LABEL_25": 25,
|
65 |
"LABEL_26": 26,
|
66 |
"LABEL_27": 27,
|
|
|
|
|
67 |
"LABEL_3": 3,
|
68 |
"LABEL_4": 4,
|
69 |
"LABEL_5": 5,
|
model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:aa395952bb223062e44ed60140de13740adb62f98ca67d1d76806ddd7130fe4d
|
3 |
+
size 498692800
|
runs/Aug29_10-13-16_n8ix64biat/events.out.tfevents.1724926397.n8ix64biat.125.1
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ff61c76cbc50d60f3f73940e6ba259257fc1f1b1b6ec153387a72386b355bda6
|
3 |
+
size 5913
|
runs/Aug31_21-02-06_n2js55wtpp/events.out.tfevents.1725138134.n2js55wtpp.292.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c9f11f2cf84d3ce9e86b8ce992992a8779192583e5adcff1ebd72d968dedca92
|
3 |
+
size 13487
|
runs/Sep01_09-31-58_nnh10dp4rb/events.out.tfevents.1725183123.nnh10dp4rb.76.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:08e18f04c908eb13b81e850df20751f09c72acf0366e0035f4d32070c7f86c82
|
3 |
+
size 19385
|
runs/Sep01_09-31-58_nnh10dp4rb/events.out.tfevents.1725188101.nnh10dp4rb.76.1
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:650dcbd6a830f184b4bbb8b59f615d352afbb55c9cacbde550a9a61eac32e551
|
3 |
+
size 516
|
training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2f72e5bd54e77557f68a2005ed8d6c46a6a1ca478fcd1a9d2b0e6b1de00f6cb4
|
3 |
+
size 4664
|