Update README.md
Browse files
README.md
CHANGED
@@ -76,16 +76,16 @@ The model was trained on 1201 training samples and 100 validation samples of the
|
|
76 |
### Training hyperparameters
|
77 |
|
78 |
The following hyperparameters were used during training:
|
79 |
-
-
|
80 |
-
-
|
81 |
-
-
|
82 |
-
-
|
83 |
-
-
|
84 |
-
-
|
85 |
-
-
|
86 |
-
-
|
87 |
-
-
|
88 |
-
-
|
89 |
|
90 |
### Training results
|
91 |
|
|
|
76 |
### Training hyperparameters
|
77 |
|
78 |
The following hyperparameters were used during training:
|
79 |
+
- ```learning_rate```: 2e-05
|
80 |
+
- ```train_batch_size```: 1
|
81 |
+
- ```eval_batch_size```: 1
|
82 |
+
- ```seed```: 42
|
83 |
+
- ```gradient_accumulation_steps```: 2
|
84 |
+
- ```total_train_batch_size```: 2
|
85 |
+
- ```optimizer```: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
86 |
+
- ```lr_scheduler_type```: linear
|
87 |
+
- ```num_epochs```: 3
|
88 |
+
- ```mixed_precision_training```: Native AMP
|
89 |
|
90 |
### Training results
|
91 |
|