janbakker commited on
Commit
0172c8f
1 Parent(s): 4fec29c

Upload RobertaForSequenceClassification

Browse files
Files changed (2) hide show
  1. config.json +123 -0
  2. pytorch_model.bin +3 -0
config.json ADDED
@@ -0,0 +1,123 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "roberta-base",
3
+ "accelerator": null,
4
+ "accumulate_grad_batches": null,
5
+ "add_context": false,
6
+ "amp_backend": null,
7
+ "amp_level": null,
8
+ "architectures": [
9
+ "RobertaForSequenceClassification"
10
+ ],
11
+ "attention_probs_dropout_prob": 0.1,
12
+ "auto_lr_find": false,
13
+ "auto_scale_batch_size": false,
14
+ "auto_select_gpus": null,
15
+ "batch_size": 32,
16
+ "benchmark": null,
17
+ "binary_clf": false,
18
+ "bos_token_id": 0,
19
+ "check_val_every_n_epoch": 1,
20
+ "checkpoint": null,
21
+ "ckpt_metric": "val_macro_f1",
22
+ "classifier_dropout": null,
23
+ "context_dir": null,
24
+ "context_doc_id": null,
25
+ "context_window": 5,
26
+ "default_root_dir": null,
27
+ "detect_anomaly": false,
28
+ "devices": "2",
29
+ "doc_pos_embeds": false,
30
+ "enable_checkpointing": true,
31
+ "enable_model_summary": true,
32
+ "enable_progress_bar": true,
33
+ "eos_token_id": 2,
34
+ "fast_dev_run": false,
35
+ "gpus": null,
36
+ "gradient_clip_algorithm": null,
37
+ "gradient_clip_val": null,
38
+ "hidden_act": "gelu",
39
+ "hidden_dropout_prob": 0.1,
40
+ "hidden_size": 768,
41
+ "id2label": {
42
+ "0": "LABEL_0",
43
+ "1": "LABEL_1",
44
+ "2": "LABEL_2",
45
+ "3": "LABEL_3"
46
+ },
47
+ "inference_mode": true,
48
+ "initializer_range": 0.02,
49
+ "intermediate_size": 3072,
50
+ "ipus": null,
51
+ "label2id": {
52
+ "LABEL_0": 0,
53
+ "LABEL_1": 1,
54
+ "LABEL_2": 2,
55
+ "LABEL_3": 3
56
+ },
57
+ "layer_norm_eps": 1e-05,
58
+ "learning_rate": 1e-05,
59
+ "left_z_only": false,
60
+ "limit_predict_batches": null,
61
+ "limit_test_batches": null,
62
+ "limit_train_batches": null,
63
+ "limit_val_batches": null,
64
+ "log_class_acc": false,
65
+ "log_every_n_steps": 50,
66
+ "logger": true,
67
+ "lr_scheduler": false,
68
+ "max_epochs": 10,
69
+ "max_length": 128,
70
+ "max_position_embeddings": 514,
71
+ "max_samples": -1,
72
+ "max_steps": -1,
73
+ "max_time": null,
74
+ "min_epochs": null,
75
+ "min_steps": null,
76
+ "model_type": "roberta",
77
+ "move_metrics_to_cpu": false,
78
+ "multiple_trainloader_mode": "max_size_cycle",
79
+ "name": "classifier",
80
+ "no_context_pos": false,
81
+ "num_attention_heads": 12,
82
+ "num_hidden_layers": 12,
83
+ "num_nodes": 1,
84
+ "num_processes": null,
85
+ "num_sanity_val_steps": 2,
86
+ "overfit_batches": 0.0,
87
+ "pad_token_id": 1,
88
+ "plugins": null,
89
+ "position_embedding_type": "absolute",
90
+ "precision": 32,
91
+ "profiler": null,
92
+ "project": "planning_models",
93
+ "reading_lvl": null,
94
+ "regression": false,
95
+ "reload_dataloaders_every_n_epochs": 0,
96
+ "replace_sampler_ddp": true,
97
+ "resume_from_checkpoint": null,
98
+ "save_dir": null,
99
+ "simple_context_dir": null,
100
+ "simple_context_doc_id": null,
101
+ "src_lvl": null,
102
+ "strategy": null,
103
+ "sync_batchnorm": false,
104
+ "torch_dtype": "float32",
105
+ "tpu_cores": null,
106
+ "track_grad_norm": -1,
107
+ "train_check_interval": 0.2,
108
+ "train_file": "data/wikiauto_sents_train.csv",
109
+ "train_split": 0.9,
110
+ "train_workers": 8,
111
+ "transformers_version": "4.29.1",
112
+ "type_vocab_size": 1,
113
+ "upsample_classes": false,
114
+ "use_cache": true,
115
+ "val_check_interval": null,
116
+ "val_file": "data/wikiauto_sents_valid.csv",
117
+ "val_split": 0.05,
118
+ "val_workers": 8,
119
+ "vocab_size": 50274,
120
+ "wandb_id": null,
121
+ "x_col": "complex",
122
+ "y_col": "label"
123
+ }
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:965766d87d6567436b4464f9713b470d8a90de3c08070966f1ff02b8a0435f53
3
+ size 498695861