serbog commited on
Commit
9489013
1 Parent(s): 6ead039

Training in progress, epoch 1

Browse files
config.json ADDED
@@ -0,0 +1,48 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "distilbert-base-multilingual-cased",
3
+ "activation": "gelu",
4
+ "architectures": [
5
+ "DistilBertForSequenceClassification"
6
+ ],
7
+ "attention_dropout": 0.1,
8
+ "dim": 768,
9
+ "dropout": 0.1,
10
+ "hidden_dim": 3072,
11
+ "id2label": {
12
+ "0": "C1",
13
+ "1": "C2",
14
+ "2": "C3",
15
+ "3": "C4",
16
+ "4": "C5",
17
+ "5": "C6",
18
+ "6": "C7",
19
+ "7": "C8",
20
+ "8": "C9"
21
+ },
22
+ "initializer_range": 0.02,
23
+ "label2id": {
24
+ "C1": 0,
25
+ "C2": 1,
26
+ "C3": 2,
27
+ "C4": 3,
28
+ "C5": 4,
29
+ "C6": 5,
30
+ "C7": 6,
31
+ "C8": 7,
32
+ "C9": 8
33
+ },
34
+ "max_position_embeddings": 512,
35
+ "model_type": "distilbert",
36
+ "n_heads": 12,
37
+ "n_layers": 6,
38
+ "output_past": true,
39
+ "pad_token_id": 0,
40
+ "problem_type": "multi_label_classification",
41
+ "qa_dropout": 0.1,
42
+ "seq_classif_dropout": 0.2,
43
+ "sinusoidal_pos_embds": false,
44
+ "tie_weights_": true,
45
+ "torch_dtype": "float32",
46
+ "transformers_version": "4.33.1",
47
+ "vocab_size": 119547
48
+ }
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5279c4ba7f35acd33597b4b290ab5535b547010ad962c5ad9103c6fe8d684b28
3
+ size 541361837
special_tokens_map.json CHANGED
@@ -1,15 +1,7 @@
1
  {
2
- "bos_token": "<s>",
3
- "cls_token": "<s>",
4
- "eos_token": "</s>",
5
- "mask_token": {
6
- "content": "<mask>",
7
- "lstrip": true,
8
- "normalized": true,
9
- "rstrip": false,
10
- "single_word": false
11
- },
12
- "pad_token": "<pad>",
13
- "sep_token": "</s>",
14
- "unk_token": "<unk>"
15
  }
 
1
  {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
 
 
 
 
 
 
 
 
7
  }
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:592ccc69de22c051c7971596a69df9932829344473990c7bef3b3faf5705d488
3
- size 17082728
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bf1b59b7b11c95f194f51708d918eea378e09d05f84c0e1656dc5180e8117088
3
+ size 2919362
tokenizer_config.json CHANGED
@@ -1,20 +1,13 @@
1
  {
2
- "bos_token": "<s>",
3
  "clean_up_tokenization_spaces": true,
4
- "cls_token": "<s>",
5
- "eos_token": "</s>",
6
- "mask_token": {
7
- "__type": "AddedToken",
8
- "content": "<mask>",
9
- "lstrip": true,
10
- "normalized": true,
11
- "rstrip": false,
12
- "single_word": false
13
- },
14
- "model_max_length": 1000000000000000019884624838656,
15
- "pad_token": "<pad>",
16
- "sep_token": "</s>",
17
- "sp_model_kwargs": {},
18
- "tokenizer_class": "XLMRobertaTokenizer",
19
- "unk_token": "<unk>"
20
  }
 
1
  {
 
2
  "clean_up_tokenization_spaces": true,
3
+ "cls_token": "[CLS]",
4
+ "do_lower_case": false,
5
+ "mask_token": "[MASK]",
6
+ "model_max_length": 512,
7
+ "pad_token": "[PAD]",
8
+ "sep_token": "[SEP]",
9
+ "strip_accents": null,
10
+ "tokenize_chinese_chars": true,
11
+ "tokenizer_class": "DistilBertTokenizer",
12
+ "unk_token": "[UNK]"
 
 
 
 
 
 
13
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ccf1ef2b8e30ec25f70150630bb43738c7ab901bdd28685498430d91c26d6978
3
- size 4091
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:90b2311cf20a68a9c8a3cfe0046953a0d77d6d85d05cc8abe6a3e5169bdc013a
3
+ size 4155
vocab.txt ADDED
The diff for this file is too large to render. See raw diff