vitouphy commited on
Commit
5ec17aa
1 Parent(s): 0cd0772

add readme file

Browse files
.ipynb_checkpoints/added_tokens-checkpoint.json DELETED
@@ -1 +0,0 @@
1
- {"<s>": 42, "</s>": 43}
 
 
.ipynb_checkpoints/config-checkpoint.json DELETED
@@ -1,108 +0,0 @@
1
- {
2
- "_name_or_path": "facebook/wav2vec2-xls-r-300m",
3
- "activation_dropout": 0.0,
4
- "adapter_kernel_size": 3,
5
- "adapter_stride": 2,
6
- "add_adapter": false,
7
- "apply_spec_augment": true,
8
- "architectures": [
9
- "Wav2Vec2ForCTC"
10
- ],
11
- "attention_dropout": 0.1,
12
- "bos_token_id": 1,
13
- "classifier_proj_size": 256,
14
- "codevector_dim": 768,
15
- "contrastive_logits_temperature": 0.1,
16
- "conv_bias": true,
17
- "conv_dim": [
18
- 512,
19
- 512,
20
- 512,
21
- 512,
22
- 512,
23
- 512,
24
- 512
25
- ],
26
- "conv_kernel": [
27
- 10,
28
- 3,
29
- 3,
30
- 3,
31
- 3,
32
- 2,
33
- 2
34
- ],
35
- "conv_stride": [
36
- 5,
37
- 2,
38
- 2,
39
- 2,
40
- 2,
41
- 2,
42
- 2
43
- ],
44
- "ctc_loss_reduction": "mean",
45
- "ctc_zero_infinity": false,
46
- "diversity_loss_weight": 0.1,
47
- "do_stable_layer_norm": true,
48
- "eos_token_id": 2,
49
- "feat_extract_activation": "gelu",
50
- "feat_extract_dropout": 0.0,
51
- "feat_extract_norm": "layer",
52
- "feat_proj_dropout": 0.0,
53
- "feat_quantizer_dropout": 0.0,
54
- "final_dropout": 0.0,
55
- "gradient_checkpointing": false,
56
- "hidden_act": "gelu",
57
- "hidden_dropout": 0.1,
58
- "hidden_size": 1024,
59
- "initializer_range": 0.02,
60
- "intermediate_size": 4096,
61
- "layer_norm_eps": 1e-05,
62
- "layerdrop": 0.0,
63
- "mask_feature_length": 64,
64
- "mask_feature_min_masks": 0,
65
- "mask_feature_prob": 0.25,
66
- "mask_time_length": 10,
67
- "mask_time_min_masks": 2,
68
- "mask_time_prob": 0.75,
69
- "model_type": "wav2vec2",
70
- "num_adapter_layers": 3,
71
- "num_attention_heads": 16,
72
- "num_codevector_groups": 2,
73
- "num_codevectors_per_group": 320,
74
- "num_conv_pos_embedding_groups": 16,
75
- "num_conv_pos_embeddings": 128,
76
- "num_feat_extract_layers": 7,
77
- "num_hidden_layers": 24,
78
- "num_negatives": 100,
79
- "output_hidden_size": 1024,
80
- "pad_token_id": 41,
81
- "proj_codevector_dim": 768,
82
- "tdnn_dilation": [
83
- 1,
84
- 2,
85
- 3,
86
- 1,
87
- 1
88
- ],
89
- "tdnn_dim": [
90
- 512,
91
- 512,
92
- 512,
93
- 512,
94
- 1500
95
- ],
96
- "tdnn_kernel": [
97
- 5,
98
- 3,
99
- 3,
100
- 1,
101
- 1
102
- ],
103
- "torch_dtype": "float32",
104
- "transformers_version": "4.17.0.dev0",
105
- "use_weighted_layer_sum": false,
106
- "vocab_size": 44,
107
- "xvector_output_dim": 512
108
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
.ipynb_checkpoints/preprocessor_config-checkpoint.json DELETED
@@ -1,9 +0,0 @@
1
- {
2
- "do_normalize": true,
3
- "feature_extractor_type": "Wav2Vec2FeatureExtractor",
4
- "feature_size": 1,
5
- "padding_side": "right",
6
- "padding_value": 0.0,
7
- "return_attention_mask": true,
8
- "sampling_rate": 16000
9
- }
 
 
 
 
 
 
 
 
 
 
.ipynb_checkpoints/special_tokens_map-checkpoint.json DELETED
@@ -1 +0,0 @@
1
- {"bos_token": "<s>", "eos_token": "</s>", "unk_token": "[UNK]", "pad_token": "[PAD]", "additional_special_tokens": [{"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}]}
 
 
.ipynb_checkpoints/tokenizer_config-checkpoint.json DELETED
@@ -1 +0,0 @@
1
- {"unk_token": "[UNK]", "bos_token": "<s>", "eos_token": "</s>", "pad_token": "[PAD]", "do_lower_case": false, "word_delimiter_token": "|", "special_tokens_map_file": null, "tokenizer_file": null, "name_or_path": "/workspace/container_0/model/", "tokenizer_class": "Wav2Vec2CTCTokenizer"}
 
 
.ipynb_checkpoints/vocab-checkpoint.json DELETED
@@ -1 +0,0 @@
1
- {"aa": 1, "ae": 2, "ah": 3, "aw": 4, "ay": 5, "b": 6, "ch": 7, "d": 8, "dh": 9, "dx": 10, "eh": 11, "er": 12, "ey": 13, "f": 14, "g": 15, "h#": 16, "hh": 17, "ih": 18, "iy": 19, "jh": 20, "k": 21, "l": 22, "m": 23, "n": 24, "ng": 25, "ow": 26, "oy": 27, "p": 28, "r": 29, "s": 30, "sh": 31, "t": 32, "th": 33, "uh": 34, "uw": 35, "v": 36, "w": 37, "y": 38, "z": 39, "|": 0, "[UNK]": 40, "[PAD]": 41}
 
 
README.md ADDED
@@ -0,0 +1,57 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ license: apache-2.0
5
+ tags:
6
+ - automatic-speech-recognition
7
+ - pytorch
8
+ - transformers
9
+ - en
10
+ - generated_from_trainer
11
+ model-index:
12
+ - name: model
13
+ results: []
14
+ ---
15
+
16
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
17
+ should probably proofread and complete it, then remove this comment. -->
18
+
19
+ # model
20
+
21
+ This model is a fine-tuned version of [facebook/wav2vec2-xls-r-300m](https://huggingface.co/facebook/wav2vec2-xls-r-300m) on the Timit dataset.
22
+
23
+ ## Model description
24
+
25
+ More information needed
26
+
27
+ ## Intended uses & limitations
28
+
29
+ More information needed
30
+
31
+ ## Training and evaluation data
32
+
33
+ More information needed
34
+
35
+ ## Training procedure
36
+
37
+ ### Training hyperparameters
38
+
39
+ The following hyperparameters were used during training:
40
+ - learning_rate: 3e-05
41
+ - train_batch_size: 8
42
+ - eval_batch_size: 8
43
+ - seed: 42
44
+ - gradient_accumulation_steps: 4
45
+ - total_train_batch_size: 32
46
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
47
+ - lr_scheduler_type: linear
48
+ - lr_scheduler_warmup_steps: 2000
49
+ - training_steps: 10000
50
+ - mixed_precision_training: Native AMP
51
+
52
+ ### Framework versions
53
+
54
+ - Transformers 4.17.0.dev0
55
+ - Pytorch 1.10.2+cu102
56
+ - Datasets 1.18.2.dev0
57
+ - Tokenizers 0.11.0
vocab.json CHANGED
@@ -1 +1 @@
1
- {"\u0251": 1, "\u00e6": 2, "\u0259": 3, "a\u028a": 4, "a\u026a": 5, "b": 6, "\u02a7": 7, "d": 8, "\u00f0": 9, "dx": 10, "\u025b": 11, "\u0259r": 12, "e\u026a": 13, "f": 14, "g": 15, "": 16, "h": 17, "\u026a": 18, "i": 19, "\u02a4": 20, "k": 21, "l": 22, "m": 23, "n": 24, "\u014b": 25, "o\u028a": 26, "\u0254\u026a": 27, "p": 28, "r": 29, "s": 30, "\u0283": 31, "t": 32, "\u03b8": 33, "\u028a": 34, "u": 35, "v": 36, "w": 37, "j": 38, "z": 39, "|": 0, "[UNK]": 40, "[PAD]": 41}
 
1
+ {"\u0251": 1, "\u00e6": 2, "\u0259": 3, "a\u028a": 4, "a\u026a": 5, "b": 6, "\u02a7": 7, "d": 8, "\u00f0": 9, "dx": 10, "\u025b": 11, "\u0259r": 12, "e\u026a": 13, "f": 14, "g": 15, " ": 16, "h": 17, "\u026a": 18, "i": 19, "\u02a4": 20, "k": 21, "l": 22, "m": 23, "n": 24, "\u014b": 25, "o\u028a": 26, "\u0254\u026a": 27, "p": 28, "r": 29, "s": 30, "\u0283": 31, "t": 32, "\u03b8": 33, "\u028a": 34, "u": 35, "v": 36, "w": 37, "j": 38, "z": 39, "|": 0, "[UNK]": 40, "[PAD]": 41}