speech31 commited on
Commit
586ac9f
1 Parent(s): f765844

upload model

Browse files
config.json ADDED
@@ -0,0 +1,99 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "microsoft/wavlm-large",
3
+ "activation_dropout": 0.0,
4
+ "adapter_kernel_size": 3,
5
+ "adapter_stride": 2,
6
+ "add_adapter": false,
7
+ "apply_spec_augment": true,
8
+ "architectures": [
9
+ "Wav2Vec2ForCTC"
10
+ ],
11
+ "attention_dropout": 0.0,
12
+ "bos_token_id": 1,
13
+ "classifier_proj_size": 256,
14
+ "codevector_dim": 768,
15
+ "contrastive_logits_temperature": 0.1,
16
+ "conv_bias": false,
17
+ "conv_dim": [
18
+ 512,
19
+ 512,
20
+ 512,
21
+ 512,
22
+ 512,
23
+ 512,
24
+ 512
25
+ ],
26
+ "conv_kernel": [
27
+ 10,
28
+ 3,
29
+ 3,
30
+ 3,
31
+ 3,
32
+ 2,
33
+ 2
34
+ ],
35
+ "conv_stride": [
36
+ 5,
37
+ 2,
38
+ 2,
39
+ 2,
40
+ 2,
41
+ 2,
42
+ 2
43
+ ],
44
+ "ctc_loss_reduction": "mean",
45
+ "ctc_zero_infinity": false,
46
+ "diversity_loss_weight": 0.1,
47
+ "do_stable_layer_norm": true,
48
+ "eos_token_id": 2,
49
+ "feat_extract_activation": "gelu",
50
+ "feat_extract_dropout": 0.0,
51
+ "feat_extract_norm": "layer",
52
+ "feat_proj_dropout": 0.0,
53
+ "feat_quantizer_dropout": 0.0,
54
+ "final_dropout": 0.0,
55
+ "gradient_checkpointing": false,
56
+ "hidden_act": "gelu",
57
+ "hidden_dropout": 0.0,
58
+ "hidden_size": 1024,
59
+ "initializer_range": 0.02,
60
+ "intermediate_size": 4096,
61
+ "layer_norm_eps": 1e-05,
62
+ "layerdrop": 0.0,
63
+ "mask_channel_length": 10,
64
+ "mask_channel_min_space": 1,
65
+ "mask_channel_other": 0.0,
66
+ "mask_channel_prob": 0.0,
67
+ "mask_channel_selection": "static",
68
+ "mask_feature_length": 10,
69
+ "mask_feature_min_masks": 0,
70
+ "mask_feature_prob": 0.0,
71
+ "mask_time_length": 10,
72
+ "mask_time_min_masks": 2,
73
+ "mask_time_min_space": 1,
74
+ "mask_time_other": 0.0,
75
+ "mask_time_prob": 0.05,
76
+ "mask_time_selection": "static",
77
+ "max_bucket_distance": 800,
78
+ "model_type": "wav2vec2",
79
+ "num_adapter_layers": 3,
80
+ "num_attention_heads": 16,
81
+ "num_buckets": 320,
82
+ "num_codevector_groups": 2,
83
+ "num_codevectors_per_group": 320,
84
+ "num_conv_pos_embedding_groups": 16,
85
+ "num_conv_pos_embeddings": 128,
86
+ "num_ctc_classes": 80,
87
+ "num_feat_extract_layers": 7,
88
+ "num_hidden_layers": 24,
89
+ "num_negatives": 100,
90
+ "output_hidden_size": 1024,
91
+ "pad_token_id": 38,
92
+ "proj_codevector_dim": 768,
93
+ "replace_prob": 0.5,
94
+ "tokenizer_class": "Wav2Vec2CTCTokenizer",
95
+ "torch_dtype": "float32",
96
+ "transformers_version": "4.11.3",
97
+ "use_weighted_layer_sum": false,
98
+ "vocab_size": 39
99
+ }
optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:91efd24a2005c4045520d38f151d2f08794bbf78537813f6eb37829195c38d7d
3
+ size 2490378705
preprocessor_config.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "do_normalize": true,
3
+ "feature_extractor_type": "Wav2Vec2FeatureExtractor",
4
+ "feature_size": 1,
5
+ "padding_side": "right",
6
+ "padding_value": 0.0,
7
+ "return_attention_mask": true,
8
+ "sampling_rate": 16000
9
+ }
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e1986049afcd821c8cb272686718253215d26316df10ce87fb33d1315050ac15
3
+ size 1262066879
rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:de7c509bd86577b8913d86be912c77a1095254b105dd4f670cdd2b78b3018d35
3
+ size 14503
scaler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:947b6a8270538e237b7ce8f531749b4e87461b5205448c391e115b5f0c596643
3
+ size 559
scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:941c0c422382fb4856d4f9560e18421fb4ea5d3c208554cecaed7b7c8275bae4
3
+ size 623
special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"bos_token": "<s>", "eos_token": "</s>", "unk_token": "[UNK]", "pad_token": "[PAD]"}
tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"unk_token": "[UNK]", "bos_token": "<s>", "eos_token": "</s>", "pad_token": "[PAD]", "do_lower_case": false, "word_delimiter_token": "|", "tokenizer_class": "Wav2Vec2CTCTokenizer"}
trainer_state.json ADDED
@@ -0,0 +1,94 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.5534710884094238,
3
+ "best_model_checkpoint": "wavlm-tamil-ipa/checkpoint-1000",
4
+ "epoch": 0.7136485280999108,
5
+ "global_step": 1000,
6
+ "is_hyper_param_search": false,
7
+ "is_local_process_zero": true,
8
+ "is_world_process_zero": true,
9
+ "log_history": [
10
+ {
11
+ "epoch": 0.07,
12
+ "learning_rate": 2.97e-05,
13
+ "loss": 6.4757,
14
+ "step": 100
15
+ },
16
+ {
17
+ "epoch": 0.14,
18
+ "learning_rate": 5.94e-05,
19
+ "loss": 2.7305,
20
+ "step": 200
21
+ },
22
+ {
23
+ "epoch": 0.21,
24
+ "learning_rate": 8.939999999999999e-05,
25
+ "loss": 2.5809,
26
+ "step": 300
27
+ },
28
+ {
29
+ "epoch": 0.29,
30
+ "learning_rate": 0.0001191,
31
+ "loss": 2.53,
32
+ "step": 400
33
+ },
34
+ {
35
+ "epoch": 0.36,
36
+ "learning_rate": 0.00014879999999999998,
37
+ "loss": 2.0147,
38
+ "step": 500
39
+ },
40
+ {
41
+ "epoch": 0.36,
42
+ "eval_cer": 0.6800635831193839,
43
+ "eval_loss": 1.4391984939575195,
44
+ "eval_runtime": 1401.2346,
45
+ "eval_samples_per_second": 8.599,
46
+ "eval_steps_per_second": 1.075,
47
+ "step": 500
48
+ },
49
+ {
50
+ "epoch": 0.43,
51
+ "learning_rate": 0.00017849999999999997,
52
+ "loss": 0.934,
53
+ "step": 600
54
+ },
55
+ {
56
+ "epoch": 0.5,
57
+ "learning_rate": 0.00020819999999999996,
58
+ "loss": 0.597,
59
+ "step": 700
60
+ },
61
+ {
62
+ "epoch": 0.57,
63
+ "learning_rate": 0.0002382,
64
+ "loss": 0.4904,
65
+ "step": 800
66
+ },
67
+ {
68
+ "epoch": 0.64,
69
+ "learning_rate": 0.0002676,
70
+ "loss": 0.4568,
71
+ "step": 900
72
+ },
73
+ {
74
+ "epoch": 0.71,
75
+ "learning_rate": 0.00029759999999999997,
76
+ "loss": 0.3961,
77
+ "step": 1000
78
+ },
79
+ {
80
+ "epoch": 0.71,
81
+ "eval_cer": 0.3502788029525032,
82
+ "eval_loss": 0.5534710884094238,
83
+ "eval_runtime": 1403.4751,
84
+ "eval_samples_per_second": 8.585,
85
+ "eval_steps_per_second": 1.074,
86
+ "step": 1000
87
+ }
88
+ ],
89
+ "max_steps": 70050,
90
+ "num_train_epochs": 50,
91
+ "total_flos": 6.366682745104583e+18,
92
+ "trial_name": null,
93
+ "trial_params": null
94
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:935d121a78eab9fc9e6e8032f0ee448b5799e8cfec09aaceb398163a8ac4331b
3
+ size 2863
vocab.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"ʂ": 1, "ɲ": 2, "r": 3, "u": 4, "a": 5, "ʈ": 6, "t͡ʃ": 7, "oː": 8, "m": 9, "p": 10, "aj": 11, "ʋ": 12, "uː": 13, "ɾ": 14, "ŋ": 15, "ɳ": 16, "ɻ": 17, "kʂ": 18, "t̪": 19, "l": 20, "e": 21, "aʋ": 22, "eː": 23, "h": 24, "aː": 25, "k": 26, "ɭ": 27, "்": 28, "i": 29, "n": 30, "o": 31, "iː": 32, "d͡ʒ": 33, "j": 34, "s": 35, "n̪": 36, "|": 0, "[UNK]": 37, "[PAD]": 38}