wav2vec2-base-turkish-artificial-cv / tokenizer_config.json
cahya's picture
add language model
24eecc3
raw
history blame contribute delete
474 Bytes
{"unk_token": "[UNK]", "bos_token": null, "eos_token": null, "pad_token": "[PAD]", "do_lower_case": false, "word_delimiter_token": "|", "special_tokens_map_file": "/home/cahya/.cache/huggingface/transformers/2916f85ca1c8db02f6dffab6c5680b457e26471b99cac401f31d9d01fd7af256.a21d51735cf8667bcd610f057e88548d5d6a381401f6b4501a8bc6c1a9dc8498", "tokenizer_file": null, "name_or_path": "./", "processor_class": "Wav2Vec2ProcessorWithLM", "tokenizer_class": "Wav2Vec2CTCTokenizer"}