File size: 1,560 Bytes
cade5bb
2a24b3a
 
 
 
 
 
210d6ad
2a24b3a
 
cade5bb
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
210d6ad
 
 
 
 
 
2a24b3a
210d6ad
 
 
 
 
 
 
 
 
2a24b3a
210d6ad
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
---
license: cc-by-4.0
task_categories:
- automatic-speech-recognition
- text-to-speech
language:
- vi
pretty_name: VLSP 2020 - VinAI - ASR challenge dataset
size_categories:
- 10K<n<100K
dataset_info:
  features:
  - name: audio
    dtype: audio
  - name: transcription
    dtype: string
  splits:
  - name: train
    num_bytes: 17159347574.893
    num_examples: 56427
  download_size: 11649243045
  dataset_size: 17159347574.893
configs:
- config_name: default
  data_files:
  - split: train
    path: data/train-*
---

# unofficial mirror of VLSP 2020 - VinAI - ASR challenge dataset

official announcement:
- tiếng việt: https://institute.vinbigdata.org/events/vinbigdata-chia-se-100-gio-du-lieu-tieng-noi-cho-cong-dong/
- in eglish: https://institute.vinbigdata.org/en/events/vinbigdata-shares-100-hour-data-for-the-community/
- VLSP 2020 workshop: https://vlsp.org.vn/vlsp2020

official download: https://drive.google.com/file/d/1vUSxdORDxk-ePUt-bUVDahpoXiqKchMx/view?usp=sharing

contact: [email protected]

100h, 56.4k samples, accuracy 96%

pre-process: merge all transcript text files into 1, remove token `<unk>`

need to do: check misspelling, restore foreign words phonetised to vietnamese

usage with HuggingFace:
```python
# pip install -q "datasets[audio]"
from datasets import load_dataset
from torch.utils.data import DataLoader

dataset = load_dataset("doof-ferb/vlsp2020_vinai_100h", split="train", streaming=True)
dataset.set_format(type="torch", columns=["audio", "transcription"])
dataloader = DataLoader(dataset, batch_size=4)
```