Datasets:
You need to agree to share your contact information to access this dataset
This repository is publicly accessible, but you have to accept the conditions to access its files and content.
Please enter your affiliation and agree to the following terms to use the dataset.
Log in or Sign Up to review the conditions and access this dataset content.
SpeechBSD Dataset
This is an extension of the BSD corpus, a Japanese--English dialogue translation corpus, with audio files and speaker attribute information. Although the primary intended usage is for speech-to-text translation, it can be viewed as text/speech Japanese/English/cross-language dialogue corpus and can be used for various tasks.
Dataset Statistics
Train | Dev. | Test | |
---|---|---|---|
Scenarios | 670 | 69 | 69 |
Sentences | 20,000 | 2,051 | 2,120 |
En audio (h) | 20.1 | 2.1 | 2.1 |
Ja audio (h) | 25.3 | 2.7 | 2.7 |
En audio gender (male % / female %) | 47.2 / 52.8 | 50.1 / 49.9 | 44.4 / 55.6 |
Ja audio gender (male % / female %) | 68.0 / 32.0 | 62.3 / 37.7 | 69.0 / 31.0 |
Data Structure
We also provide the dataset at the GitHub repository. The data structure there is similar to the original BSD corpus. Here, the structure is changed to be represented in the JSONL format where one instance contains one audio file.
Data Instances
There are two types of instances, one that contains English wav file, and one that contains Japanese wav file.
A typical instance that contains an English wav file looks like this:
{
'audio': {'path': '/path/to/speech-bsd-hf/data/dev/190315_E001_17_spk0_no10_en.wav', 'array': array([ 0.00000000e+00, 0.00000000e+00, 0.00000000e+00, ...,
-6.10351562e-05, -1.83105469e-04, 6.10351562e-05]), 'sampling_rate': 16000},
'no': 10,
'en_speaker': 'Mr. Ben Sherman',
'en_sentence': 'You can also check out major institutions like banks, accounting companies, and market research companies.',
'en_spkid': '190315_E001_17_spk0_en',
'en_wav': '190315_E001_17_spk0_no10_en.wav',
'en_spk_gender': 'M',
'en_spk_state': 'CA',
'scenario_id': '190315_E001_17',
'scenario_tag': 'training',
'scenario_title': 'Training: How to do research',
'scenario_original_language': 'en',
'ja_speaker': None,
'ja_sentence': None,
'ja_spkid': None,
'ja_wav': None,
'ja_spk_gender': None,
'ja_spk_prefecture': None
}
In the corresponding instance that contains a Japanese wav file, en_speaker
, en_sentence
, en_spkid
, en_wav
, en_spk_gender
, and en_spk_state
are None
and corresponding Japanese ones are filled instead.
Data Fields
Each utterance is a part of a scenario.
The scenario information is shown with scenario_id
, scenario_tag
, scenario_title
, and scenario_original_language
, which corresponds respectively to id
, tag
, title
, and original_language
in the original BSD corpus.
The utterance-specific fields are the following:
no
,ja_speaker
,en_speaker
,ja_sentence
,en_sentence
are identical to the ones of the BSD corpus.ja_spkid
anden_spkid
show speaker IDs consistent throughout the conversation.ja_wav
anden_wav
show the wavfile names.ja_spk_gender
anden_spk_gender
show the gender of the speaker of the corresponding wav file (either "M" or "F").ja_spk_prefecture
anden_spk_state
show the speaker's region of origin.- As any other huggingface audio datasets,
audio
contains full path to the audio file in your environment, audio array, and sampling rate (which is 16000). You will needlibrosa
andsoundfile
packages to have access to these fields.
Here are some additional notes:
Speakers are different if speaker ID is different. For example, if a conversation is spoken by two speakers taking turns, there would be 4 speakers (2 Japanese speakers and 2 English speakers). However, it's possible that speakers with different speaker ID is actually spoken by the same person because of the way audio is collected.
Gender information of audio does not necessarily match with the one inferrable from text. For example, even if the
en_speaker
is "Mr. Sam Lee", the audio may contain female voice. This is because no explicit gender information is given in the original BSD corpus.Japanese speech is collected from Japanese speakers who are from Japan.
ja_spk_prefecture
is one of the 47 prefectures or "不明" (unknown). Japanese prefectures have four different ending characters, "県", "府", "都", and "道".- Prefectures that ends with "県" or "府" does not contain those characters (e.g., "神奈川", "京都").
- Tokyo is "東京" without "都".
- Hokkaido is "北海道".
English speech is collected from English speakers who are from the US.
en_spk
is one of the 50 states, written in postal abbreviation.
Citation
If you find the dataset useful, please cite our ACL 2023 Findings paper: Towards Speech Dialogue Translation Mediating Speakers of Different Languages.
@inproceedings{shimizu-etal-2023-towards,
title = "Towards Speech Dialogue Translation Mediating Speakers of Different Languages",
author = "Shimizu, Shuichiro and
Chu, Chenhui and
Li, Sheng and
Kurohashi, Sadao",
booktitle = "Findings of the Association for Computational Linguistics: ACL 2023",
month = jul,
year = "2023",
address = "Toronto, Canada",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2023.findings-acl.72",
pages = "1122--1134",
abstract = "We present a new task, speech dialogue translation mediating speakers of different languages. We construct the SpeechBSD dataset for the task and conduct baseline experiments. Furthermore, we consider context to be an important aspect that needs to be addressed in this task and propose two ways of utilizing context, namely monolingual context and bilingual context. We conduct cascaded speech translation experiments using Whisper and mBART, and show that bilingual context performs better in our settings.",
}
License
This dataset is licensed under CC-BY-NC-SA 4.0.
- Downloads last month
- 34