--- dataset_info: features: - name: audio dtype: audio: sampling_rate: 16000 - name: timestamps_start sequence: float64 - name: timestamps_end sequence: float64 - name: speakers sequence: string splits: - name: train num_bytes: 186825817 num_examples: 61 download_size: 172710554 dataset_size: 186825817 configs: - config_name: default data_files: - split: train path: data/train-* tags: - speaker diarization - speaker-segmentation - voice-activity-detection language: - fr license: mit --- # Dataset Card for the Simsamu dataset This repository contains recordings of simulated medical dispatch dialogs in the french language, annotated for diarization and transcription. It is published under the MIT license. These dialogs were recorded as part of the training of emergency medicine interns, which consisted in simulating a medical dispatch call where the interns took turns playing the caller and the regulating doctor. Each situation was decided randomly in advance, blind to who was playing the medical dispatcher (e.g., road accident, chest pain, burns, etc.). The affiliations between the caller and the patient (family, friend, colleague...) and the caller's communication mode is then randomly selected. The caller had to adapt his or her performance to the communication mode associated with the situation. Seven communication modes were defined: shy, procedural, angry, cooperative, frightened, impassive, incomprehensible. Regarding sound quality, the voice of the regulating doctor is directly picked up by a microphone, whereas the voice of the caller is transmitted through the phone network and re-emitted by a phone speaker before being picked up by the microphone. This leads to different acoustic characteristics between the caller's voice and the regulator's, the later one often being much clearer. This phenomena is also present in actual dispatch services recordings, where the regulator's voice is directly recorded in a quiet room whereas the caller is often calling from noisier environments and its voice is altered by the phone network compression. The dataset is composed of 61 audio recordings with a total duration of 3h 15 and an average duration per recording of 3 minutes 11 seconds Note: This dataset has been preprocessed using [diarizers](https://github.com/huggingface/diarizers). It makes the dataset compatible with diarizers to fine-tune [pyannote](https://huggingface.co/pyannote/segmentation-3.0) segmentation models. # Example Usage ``` from datasets import load_dataset ds = load_dataset("diarizers-community/simsamu") print(ds) ``` gives: ``` DatasetDict({ train: Dataset({ features: ['audio', 'timestamps_start', 'timestamps_end', 'speakers'], num_rows: 61 }) }) ``` # Dataset source - Repository: https://huggingface.co/datasets/medkit/simsamu - Preprocessed using [diarizers](https://github.com/kamilakesbi/diarizers/tree/main/datasets) # Contribution Thanks to [@kamilakesbi](https://huggingface.co/kamilakesbi) and [@sanchit-gandhi](https://huggingface.co/sanchit-gandhi) for adding this dataset.