Data

The CHiME-5 data consists of 20 parties each recorded in a different home.

To refer to these data in a publication, please cite:

    Jon Barker, Shinji Watanabe, Emmanuel Vincent, and Jan Trmal
    The fifth `CHiME’ Speech Separation and Recognition Challenge: Dataset, task and baselines
    Interspeech, 2018.

The data have been split into training, development test, and evaluation test sets as follows.

Dataset Parties Speakers Hours Utterances
Train 16 32 40:33 79,980
Dev 2 8 4:27 7,440
Eval 2 8 5:12 11,028

The audio data and the transcriptions follow this directory structure:

CHiME5/
├── audio
│  ├── dev
│  ├── eval
│  └── train
└── transcriptions
   ├── dev
   ├── eval
   └── train

Each audio/transcription directory has subdirectories for training, development, and evaluation sets. The evaluation data will be released later.

Audio

All audio data are distributed as WAV files with a sampling rate of 16 kHz. Each session consists of the recordings made by the binaural microphones worn by each participant (4 participants per session), and by 6 microphone arrays with 4 microphones each. Therefore, the total number of microphones per session is 32 (2 x 4 + 4 x 6). These WAV files are named as follows:

  • Binaural microphones
    <session ID>_<speaker ID>.wav , e.g., S02_P05.wav
  • Array microphone
    <session ID>_<array ID>.CH<channel ID>.wav , e.g., S02_U05.CH1.wav

Note:

  • The recordings made by the binaural microphones are stereo WAV files which include both left and right channels, while the recordings made by array microphones are decomposed into one mono WAV file per channel.
  • We will not distribute the binaural microphone recordings for the evaluation set during the challenge.

The following tables provide more detailed statistics and notes about each session:

Training sessions

Session IDParticipants (Bold=Male)Duration#UttsNotes
S03P09, P10, P11, P122:11:224,090P11 dropped from min ~15 to ~30
S04P09, P10, P11, P122:29:365,563
S05P13, p14, p15, P162:31:444,939U03 missing (crashed)
S06P13, p14, p15, P162:30:065,097
S07p17, P18, p19, P202:26:533,656
S17p17, P18, p19, P202:32:165,892
S08P21, P22, P23, P242:31:356,175
S16P21, P22, P23, P242:32:195,004
S12P33, P34, P35, p362:29:243,300Last 15 minutes of U05 missing
(Kinect was accidentally turned off)
S13P33, P34, P35, p362:30:114,193
S19p49, P50, P51, p522:32:384,292P52 mic unreliable
S20p49, P50, P51, p522:18:045,365
S18p41, P42, p43, p442:42:234,907
S22p41, P42, p43, p442:35:444,758U03 missing
S23p53, P54, P55, p562:58:437,054Neighbour interrupts
S24p53, P54, P55, p562:37:095,695P54 mic unreliable,
P53 disconnects for bathroom

Development sessions

Session IDParticipants (Bold=Male)Duration#UttsNotes
S02p05, P06, P07, p082:28:243,822
S09p25, p26, p27, p281:59:213,618U05 missing

Evaluation sessions

Session IDParticipants (Bold=Male)Duration#UttsNotes
S01p01, p02, P03, p042:39:045,797No registration tone
S21P45, P46, P47, p482:33:205,231

Transcriptions

The transcriptions are provided in JSON format for each session as <session ID>.json. The JSON file includes the following pieces of information for each utterance:

  • Session ID ("session_id")
  • Location ("kitchen", "dining", or "living")
  • Speaker ID ("speaker")
  • Transcription ("words")
  • Start time ("start_time")
    • For the binaural microphone recording of that speaker ("original")
    • For all array recordings ("U01", etc.)
    • For all binaural microphone recordings ("P01", etc.)
  • End time ("end_time")
  • Reference microphone array ID ("ref")

The following is an example annotation of one utterance in a JSON file:


    {
        "end_time": {
            "original": "0:00:43.82",
            "U01": "0:00:43.85",
            "U02": "0:00:43.84",
            "U03": "0:00:43.83",
            "U04": "0:00:43.83",
            "U05": "0:00:43.82",
            "U06": "0:00:43.82",
            "P05": "0:00:43.82",
            "P06": "0:00:43.82",
            "P07": "0:00:43.82",
            "P08": "0:00:43.82"
        },
        "start_time": {
            "original": "0:00:40.60",
            "U01": "0:00:40.63",
            "U02": "0:00:40.62",
            "U03": "0:00:40.61",
            "U04": "0:00:40.61",
            "U05": "0:00:40.60",
            "U06": "0:00:40.60",
            "P05": "0:00:40.60",
            "P06": "0:00:40.60",
            "P07": "0:00:40.60",
            "P08": "0:00:40.60"
        },
        "words": "[laughs] It's the blue, I think. I think.",
        "speaker": "P05",
        "ref": "U02",
        "location": "kitchen",
        "session_id": "S02"
    },

Note

  • "location" and "ref" are provided for development and evaluation data only (not for training), due to the fact that you are allowed to use all array channels and binaural microphones for training.
  • Transcriptions include the following tags:
    • [noise] denotes any non-language noise made by the speaker (ex: grunts, coughing, loud chewing, loud lip smacking etc.)
    • [inaudible] denotes speech that is audible but not clear enough to be transcribed.
    • [laughs] denotes an instance where the participant laughs.
    • [redacted] are parts of the signals that have been zeroed out for privacy reasons
  • For redacted utterances, "speaker" information is not provided.
  • The start and end times have been manually annotated for the binaural microphone pair worn by the speaker ("original"). The other microphones are not aligned. Therefore, we also provide the start and end times for each array and each binaural microphone pair by shifting the original times by the appropriate delays. These delays have been estimated by means of simple cross-correlation between the signals and may not be fully accurate. This approach should be considered as a baseline to be improved as part of the challenge.
  • The reference array microphone array ID ("ref") is fixed for a given session and location.

All data is available under licence via the download page.