Two-channel mixtures of speech and real-world background noise
We propose to repeat the Two-channel mixtures of speech and real-world background noise without Chime corpus because the reference speech data has been already provided in the second ChiME challenge.
Introduction
This task aims to evaluate denoising and DOA estimation techniques by the SiSEC 2010 noisy speech dataset.
Description of the dataset
We consider two-channel mixtures of one speech source and real-world background noise sampled at 16 kHz.
These data are part of the SiSEC 2010 noisy speech dataset. Background noise signals were recorded via a pair of omnidirectional microphones spaced by 8.6 cm in six different public environments:
Su1
: subway car movingSu2
: subway car standing at stationCa1
: cafeteria 1Ca2
: cafeteria 2 (another cafeteria than Ca1)Sq1
: square 1Sq2
: square 2 (another square than Sq1)
and in two different positions within each environment:
Ce
: centerCo
: corner (except inSu1
andSu2
)
Two recordings identified by a letter (A or B) were made in each case. Mixtures were then generated by adding a speech signal to the background noise signal. For the reverberant environments Su
and Ca
, the speech signals were recorded in an office room using the same microphone pair. For the outdoor environment Sq
, the speech signals were mixed anechoically through simulation. The distance between the sound source and the array centroid was 1.0 m for female speech and 0.8 m for male speech. The direction of arrival (DOA) of the speech source was different in each mixture and the signal-to-noise ratio (SNR) was drawn randomly between -17 and +12 dB.
Test data
Download the test set (13 MB)
The data consist of 20 stereo WAV audio files that can be imported in Matlab using the wavread command. These files are named test____mix.wav
, where
: noise environment (
Su1
,Su2
,Ca1
,Ca2
,Sq1
,Sq2
): recording condition (
Ce
,Co
): take (
A
,B
)
Development data
Download the development set (11 MB)
The data consists of 36 WAV audio files and 10 text files. These files are named as follows:
dev____src.wav
: single-channel speech signaldev____sim.wav
: two-channel spatial image of the speech sourcedev____noi.wav
: two-channel spatial image of the background noisedev____mix.wav
: two-channel mixture signaldev____DOA.txt
: DOA of the speech source (see the SiSEC 2010 wiki for the convention adopted to measure DOA)
where
: noise environment (
Su1
,Ca1
,Sq1
): recording condition (
Ce
,Co
): take (
A
,B
)
Since the source DOAs were measured geometrically in the Su1
and Ca1
environments, they might contain a measurement error up to a few degrees; on the contrary, there is no such error in the Sq
environment, because the spatial images of the speech source were simulated. The Co
condition of the Ca1
environment has take A
only.
Tasks and reference software
We propose the following 3 tasks:
- speaker DOA estimation: estimate the DOA of the speech source
- speech signal estimation: estimate the single-channel speech source
- speech and noise spatial image estimation: decompose the mixture signal into two two-channel signals corresponding to the speech source and the background noise
Participants are welcome to use some of the Matlab reference software below to build their own algorithms:
- stft_multi.m: multichannel STFT
- istft_multi.m: multichannel inverse STFT
- example_denoising.m: TDOA estimation by GCC-PHATmax, ML target and noise variance estimation under a diffuse noise model, and multichannel Wiener filtering
Submission
Each participant is asked to submit the results of his/her algorithm for task 2 and/or 3 over all or part of the mixtures in the development dataset and the test dataset. The results for task 1 may also be submitted if possible.
Each participant should make his results available online in the form of a tarball with the following file naming convention:
____src.wav
: single-channel speech signal____sim.wav
: two-channel spatial image of the speech source____noi.wav
: two-channel spatial image of the background noise____DOA.txt
: DOA of the speech source
where is
test
or dev
.
Each participant should then send an email to “onono (at) nii.ac.jp” and “zbynek.koldovsky (at) tul.cz” providing:
- contact information (name, affiliation)
- basic information about his/her algorithm, including its average running time (in seconds per test excerpt and per GHz of CPU) and a bibliographical reference if possible
- the URL of the tarball
The submitted audio files will be made available on a website under the terms of the Licensing section below.
Evaluation criteria
We propose to use the same evaluation criteria as in SiSEC 2010, except that the order of the estimated sources must be recovered.
The estimated speaker DOAs in task 1 will be evaluated in terms of absolute difference with the true DOAs.
The estimated speech signals in task 2 will be evaluated via the energy ratio criteria defined in the BSS_EVAL toolbox allowing arbitrary filtering between the estimated source and the true source.
The estimated speech and noise spatial image signals in task 3 will be evaluated via the energy ratio criteria introduced for the Stereo Audio Source Separation Evaluation Campaign and via the perceptually-motivated criteria in the PEASS toolkit.
Performance will be compared to that of ideal binary masking as a benchmark (i.e. binary masks providing maximum SDR), computed over a STFT or a cochleagram.
The above performance criteria and benchmarks are respectively implemented in
An example use is given in example_denoising.m.
Licensing issues
All files are distributed under the terms of the Creative Commons Attribution-Noncommercial-ShareAlike 3.0 license. The files to be submitted by participants will be made available on a website under the terms of the same license.
Public environment data were authored by Ngoc Q. K. Duong and Nobutaka Ito.
Task proposed by the Audio Committee
Recent Comments