Signal Processing and Speech Communication Laboratory
hometheses & projects › Single-channel source separation for conversational speech in realistic environments

Single-channel source separation for conversational speech in realistic environments

Status
Finished
Type
Bachelor Project
Announcement date
18 Oct 2023
Student
Emil Berger
Mentors
Research Areas

Short Description

Recent progress in Single Channel Source Separation (SCSS) using deep neural networks led to impressive performance gains while also increasing the model sizes, requiring tremendous data resources. This demand is covered by artificially composed speech and noise mixtures that do not capture real-life characteristics of conversations taking place in noisy environments. This paper introduces a new dataset containing task-oriented dialogues spoken in a realistic environment and presents experimental results for two SCSS architectures - the Conv-TasNet and the transformer-based MossFormer. Overall, we observe a severe drop in performance of up to 4.3dB (SI-SDR improvement) for the 8kHz variant of the Conv-TasNet. For speaker pairs of homogeneous sex, the difference is even higher of up to 6dB. Only the model using 16kHz sample rate performs on a comparable level for speaker pairs of mixed sex. Our findings illustrate the need of using realistic data for both, training and evaluating. The results of this bachelor project contributed to the following publication: https://easychair.org/publications/preprint/Cdb4