Machine Learning Based Speech (or Music) Separation
- Master Thesis
- Announcement date
- 01 Oct 2019
- Research Areas
Assume a single-channel multiple (two) speaker recording. Speech separation for such tasks can be formulated as classification or regression problem in the time-frequency domain. Recently, we used deep neural networks to accomplish this task. One task of this thesis is to extend the system by using recently developed neural network architectures such as DenseNets. The system is evaluated on available data using commonly used performance measures such as SIR, SDR, SAR, or PESQ. Similar approaches work for music signals.
- Extend the available prototype system.
- Test the system on available data set.
- Implement these models in Tensorflow
- Empirical Verification of these algorithms
The candidate should be interested in machine learning, speech processing, and neural networks. Excellent programming skills in C++, python etc. are required. Interested candidates are encouraged to ask for further information. Additionally, the supervision of own projects in one of the above mention fields is possible.
Franz Pernkopf (firstname.lastname@example.org or 0316/873 4436) and Pejman Mowlaee