Deep neural networks for auditory scene analysis
Deep Learning (DL) has produced state-of-art results in many audio tasks. We are currently using DL approaches in neuromorphic auditory scene analysis. In this project, the student will implement a recurrent neural network model for estimating the number and identity of speakers in a conversation scenario. In particular, we focus on networks with a small memory footprint and low computational cost so that they can be implement on embedded systems. The use of the silicon DAS cochlea sensor and DVS vision sensor can also be incorporated as an additional sensor for this system. The project is suitable for a Semester or Master Thesis project.
Requirements
Knowledge of Python, Matlab, some knowledge of deep networks, interest in hardware.
Contact
For more details, contact Shih-Chii Liu (shih@ini.uzh.ch) and Enea Ceolini (eceoli@ini.uzh.ch) and go to http://sensors.ini.uzh.ch for more information.