Last Update:29.01.2026
Status: OPEN
Type: masters (short/long)
Contact: Zixiao Li (zixili@ethz.ch), Prof. Shih-Chii Liu (shih@ini.uzh.ch)
___________________________________________
Speech enhancement (SE) aims to improve the input speech quality by removing the ambient background noise from the audio. While deep learning-based algorithms produce impressive SE quality (check out two demos: RNNoise and DeepFilter), their deployment on resource-constrained edge devices is still challenging due to the high computational complexity and large size of the models. The goal of this project is to implement a complete end-to-end RNN FPGA accelerator for SE and deploy the platform for real-world demonstration.
Check out demonstrations of previous audio processing accelerators from Sensors Group here.
Prerequisite
Familiar with python, in the context of machine learning.
Familiar with Verilog or VHDL (taken VLSI I or equivalent course).
FPGA development experience is preferred, but not necessary.
Audio processing algorithms knowledge is preferred, but not necessary.
Background
[1] I. Fedorov, et al. "TinyLSTMs: Efficient Neural Speech Enhancement for Hearing Aids," Interspeech, 2020.
[2] C. Gao, et al. “EdgeDRNN: Recurrent Neural Network Accelerator for Edge Inference,” IEEE JETCAS, 2020.
Last Update:29.01.2026
Status: OPEN
Type: masters (short/long)
Contact: Zixiao Li (zixili@ethz.ch), Prof. Shih-Chii Liu (shih@ini.uzh.ch)
___________________________________________
Interested in audio prosthesis algorithms using audio and visual inputs? Or implementing these algorithms on embedded or augmented reality platforms?
Check out demonstrations of audio processing systems from Sensors Group here.
Prerequisite
Familiar with python, in the context of machine learning.
Audio processing algorithms knowledge is preferred, but not necessary.
Background
[1] I. Fedorov, et al. "TinyLSTMs: Efficient Neural Speech Enhancement for Hearing Aids," Interspeech, 2020.