Deep recurrent mixture of experts for speech enhancement

Shlomo E. Chazan, Jacob Goldberger, Sharon Gannot

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

10 Scopus citations

Abstract

Deep neural networks (DNNs) have recently became a viable methodology for single microphone speech enhancement. The most common approach, is to feed the noisy speech features into a fully-connected DNN to either directly enhance the speech signal or to infer a mask which can be used for the speech enhancement. In this case, one network has to deal with the large variability of the speech signal. Most approaches also discard the speech continuity. In this paper, we propose a deep recurrent mixture of experts (DRMoE) architecture that addresses these two issues. In order to reduce the large speech variability, we split the network into a mixture of networks (denoted experts), each of which specializes in a specific and simpler task and a gating network. The time-continuity of the speech signal is taken into account by implementing the experts and the gating network as a recurrent neural network (RNN). Experimental study shows that the proposed algorithm produces higher objective measurements scores compared to both a single RNN and a deep mixture of experts (DMoE) architectures.

Original languageEnglish
Title of host publication2017 IEEE Workshop on Applications of Signal Processing to Audio and Acoustics, WASPAA 2017
PublisherInstitute of Electrical and Electronics Engineers Inc.
Pages359-363
Number of pages5
ISBN (Electronic)9781538616321
DOIs
StatePublished - 7 Dec 2017
Event2017 IEEE Workshop on Applications of Signal Processing to Audio and Acoustics, WASPAA 2017 - New Paltz, United States
Duration: 15 Oct 201718 Oct 2017

Publication series

NameIEEE Workshop on Applications of Signal Processing to Audio and Acoustics
Volume2017-October

Conference

Conference2017 IEEE Workshop on Applications of Signal Processing to Audio and Acoustics, WASPAA 2017
Country/TerritoryUnited States
CityNew Paltz
Period15/10/1718/10/17

Bibliographical note

Publisher Copyright:
© 2017 IEEE.

Keywords

  • long short-Term memory
  • recurrent neural network
  • speech presence probability

Fingerprint

Dive into the research topics of 'Deep recurrent mixture of experts for speech enhancement'. Together they form a unique fingerprint.

Cite this