Leveraging upon transfer learning, we distill the knowledge in a conventional wide and deep neural network (DNN) into a narrower yet deeper model with fewer parameters and comparable system performance for speech enhancement. We present three transfer-learning solutions to accomplish our goal. First, the knowledge embedded in the form of the output values of a high-performance DNN is used to guide the training of a smaller DNN model in sequential transfer learning. In the second multi-task transfer learning solution, the smaller DNN is trained to learn the output value of the larger DNN, and the speech enhancement task in parallel. Finally, a progressive stacking transfer learning is accomplished through multi-task learning, and DNN stacking. Our experimental evidences demonstrate 5 times parameter reduction while maintaining similar enhancement performance with the proposed framework

Wang, S., Li, K., Huang, Z., SINISCALCHI, S.M., Lee, C.H. (2017). A transfer learning and progressive stacking approach to reducing deep model sizes with an application to speech enhancement. In 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) (pp. 5575-5579). Institute of Electrical and Electronics Engineers Inc. [10.1109/ICASSP.2017.7953223].

A transfer learning and progressive stacking approach to reducing deep model sizes with an application to speech enhancement

SINISCALCHI, SABATO MARCO;
2017-01-01

Abstract

Leveraging upon transfer learning, we distill the knowledge in a conventional wide and deep neural network (DNN) into a narrower yet deeper model with fewer parameters and comparable system performance for speech enhancement. We present three transfer-learning solutions to accomplish our goal. First, the knowledge embedded in the form of the output values of a high-performance DNN is used to guide the training of a smaller DNN model in sequential transfer learning. In the second multi-task transfer learning solution, the smaller DNN is trained to learn the output value of the larger DNN, and the speech enhancement task in parallel. Finally, a progressive stacking transfer learning is accomplished through multi-task learning, and DNN stacking. Our experimental evidences demonstrate 5 times parameter reduction while maintaining similar enhancement performance with the proposed framework
2017
978-1-5090-4117-6
Wang, S., Li, K., Huang, Z., SINISCALCHI, S.M., Lee, C.H. (2017). A transfer learning and progressive stacking approach to reducing deep model sizes with an application to speech enhancement. In 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) (pp. 5575-5579). Institute of Electrical and Electronics Engineers Inc. [10.1109/ICASSP.2017.7953223].
File in questo prodotto:
File Dimensione Formato  
07953223.pdf

Solo gestori archvio

Tipologia: Versione Editoriale
Dimensione 477.15 kB
Formato Adobe PDF
477.15 kB Adobe PDF   Visualizza/Apri   Richiedi una copia

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/10447/649514
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 12
  • ???jsp.display-item.citation.isi??? 12
social impact