Leveraging upon transfer learning, we distill the knowledge in a conventional wide and deep neural network (DNN) into a narrower yet deeper model with fewer parameters and comparable system performance for speech enhancement. We present three transfer-learning solutions to accomplish our goal. First, the knowledge embedded in the form of the output values of a high-performance DNN is used to guide the training of a smaller DNN model in sequential transfer learning. In the second multi-task transfer learning solution, the smaller DNN is trained to learn the output value of the larger DNN, and the speech enhancement task in parallel. Finally, a progressive stacking transfer learning is accomplished through multi-task learning, and DNN stacking. Our experimental evidences demonstrate 5 times parameter reduction while maintaining similar enhancement performance with the proposed framework
Wang, S., Li, K., Huang, Z., SINISCALCHI, S.M., Lee, C.H. (2017). A transfer learning and progressive stacking approach to reducing deep model sizes with an application to speech enhancement. In 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) (pp. 5575-5579). Institute of Electrical and Electronics Engineers Inc. [10.1109/ICASSP.2017.7953223].
A transfer learning and progressive stacking approach to reducing deep model sizes with an application to speech enhancement
SINISCALCHI, SABATO MARCO;
2017-01-01
Abstract
Leveraging upon transfer learning, we distill the knowledge in a conventional wide and deep neural network (DNN) into a narrower yet deeper model with fewer parameters and comparable system performance for speech enhancement. We present three transfer-learning solutions to accomplish our goal. First, the knowledge embedded in the form of the output values of a high-performance DNN is used to guide the training of a smaller DNN model in sequential transfer learning. In the second multi-task transfer learning solution, the smaller DNN is trained to learn the output value of the larger DNN, and the speech enhancement task in parallel. Finally, a progressive stacking transfer learning is accomplished through multi-task learning, and DNN stacking. Our experimental evidences demonstrate 5 times parameter reduction while maintaining similar enhancement performance with the proposed frameworkFile | Dimensione | Formato | |
---|---|---|---|
07953223.pdf
Solo gestori archvio
Tipologia:
Versione Editoriale
Dimensione
477.15 kB
Formato
Adobe PDF
|
477.15 kB | Adobe PDF | Visualizza/Apri Richiedi una copia |
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.