Speech recognition has become common in many application domains. Incorporating acoustic-phonetic knowledge into Automatic Speech Recognition (ASR) systems design has been proven a viable approach to rise ASR accuracy. Manner of articulation attributes such as vowel, stop, fricative, approximant, nasal, and silence are examples of such knowledge. Neural networks have already been used successfully as detectors for manner of articulation attributes starting from representations of speech signal frames. In this paper, a set of six detectors for the above mentioned attributes is designed based on the E-αNet model of neural networks. This model was chosen for its capability to learn hidden activation functions that results in better generalization properties. Experimental set-up and results are presented that show an average 3.5% improvement over a baseline neural network implementation.

Speech recognition has become common in many application domains. Incorporating acoustic-phonetic knowledge into Automatic Speech Recognition (ASR) systems design has been proven a viable approach to rise ASR accuracy. Manner of articulation attributes such as vowel, stop, fricative, approximant, nasal, and silence are examples of such knowledge. Neural networks have already been used successfully as detectors for manner of articulation attributes starting from representations of speech signal frames. In this paper, a set of six detectors for the above mentioned attributes is designed based on the E-αNet model of neural networks. This model was chosen for its capability to learn hidden activation functions that results in better generalization properties. Experimental set-up and results are presented that show an average 3.5% improvement over a baseline neural network implementation

Siniscalchi, S.M., Li, J., Pilato, G., Vassallo, g., Clements, M.A., Gentile, A., et al. (2006). Application of EalphaNets to Feature Recognition of Articulation Manner in Knowledge-Based Automatic Speech Recognition. In M.M. Bruno Apolloni (a cura di), Lecture Notes in Computer Science (pp. 140-146). Springer Verlag [10.1007/11731177_21].

Application of EalphaNets to Feature Recognition of Articulation Manner in Knowledge-Based Automatic Speech Recognition

VASSALLO, Giorgio;GENTILE, Antonio;SORBELLO, Filippo;SINISCALCHI, Sabato Marco
2006-01-01

Abstract

Speech recognition has become common in many application domains. Incorporating acoustic-phonetic knowledge into Automatic Speech Recognition (ASR) systems design has been proven a viable approach to rise ASR accuracy. Manner of articulation attributes such as vowel, stop, fricative, approximant, nasal, and silence are examples of such knowledge. Neural networks have already been used successfully as detectors for manner of articulation attributes starting from representations of speech signal frames. In this paper, a set of six detectors for the above mentioned attributes is designed based on the E-αNet model of neural networks. This model was chosen for its capability to learn hidden activation functions that results in better generalization properties. Experimental set-up and results are presented that show an average 3.5% improvement over a baseline neural network implementation
2006
Settore ING-INF/05 - Sistemi Di Elaborazione Delle Informazioni
Siniscalchi, S.M., Li, J., Pilato, G., Vassallo, g., Clements, M.A., Gentile, A., et al. (2006). Application of EalphaNets to Feature Recognition of Articulation Manner in Knowledge-Based Automatic Speech Recognition. In M.M. Bruno Apolloni (a cura di), Lecture Notes in Computer Science (pp. 140-146). Springer Verlag [10.1007/11731177_21].
File in questo prodotto:
File Dimensione Formato  
2121565-pages-.pdf

Solo gestori archvio

Descrizione: pdf
Dimensione 1.69 MB
Formato Adobe PDF
1.69 MB Adobe PDF   Visualizza/Apri   Richiedi una copia

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/10447/59647
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 1
  • ???jsp.display-item.citation.isi??? 0
social impact