Phoneme classification using the auditory neurogram

Md Shariful Alam*, Muhammad S.A. Zilany, Wissam A. Jassim, Mohd Yazed Ahmad

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

9 Citations (Scopus)

Abstract

In order to mimic the capability of human listeners identifying speech in noisy environments, this paper proposes a phoneme classification technique using simulated neural responses from a physiologically based computational model of the auditory periphery instead of using features directly from the acoustic signal. The 2-D neurograms were constructed from the simulated responses of the auditory-nerve fibers to speech phonemes. The features of the neurograms were extracted using the Radon transform and used to train the classification system using a deep neural network classifier. Classification performance was evaluated in quiet and under noisy conditions for different types of phonemes extracted from the TIMIT database. Based on simulation results, the proposed method outperformed most of the traditional acoustic-property-based phoneme classification methods for both in quiet and under noisy conditions. The proposed method could easily be extended to develop an automatic speech recognition system.

Original languageEnglish
Article number7815325
Pages (from-to)633-642
Number of pages10
JournalIEEE Access
Volume5
DOIs
Publication statusPublished - 2017
Externally publishedYes

Keywords

  • Auditory-nerve model
  • DNN
  • FDLP
  • MFCC
  • neurogram
  • phoneme classification.

Fingerprint

Dive into the research topics of 'Phoneme classification using the auditory neurogram'. Together they form a unique fingerprint.

Cite this