ZHANG Xin-ran, ZHA Cheng, SONG Peng, TAO Hua-wei, ZHAO Li. Spectrogram Speech Emotion Recognition Method Based on Auditory Attention Model[J]. JOURNAL OF SIGNAL PROCESSING, 2016, 32(9): 1117-1125. DOI: 10.16798/j.issn.1003-0530.2016.09.15
Citation: ZHANG Xin-ran, ZHA Cheng, SONG Peng, TAO Hua-wei, ZHAO Li. Spectrogram Speech Emotion Recognition Method Based on Auditory Attention Model[J]. JOURNAL OF SIGNAL PROCESSING, 2016, 32(9): 1117-1125. DOI: 10.16798/j.issn.1003-0530.2016.09.15

Spectrogram Speech Emotion Recognition Method Based on Auditory Attention Model

  • When there exists mismatch between the trained acoustic models and the test utterances due to noise conditions, speaking styles and speaker traits, unmatched features may appear in cross-corpus. The resulting is the drastic degression in the performance of speech emotion recognition. Hence, the auditory attention model is found to be very effective for variational emotion features detection in our work. Therefore, Chirplet has been adopted to obtain salient gist features which show their relation to the expected performance in cross-corpus testing. In our experimental results, the prototypical classifier with the proposed feature extraction approach can deliver a gain of up to 9.6% accuracy in cross-corpus speech recognition, which is observed insensitive to different databases.
  • loading

Catalog

    /

    DownLoad:  Full-Size Img  PowerPoint
    Return
    Return