Emotion Recognition using Mel-Frequency Cepstral Coefficients
スポンサーリンク
概要
- 論文の詳細を見る
In this paper, we propose a new approach to emotion recognition. Prosodic features are currently used in most emotion recognition algorithms. However, emotion recognition algorithms using prosodic features are not sufficiently accurate. Therefore, we focused on the phonetic features of speech for emotion recognition. In particular, we describe the effectiveness of Mel-frequency Cepstral Coefficients (MFCCs) as the feature for emotion recognition. We focus on the precise classification of MFCC feature vectors, rather than their dynamic nature over an utterance. To realize such an approach, the proposed algorithm employs multi-template emotion classification of the analysis frames. Experimental evaluations show that the proposed algorithm produces 66.4% recognition accuracy in speaker-independent emotion recognition experiments for four specific emotions. This recognition accuracy is higher than the accuracy obtained by the conventional prosody-based and MFCC-based emotion recognition algorithms, which confirms the potential of the proposed algorithm.
- Information and Media Technologies 編集運営会議の論文
著者
-
Obuchi Yasunari
Central Research Laboratory Hitachi Ltd.
-
Sato Nobuo
Advanced Research Laboratory Hitachi Ltd.
関連論文
- Multi-Input Feature Combination in the Cepstral Domain for Practical Speech Recognition Systems
- Intentional Voice Command Detection for Trigger-Free Speech Interface
- Emotion Recognition using Mel-Frequency Cepstral Coefficients
- Stepwise Phase Difference Restoration Method for DOA Estimation of Multiple Sources
- Multichannel Two-Stage Beamforming with Unconstrained Beamformer and Distortion Reduction
- Noise suppression method for preprocessor of time-lag speech recognition system based on bidirectional optimally modified log spectral amplitude estimation
- Emotion Recognition using Mel-Frequency Cepstral Coefficients