Speech Emotion Verification Using Emotion Variance And Discriminate Scale Frequency Maps

P. Sunitha, E. Lakshmi Shravya

Abstract


The problems of emotion variances and emotion blending are solved using the emotion verification system. In this methodology consists of two parts. The first part, feature extraction is used to lessen the huge input data into smaller data. The second part, emotion verification can automatically verify and authenticate users from their voice or speech. For each sound frames, vital atoms from the Gabor dictionary are selected by the using of vector quantization. Each atom within the dictionary takes the form of a Gabor function which includes frequency, scale etc.Sparse representation is used to convert scale-frequency maps into sparse coefficients to enhance the robustness towards emotion variance.After the feature extraction, a fusion mechanism is applied the subsequent two strategies. The first is the sparse representation verification based on Gaussian modeled residual errors which can be accommodate emotion variance within the voices of speakers. Such a classifier can be reduce emotion variance and improve recognition accuracy. The second, is to involves the use of an indicator, the EAI, for the measuring the degree of blended emotions. The two scores have been calculated they are fused with each other. It also verifies the performance as well as feasibility.


References


W. H. Abdulla and N. Kasabov, “Improving speech recognition performance through gender separation,” in Proceedings of the International Conference on Artificial Neural Networks and Expert Systems (ANNES), Dunedin, New Zealand, 2001, pp. 218–222.

D. W. Aha, D. Kibler, and M. K. Albert, “Instance-based learning algorithms,” Machine Learning, vol. 6, no. 1, pp. 37–66, January 1991.

H. Ai, D. J. Litman, K. Forbes-Riley, M. Rotaru, J. Tetreault, and A. Purandare, “Using system and user performance features to improve emotion detection in spoken tutoring dialogs,” in Proceedings of the 9th International Conference on Spoken Language Processing (INTERSPEECH), Pittsburgh, PA, USA, September 2006.

K. Alter, E. Rank, S. A. Kotz, U. Toepel, M. Besson, A. Schirmer, and A. D. Friderici, “Affective encoding in the speech signal and in event-related brain potentials,” Speech Communication, vol. 40, no. 1–2, pp. 61–70, April 2003.

A. Álvarez, I. Cearreta, J. M. López, A. Arruti, E. Lazkano, B. Sierra, and N. Garay, “Feature subset selection based on evolutionary algorithms for automatic emotion recognition in spoken Spanish and standard Basque language,” in Proceedings of the 9th International Conference on Text, Speech and Dialogue (TSD), Brno, Czech Republic, September 2006, pp. 565–572.


Full Text: PDF [Full Text]

Refbacks

  • There are currently no refbacks.


Copyright © 2013, All rights reserved.| ijseat.com

Creative Commons License
International Journal of Science Engineering and Advance Technology is licensed under a Creative Commons Attribution 3.0 Unported License.Based on a work at IJSEat , Permissions beyond the scope of this license may be available at http://creativecommons.org/licenses/by/3.0/deed.en_GB.