Redirigiendo al acceso original de articulo en 15 segundos...
ARTÍCULO
TITULO

Speech Emotion Recognition System Using Gaussian Mixture Model and Improvement proposed via Boosted GMM

Pavitra Patel    
A. A. Chaudhari    
M. A. Pund    
D. H. Deshmukh    

Resumen

Speech emotion recognition is an important issue which affects the human machine interaction. Automatic recognition of human emotion in speech aims at recognizing the underlying emotional state of a speaker from the speech signal. Gaussian mixture models (GMMs) and the minimum error rate classifier (i.e. Bayesian optimal classifier) are popular and effective tools for speech emotion recognition. Typically, GMMs are used to model the class-conditional distributions of acoustic features and their parameters are estimated by the expectation maximization (EM) algorithm based on a training data set. In this paper, we introduce a boosting algorithm for reliably and accurately estimating the class-conditional GMMs. The resulting algorithm is named the Boosted-GMM algorithm. Our speech emotion recognition experiments show that the emotion recognition rates are effectively and significantly boosted by the Boosted-GMM algorithm as compared to the EM-GMM algorithm.During this interaction, human beings have some feelings that they want to convey to their communication partner with whom they are communicating, and then their communication partner may be the human or machine. This work dependent on the emotion recognition of the human beings from their speech signalEmotion recognition from the speaker?s speech is very difficult because of the following reasons: Because of the existence of the different sentences, speakers, speaking styles, speaking rates accosting variability was introduced. The same utterance may show different emotions. Therefore it is very difficult to differentiate these portions of utterance. Another problem is that emotion expression is depending on the speaker and his or her culture and environment. As the culture and environment gets change the speaking style also gets change, which is another challenge in front of the speech emotion recognition system.

Palabras claves

 Artículos similares

       
 
Zhichao Peng, Wenhua He, Yongwei Li, Yegang Du and Jianwu Dang    
Speech emotion recognition is a critical component for achieving natural human?robot interaction. The modulation-filtered cochleagram is a feature based on auditory modulation perception, which contains multi-dimensional spectral?temporal modulation repr... ver más
Revista: Applied Sciences

 
Jialin Zhang, Mairidan Wushouer, Gulanbaier Tuerhong and Hanfang Wang    
Emotional speech synthesis is an important branch of human?computer interaction technology that aims to generate emotionally expressive and comprehensible speech based on the input text. With the rapid development of speech synthesis technology based on ... ver más
Revista: Applied Sciences

 
Yingmei Qi, Heming Huang and Huiyun Zhang    
Speech emotion recognition is a crucial work direction in speech recognition. To increase the performance of speech emotion detection, researchers have worked relentlessly to improve data augmentation, feature extraction, and pattern formation. To addres... ver más
Revista: Applied Sciences

 
Sung-Woo Byun and Seok-Pil Lee    
The goal of the human interface is to recognize the user?s emotional state precisely. In the speech emotion recognition study, the most important issue is the effective parallel use of the extraction of proper speech features and an appropriate classific... ver más
Revista: Applied Sciences

 
Noé Tits, Kevin El Haddad and Thierry Dutoit    
In this paper, we study the controllability of an Expressive TTS system trained on a dataset for a continuous control. The dataset is the Blizzard 2013 dataset based on audiobooks read by a female speaker containing a great variability in styles and expr... ver más
Revista: Informatics