Acoustic features extraction for emotion recognition
Rong, J., Chen, Yi-Ping Phoebe, Chowdhury, Morshed and Li, Gang 2007, Acoustic features extraction for emotion recognition, in 6th IEEE/ACIS International Conference on Computer and Information Science : (ICIS 2007) in conjunction with 1st IEEE/ACIS International Workshop on e-Activity (IWEA 2007) : proceedings : 11-13 July, 2007, Melbourne, Australia, IEEE Xplore, Piscataway, N.J., pp. 419-424.
(Some files may be inaccessible until you login with your Deakin Research Online credentials)
6th IEEE/ACIS International Conference on Computer and Information Science : (ICIS 2007) in conjunction with 1st IEEE/ACIS International Workshop on e-Activity (IWEA 2007) : proceedings : 11-13 July, 2007, Melbourne, Australia
Lee, Roger Chowdhury, Morshed Ray, Sid Lee, Thuy
International Conference on Computer and Information Science
In the last decade, the efforts of spoken language processing have achieved significant advances, however, the work with emotional recognition has not progressed so far, and can only achieve 50% to 60% in accuracy. This is because a majority of researchers in this field have focused on the synthesis of emotional speech rather than focusing on automating human emotion recognition. Many research groups have focused on how to improve the performance of the classifier they used for emotion recognition, and few work has been done on data pre-processing, such as the extraction and selection of a set of specifying acoustic features instead of using all the possible ones they had in hand. To work with well-selected acoustic features does not mean to delay the whole job, but this will save much time and resources by removing the irrelative information and reducing the high-dimension data calculation. In this paper, we developed an automatic feature selector based on a RF2TREE algorithm and the traditional C4.5 algorithm. RF2TREE applied here helped us to solve the problems that did not have enough data examples. The ensemble learning technique was applied to enlarge the original data set by building a bagged random forest to generate many virtual examples, and then the new data set was used to train a single decision tree, which selects the most efficient features to represent the speech signals for the emotion recognition. Finally, the output of the selector was a set of specifying acoustic features, produced by RF2TREE and a single decision tree.
This material is presented to ensure timely dissemination of scholarly and technical work. Copyright and all rights therein are retained by authors or by other copyright holders. All persons copying this information are expected to adhere to the terms and constraints invoked by each author's copyright. In most cases, these works may not be reposted without the explicit permission of the copyright holder.
Unless expressly stated otherwise, the copyright for items in Deakin Research Online is owned by the author, with all rights reserved.
Every reasonable effort has been made to ensure that permission has been obtained for items included in DRO.
If you believe that your rights have been infringed by this repository, please contact firstname.lastname@example.org.