首页> 外文会议>World Multi-Conference on Systemics, Cybernetics and Informatics >Intelligent Ear for Emotion Recognition: Multi-Modal Emotion Recognition via Acoustic Features, Semantic Contents and Facial Images
【24h】

Intelligent Ear for Emotion Recognition: Multi-Modal Emotion Recognition via Acoustic Features, Semantic Contents and Facial Images

机译:情感识别智能耳朵:通过声学特征,语义内容和面部图像的多模态情感识别

获取原文

摘要

In this paper, based on the idea that humans are capable of detecting human emotions during a conversation through speech and facial expression input, an emotion recognition system that can detect the emotion from acoustic features, semantic contents, and facial expression during conversation is proposed. In the analysis of speech signals, thirty-three acoustic features are extracted from the speech input After Principle Component Analysis (PCA), 14 principle components are selected for discriminative representation. In this representation each principle component is the combination of the 33 original acoustic features and forms a feature subspace. The Support Vector Machines (SVMs) are adopted to classify the emotional states. In facial emotion recognition module, the facial image captured from CCD is provided for facial image feature extraction. An SVM model is applied for emotion recognition. Finally in text analysis, all emotional keywords and emotion modification words are manually defined. The emotion intensity levels of emotional keywords and emotion modification words are estimated from a collected emotion corpus. The final emotional state is determined based on the emotion outputs from these three modules. The experimental result shows that the emotion recognition accuracy of the integrated system is better than each of the three individual approaches.
机译:在本文中,基于人类能够通过语音和面部表情输入在对话期间能够检测人类情绪的想法,提出了一种情感识别系统,可以检测来自声学特征,语义内容和对话期间的面部表情的情感识别系统。在对语音信号的分析中,在原理分量分析(PCA)之后,从语音输入中提取三十三个声学特征,选择14个原理组分以进行辨别表示。在该表示中,每个原理组件是33原始声学功能的组合,并形成特征子空间。采用支持向量机(SVM)对情绪状态进行分类。在面部情感识别模块中,为面部图像特征提取提供了从CCD捕获的面部图像。 SVM模型适用于情感识别。最后在文本分析中,手动定义所有情绪关键字和情感修改词。情绪关键词和情感修改词的情绪强度水平估计来自收集的情绪语料库。最终的情绪状态是基于来自这三个模块的情绪输出来确定。实验结果表明,集成系统的情感识别精度优于三种单独方法中的每一种。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号