Emotional speech classification with prosodic prameters by using neural networks

H. Sato, Y. Mitsukura, M. Fukumi, N. Akamatsu
{"title":"Emotional speech classification with prosodic prameters by using neural networks","authors":"H. Sato, Y. Mitsukura, M. Fukumi, N. Akamatsu","doi":"10.1109/ANZIIS.2001.974111","DOIUrl":null,"url":null,"abstract":"Interestingly, in order to achieve a new Human Interface such that digital computers can deal with the KASEI information, the study of the KANSEI information processing recently has been approached. In this paper, we propose a new classification method of emotional speech by analyzing feature parameters obtained from the emotional speech and by learning them using neural networks, which is regarded as a KANSEI information processing. In the present research, KANSEI information is usually human emotion. The emotion is classified broadly into four patterns such as neutral, anger, sad and joy. The pitch as one of feature parameters governs voice modulation, and can be sensitive to change of emotion. The pitch is extracted from each emotional speech by the cepstrum method. Input values of neural networks (NNs) are then emotional pitch patterns, which are time-varying. It is shown that NNs can achieve classification of emotion by learning each emotional pitch pattern by means of computer simulations.","PeriodicalId":383878,"journal":{"name":"The Seventh Australian and New Zealand Intelligent Information Systems Conference, 2001","volume":"51 6","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"16","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"The Seventh Australian and New Zealand Intelligent Information Systems Conference, 2001","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ANZIIS.2001.974111","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 16

Abstract

Interestingly, in order to achieve a new Human Interface such that digital computers can deal with the KASEI information, the study of the KANSEI information processing recently has been approached. In this paper, we propose a new classification method of emotional speech by analyzing feature parameters obtained from the emotional speech and by learning them using neural networks, which is regarded as a KANSEI information processing. In the present research, KANSEI information is usually human emotion. The emotion is classified broadly into four patterns such as neutral, anger, sad and joy. The pitch as one of feature parameters governs voice modulation, and can be sensitive to change of emotion. The pitch is extracted from each emotional speech by the cepstrum method. Input values of neural networks (NNs) are then emotional pitch patterns, which are time-varying. It is shown that NNs can achieve classification of emotion by learning each emotional pitch pattern by means of computer simulations.
基于神经网络的带韵律参数的情绪语音分类
有趣的是,为了实现一种新的人机界面,使数字计算机可以处理KASEI信息,最近已经开始研究KANSEI信息处理。本文提出了一种新的情感语音分类方法,通过分析情感语音的特征参数,并利用神经网络对其进行学习,将其视为一种感性信息处理。在目前的研究中,感性信息通常是人类的情感。这种情绪大致分为四种类型,如中性、愤怒、悲伤和快乐。音高作为控制声音调制的特征参数之一,对情绪的变化非常敏感。用倒谱法从每个情感言语中提取音高。神经网络(nn)的输入值是时变的情绪音高模式。研究表明,神经网络可以通过计算机模拟来学习每种情绪音高模式,从而实现情绪的分类。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信