基于众包标签的cnn面部表情识别损失层正则化

Philip Lu, Boyi Li, S. Shama, I. King, Jonathan H. Chan
{"title":"基于众包标签的cnn面部表情识别损失层正则化","authors":"Philip Lu, Boyi Li, S. Shama, I. King, Jonathan H. Chan","doi":"10.1109/IESYS.2017.8233557","DOIUrl":null,"url":null,"abstract":"Deep, convolutional neural networks have become the state-of-the-art method for automatic Facial Expression Recognition (FER). Because of the small size and controlled conditions of most FER datasets, however, models can still overfit to the training dataset and struggle to generalize well to new data. We present a novel approach of using crowdsourced label distributions for improving the generalization performance of convolutional neural networks for FER. We implement this as a loss layer regularizer, where the ground truth labels are combined with crowdsourced labels in order to construct a noisy output distribution during training. We use a label disturbance method in which training examples are randomly replaced with incorrect labels drawn from the combined label probability distribution. We compare the performance of our disturbed and undisturbed models in cross-validation testing on the extended Cohn-Kanade dataset and cross-dataset experiments on the MMI, JAFFE, and FER2013 datasets. We find that using our proposed method, test performance is improved on both the MMI and JAFFE datasets. Our results suggest that using nonuniform probability distributions to disturb training can improve generalization performance of CNNs on other FER datasets.","PeriodicalId":429982,"journal":{"name":"2017 21st Asia Pacific Symposium on Intelligent and Evolutionary Systems (IES)","volume":"15 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2017-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":"{\"title\":\"Regularizing the loss layer of CNNs for facial expression recognition using crowdsourced labels\",\"authors\":\"Philip Lu, Boyi Li, S. Shama, I. King, Jonathan H. Chan\",\"doi\":\"10.1109/IESYS.2017.8233557\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Deep, convolutional neural networks have become the state-of-the-art method for automatic Facial Expression Recognition (FER). Because of the small size and controlled conditions of most FER datasets, however, models can still overfit to the training dataset and struggle to generalize well to new data. We present a novel approach of using crowdsourced label distributions for improving the generalization performance of convolutional neural networks for FER. We implement this as a loss layer regularizer, where the ground truth labels are combined with crowdsourced labels in order to construct a noisy output distribution during training. We use a label disturbance method in which training examples are randomly replaced with incorrect labels drawn from the combined label probability distribution. We compare the performance of our disturbed and undisturbed models in cross-validation testing on the extended Cohn-Kanade dataset and cross-dataset experiments on the MMI, JAFFE, and FER2013 datasets. We find that using our proposed method, test performance is improved on both the MMI and JAFFE datasets. Our results suggest that using nonuniform probability distributions to disturb training can improve generalization performance of CNNs on other FER datasets.\",\"PeriodicalId\":429982,\"journal\":{\"name\":\"2017 21st Asia Pacific Symposium on Intelligent and Evolutionary Systems (IES)\",\"volume\":\"15 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2017-11-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"3\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2017 21st Asia Pacific Symposium on Intelligent and Evolutionary Systems (IES)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/IESYS.2017.8233557\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2017 21st Asia Pacific Symposium on Intelligent and Evolutionary Systems (IES)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/IESYS.2017.8233557","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 3

摘要

深度卷积神经网络已经成为自动面部表情识别(FER)的最先进方法。然而,由于大多数FER数据集的规模小且条件可控,模型仍然可能过度拟合训练数据集,难以很好地推广到新数据。我们提出了一种使用众包标签分布来提高卷积神经网络泛化性能的新方法。我们将其实现为损失层正则化器,其中地面真值标签与众包标签相结合,以便在训练期间构建有噪声的输出分布。我们使用标签干扰方法,其中训练样本被随机替换为从组合标签概率分布中提取的错误标签。在扩展的Cohn-Kanade数据集和MMI、JAFFE和FER2013数据集的交叉验证测试中,我们比较了扰动模型和未扰动模型的性能。我们发现,使用我们的方法,在MMI和JAFFE数据集上的测试性能都得到了提高。我们的研究结果表明,使用非均匀概率分布来干扰训练可以提高cnn在其他FER数据集上的泛化性能。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Regularizing the loss layer of CNNs for facial expression recognition using crowdsourced labels
Deep, convolutional neural networks have become the state-of-the-art method for automatic Facial Expression Recognition (FER). Because of the small size and controlled conditions of most FER datasets, however, models can still overfit to the training dataset and struggle to generalize well to new data. We present a novel approach of using crowdsourced label distributions for improving the generalization performance of convolutional neural networks for FER. We implement this as a loss layer regularizer, where the ground truth labels are combined with crowdsourced labels in order to construct a noisy output distribution during training. We use a label disturbance method in which training examples are randomly replaced with incorrect labels drawn from the combined label probability distribution. We compare the performance of our disturbed and undisturbed models in cross-validation testing on the extended Cohn-Kanade dataset and cross-dataset experiments on the MMI, JAFFE, and FER2013 datasets. We find that using our proposed method, test performance is improved on both the MMI and JAFFE datasets. Our results suggest that using nonuniform probability distributions to disturb training can improve generalization performance of CNNs on other FER datasets.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信