Philip Lu, Boyi Li, S. Shama, I. King, Jonathan H. Chan
{"title":"基于众包标签的cnn面部表情识别损失层正则化","authors":"Philip Lu, Boyi Li, S. Shama, I. King, Jonathan H. Chan","doi":"10.1109/IESYS.2017.8233557","DOIUrl":null,"url":null,"abstract":"Deep, convolutional neural networks have become the state-of-the-art method for automatic Facial Expression Recognition (FER). Because of the small size and controlled conditions of most FER datasets, however, models can still overfit to the training dataset and struggle to generalize well to new data. We present a novel approach of using crowdsourced label distributions for improving the generalization performance of convolutional neural networks for FER. We implement this as a loss layer regularizer, where the ground truth labels are combined with crowdsourced labels in order to construct a noisy output distribution during training. We use a label disturbance method in which training examples are randomly replaced with incorrect labels drawn from the combined label probability distribution. We compare the performance of our disturbed and undisturbed models in cross-validation testing on the extended Cohn-Kanade dataset and cross-dataset experiments on the MMI, JAFFE, and FER2013 datasets. We find that using our proposed method, test performance is improved on both the MMI and JAFFE datasets. Our results suggest that using nonuniform probability distributions to disturb training can improve generalization performance of CNNs on other FER datasets.","PeriodicalId":429982,"journal":{"name":"2017 21st Asia Pacific Symposium on Intelligent and Evolutionary Systems (IES)","volume":"15 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2017-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":"{\"title\":\"Regularizing the loss layer of CNNs for facial expression recognition using crowdsourced labels\",\"authors\":\"Philip Lu, Boyi Li, S. Shama, I. King, Jonathan H. Chan\",\"doi\":\"10.1109/IESYS.2017.8233557\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Deep, convolutional neural networks have become the state-of-the-art method for automatic Facial Expression Recognition (FER). Because of the small size and controlled conditions of most FER datasets, however, models can still overfit to the training dataset and struggle to generalize well to new data. We present a novel approach of using crowdsourced label distributions for improving the generalization performance of convolutional neural networks for FER. We implement this as a loss layer regularizer, where the ground truth labels are combined with crowdsourced labels in order to construct a noisy output distribution during training. We use a label disturbance method in which training examples are randomly replaced with incorrect labels drawn from the combined label probability distribution. We compare the performance of our disturbed and undisturbed models in cross-validation testing on the extended Cohn-Kanade dataset and cross-dataset experiments on the MMI, JAFFE, and FER2013 datasets. We find that using our proposed method, test performance is improved on both the MMI and JAFFE datasets. Our results suggest that using nonuniform probability distributions to disturb training can improve generalization performance of CNNs on other FER datasets.\",\"PeriodicalId\":429982,\"journal\":{\"name\":\"2017 21st Asia Pacific Symposium on Intelligent and Evolutionary Systems (IES)\",\"volume\":\"15 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2017-11-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"3\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2017 21st Asia Pacific Symposium on Intelligent and Evolutionary Systems (IES)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/IESYS.2017.8233557\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2017 21st Asia Pacific Symposium on Intelligent and Evolutionary Systems (IES)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/IESYS.2017.8233557","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Regularizing the loss layer of CNNs for facial expression recognition using crowdsourced labels
Deep, convolutional neural networks have become the state-of-the-art method for automatic Facial Expression Recognition (FER). Because of the small size and controlled conditions of most FER datasets, however, models can still overfit to the training dataset and struggle to generalize well to new data. We present a novel approach of using crowdsourced label distributions for improving the generalization performance of convolutional neural networks for FER. We implement this as a loss layer regularizer, where the ground truth labels are combined with crowdsourced labels in order to construct a noisy output distribution during training. We use a label disturbance method in which training examples are randomly replaced with incorrect labels drawn from the combined label probability distribution. We compare the performance of our disturbed and undisturbed models in cross-validation testing on the extended Cohn-Kanade dataset and cross-dataset experiments on the MMI, JAFFE, and FER2013 datasets. We find that using our proposed method, test performance is improved on both the MMI and JAFFE datasets. Our results suggest that using nonuniform probability distributions to disturb training can improve generalization performance of CNNs on other FER datasets.