Bin-Bin Yang, Shengjie Zhao, Kenan Ye, Rongqing Zhang
{"title":"不平衡数据集有序回归的二次Kappa损失中的分布一致性惩罚","authors":"Bin-Bin Yang, Shengjie Zhao, Kenan Ye, Rongqing Zhang","doi":"10.1145/3507548.3507612","DOIUrl":null,"url":null,"abstract":"Ordinal regression is a typical deep learning problem, which involves inherently ordered labels that are common in practical applications, especially in medical diagnosis tasks. To overcome the neglect of ordered or non-stationary property by merely exploiting classification or regression, quadratic weighted kappa (QWK) is proposed to be employed in the QWK loss function design as an efficient evaluation metric for ordinal regression. However, the paradox that kappa will be higher with an asymmetrical marginal histogram leads the QWK loss function to get the local optimal solution with all-zero-column in the confusion matrices during training. In practice, the all-zero column problem will result in a certain category not being detected at all, which can have serious consequences for the exclusion of pathology. To address this limitation, a new form of penalty term is proposed for the QWK loss function by penalizing the distance of marginal histogram to effectively avoid all-zero-column of the models. The experiments on the category-imbalanced datasets demonstrate that our penalty terms solve all-zero-column problem. On Adience dataset our penalty terms achieve 0.915 QWK, 0.446 MAE and 0.612 accuracy, while on DR dataset our penalty terms achieve 0.744 QWK, 0.281 MAE and 0.810 accuracy. Besides, experiments on the category-balanced datasets HCI show that our penalty terms achieve 0.810 QWK, 0.499 MAE and 0.610 accuracy.","PeriodicalId":414908,"journal":{"name":"Proceedings of the 2021 5th International Conference on Computer Science and Artificial Intelligence","volume":"22 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Distribution Consistency Penalty in the Quadratic Kappa Loss for Ordinal Regression of Imbalanced Datasets\",\"authors\":\"Bin-Bin Yang, Shengjie Zhao, Kenan Ye, Rongqing Zhang\",\"doi\":\"10.1145/3507548.3507612\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Ordinal regression is a typical deep learning problem, which involves inherently ordered labels that are common in practical applications, especially in medical diagnosis tasks. To overcome the neglect of ordered or non-stationary property by merely exploiting classification or regression, quadratic weighted kappa (QWK) is proposed to be employed in the QWK loss function design as an efficient evaluation metric for ordinal regression. However, the paradox that kappa will be higher with an asymmetrical marginal histogram leads the QWK loss function to get the local optimal solution with all-zero-column in the confusion matrices during training. In practice, the all-zero column problem will result in a certain category not being detected at all, which can have serious consequences for the exclusion of pathology. To address this limitation, a new form of penalty term is proposed for the QWK loss function by penalizing the distance of marginal histogram to effectively avoid all-zero-column of the models. The experiments on the category-imbalanced datasets demonstrate that our penalty terms solve all-zero-column problem. On Adience dataset our penalty terms achieve 0.915 QWK, 0.446 MAE and 0.612 accuracy, while on DR dataset our penalty terms achieve 0.744 QWK, 0.281 MAE and 0.810 accuracy. Besides, experiments on the category-balanced datasets HCI show that our penalty terms achieve 0.810 QWK, 0.499 MAE and 0.610 accuracy.\",\"PeriodicalId\":414908,\"journal\":{\"name\":\"Proceedings of the 2021 5th International Conference on Computer Science and Artificial Intelligence\",\"volume\":\"22 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-12-04\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Proceedings of the 2021 5th International Conference on Computer Science and Artificial Intelligence\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1145/3507548.3507612\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 2021 5th International Conference on Computer Science and Artificial Intelligence","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3507548.3507612","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Distribution Consistency Penalty in the Quadratic Kappa Loss for Ordinal Regression of Imbalanced Datasets
Ordinal regression is a typical deep learning problem, which involves inherently ordered labels that are common in practical applications, especially in medical diagnosis tasks. To overcome the neglect of ordered or non-stationary property by merely exploiting classification or regression, quadratic weighted kappa (QWK) is proposed to be employed in the QWK loss function design as an efficient evaluation metric for ordinal regression. However, the paradox that kappa will be higher with an asymmetrical marginal histogram leads the QWK loss function to get the local optimal solution with all-zero-column in the confusion matrices during training. In practice, the all-zero column problem will result in a certain category not being detected at all, which can have serious consequences for the exclusion of pathology. To address this limitation, a new form of penalty term is proposed for the QWK loss function by penalizing the distance of marginal histogram to effectively avoid all-zero-column of the models. The experiments on the category-imbalanced datasets demonstrate that our penalty terms solve all-zero-column problem. On Adience dataset our penalty terms achieve 0.915 QWK, 0.446 MAE and 0.612 accuracy, while on DR dataset our penalty terms achieve 0.744 QWK, 0.281 MAE and 0.810 accuracy. Besides, experiments on the category-balanced datasets HCI show that our penalty terms achieve 0.810 QWK, 0.499 MAE and 0.610 accuracy.