{"title":"基于Mahalanobis距离的k-means聚类的收敛问题","authors":"I. Lapidot","doi":"10.1109/ICSEE.2018.8646138","DOIUrl":null,"url":null,"abstract":"Mahalanobis distance is used for clustering and appears in different scenarios. Sometimes the same covariance is shared for all the clusters. This assumption is very restricted and it might be more meaningful that each cluster will be defined not only by its centroid but also with the covariance matrix. However, its use for k-means algorithm is not appropriate for optimization. It might lead to a good and meaningful clustering, but this is a fact of empirical observation and is not due to the algorithm’s convergence. In this study we will show that the overall distance may not decrease from one iteration to another, and that, to ensure convergence, some constraints must be added. Moreover, we will show that in an unconstrained clustering, the cluster covariance matrix is not a solution of the optimization process, but a constraint.","PeriodicalId":254455,"journal":{"name":"2018 IEEE International Conference on the Science of Electrical Engineering in Israel (ICSEE)","volume":"13 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2018-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"4","resultStr":"{\"title\":\"Convergence problems of Mahalanobis distance-based k-means clustering\",\"authors\":\"I. Lapidot\",\"doi\":\"10.1109/ICSEE.2018.8646138\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Mahalanobis distance is used for clustering and appears in different scenarios. Sometimes the same covariance is shared for all the clusters. This assumption is very restricted and it might be more meaningful that each cluster will be defined not only by its centroid but also with the covariance matrix. However, its use for k-means algorithm is not appropriate for optimization. It might lead to a good and meaningful clustering, but this is a fact of empirical observation and is not due to the algorithm’s convergence. In this study we will show that the overall distance may not decrease from one iteration to another, and that, to ensure convergence, some constraints must be added. Moreover, we will show that in an unconstrained clustering, the cluster covariance matrix is not a solution of the optimization process, but a constraint.\",\"PeriodicalId\":254455,\"journal\":{\"name\":\"2018 IEEE International Conference on the Science of Electrical Engineering in Israel (ICSEE)\",\"volume\":\"13 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2018-12-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"4\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2018 IEEE International Conference on the Science of Electrical Engineering in Israel (ICSEE)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICSEE.2018.8646138\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2018 IEEE International Conference on the Science of Electrical Engineering in Israel (ICSEE)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICSEE.2018.8646138","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Convergence problems of Mahalanobis distance-based k-means clustering
Mahalanobis distance is used for clustering and appears in different scenarios. Sometimes the same covariance is shared for all the clusters. This assumption is very restricted and it might be more meaningful that each cluster will be defined not only by its centroid but also with the covariance matrix. However, its use for k-means algorithm is not appropriate for optimization. It might lead to a good and meaningful clustering, but this is a fact of empirical observation and is not due to the algorithm’s convergence. In this study we will show that the overall distance may not decrease from one iteration to another, and that, to ensure convergence, some constraints must be added. Moreover, we will show that in an unconstrained clustering, the cluster covariance matrix is not a solution of the optimization process, but a constraint.