Fault Classification Method based on Fast K-Nearest Neighbor with Hybrid Feature Generation and K - Medoids Clustering

Zhe Zhou, Fanliang Zeng, Jiacheng Huang, Jingjing Zheng, Zuxin Li
{"title":"Fault Classification Method based on Fast K-Nearest Neighbor with Hybrid Feature Generation and K - Medoids Clustering","authors":"Zhe Zhou, Fanliang Zeng, Jiacheng Huang, Jingjing Zheng, Zuxin Li","doi":"10.1109/YAC51587.2020.9337657","DOIUrl":null,"url":null,"abstract":"Fast and accurate classification of the process faults in industry is important for ensuring reliable operation. The traditional K-Nearest Neighbor (KNN) algorithm needs to calculate the distance between the sample to be classified and all the training samples obtained under normal operation condition (NOC). Large computational cost involves in this step when the number of the NOC samples is huge. Aiming at this, existing methods reduce the training samples by using the clustering algorithms, so as to reduce the computational cost of the KNN. However, the reduction of training samples usually leads to a decrease in the accuracy of fault classification. Misclassification directly affects the normal production and safety of the industrial processes. To this end, a fast KNN fault classification method based on hybrid feature generation and K-Medoids is proposed. Firstly, a hybrid feature generation method combining ReliefF algorithm and linear discriminant analysis algorithm is used to select and extract the sample features, thus enhance the separability inter-classes and improve the accuracy of fault classification. Then, K-Medoids clustering algorithm is used to select few representative training samples and reduce the computational complexity of KNN algorithm. Finally, the simulation is performed on the Tennessee- Eastman process, which verifies that the proposed algorithm is superior than other related four methods and only consumes far less running time than the basic KNN algorithm while retaining higher classification accuracy.","PeriodicalId":287095,"journal":{"name":"2020 35th Youth Academic Annual Conference of Chinese Association of Automation (YAC)","volume":"28 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-10-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2020 35th Youth Academic Annual Conference of Chinese Association of Automation (YAC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/YAC51587.2020.9337657","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Fast and accurate classification of the process faults in industry is important for ensuring reliable operation. The traditional K-Nearest Neighbor (KNN) algorithm needs to calculate the distance between the sample to be classified and all the training samples obtained under normal operation condition (NOC). Large computational cost involves in this step when the number of the NOC samples is huge. Aiming at this, existing methods reduce the training samples by using the clustering algorithms, so as to reduce the computational cost of the KNN. However, the reduction of training samples usually leads to a decrease in the accuracy of fault classification. Misclassification directly affects the normal production and safety of the industrial processes. To this end, a fast KNN fault classification method based on hybrid feature generation and K-Medoids is proposed. Firstly, a hybrid feature generation method combining ReliefF algorithm and linear discriminant analysis algorithm is used to select and extract the sample features, thus enhance the separability inter-classes and improve the accuracy of fault classification. Then, K-Medoids clustering algorithm is used to select few representative training samples and reduce the computational complexity of KNN algorithm. Finally, the simulation is performed on the Tennessee- Eastman process, which verifies that the proposed algorithm is superior than other related four methods and only consumes far less running time than the basic KNN algorithm while retaining higher classification accuracy.
基于混合特征生成和K-介质聚类的快速K近邻故障分类方法
工业生产过程故障的快速、准确分类对保证设备的可靠运行具有重要意义。传统的k -最近邻(KNN)算法需要计算待分类样本与在正常运行条件下获得的所有训练样本之间的距离。当NOC样本数量巨大时,该步骤的计算成本较大。针对这一点,现有方法通过使用聚类算法来减少训练样本,从而降低KNN的计算成本。然而,训练样本的减少往往会导致故障分类准确率的下降。分类错误直接影响工业过程的正常生产和安全。为此,提出了一种基于混合特征生成和k -媒质的快速KNN故障分类方法。首先,采用ReliefF算法和线性判别分析算法相结合的混合特征生成方法对样本特征进行选择和提取,增强了类间可分性,提高了故障分类的准确率;然后,利用K-Medoids聚类算法选取少量具有代表性的训练样本,降低KNN算法的计算复杂度;最后,在Tennessee- Eastman过程中进行了仿真,验证了该算法优于其他四种相关方法,并且在保持较高分类精度的同时,只消耗远少于基本KNN算法的运行时间。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信