Enhancing Multi-Layer Perceptron Performance with K-Means Clustering

Doughlas Pardede, Aulia Ichsan, Sugeng Riyadi
{"title":"Enhancing Multi-Layer Perceptron Performance with K-Means Clustering","authors":"Doughlas Pardede, Aulia Ichsan, Sugeng Riyadi","doi":"10.47709/cnahpc.v6i1.3600","DOIUrl":null,"url":null,"abstract":"Machine learning plays a crucial role in identifying patterns within data, with classification being a prominent application. This study investigates the use of Multilayer Perceptron (MLP) classification models and explores preprocessing techniques, particularly K-Means clustering, to enhance model performance. Overfitting, a common challenge in MLP models, is addressed through the application of K-Means clustering to streamline data preparation and improve classification accuracy. The study begins with an overview of overfitting in MLP models, highlighting the significance of mitigating this issue. Various techniques for addressing overfitting are reviewed, including regularization, dropout, early stopping, data augmentation, and ensemble methods. Additionally, the complementary role of K-Means clustering in enhancing model performance is emphasized. Preprocessing using K-Means clustering aims to reduce data complexity and prevent overfitting in MLP models. Three datasets - Iris, Wine, and Breast Cancer Wisconsin - are employed to evaluate the performance of K-Means as a preprocessing technique. Results from cross-validation demonstrate significant improvements in accuracy, precision, recall, and F1 scores when employing K-Means clustering compared to models without preprocessing. The findings highlight the efficacy of K-Means clustering in enhancing the discriminative power of MLP classification models by organizing data into clusters based on similarity. These results have practical implications, underlining the importance of appropriate preprocessing techniques in improving classification performance. Future research could explore additional preprocessing methods and their impact on classification accuracy across diverse datasets, advancing the field of machine learning and its applications","PeriodicalId":15605,"journal":{"name":"Journal Of Computer Networks, Architecture and High Performance Computing","volume":"4 2","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-02-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal Of Computer Networks, Architecture and High Performance Computing","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.47709/cnahpc.v6i1.3600","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Machine learning plays a crucial role in identifying patterns within data, with classification being a prominent application. This study investigates the use of Multilayer Perceptron (MLP) classification models and explores preprocessing techniques, particularly K-Means clustering, to enhance model performance. Overfitting, a common challenge in MLP models, is addressed through the application of K-Means clustering to streamline data preparation and improve classification accuracy. The study begins with an overview of overfitting in MLP models, highlighting the significance of mitigating this issue. Various techniques for addressing overfitting are reviewed, including regularization, dropout, early stopping, data augmentation, and ensemble methods. Additionally, the complementary role of K-Means clustering in enhancing model performance is emphasized. Preprocessing using K-Means clustering aims to reduce data complexity and prevent overfitting in MLP models. Three datasets - Iris, Wine, and Breast Cancer Wisconsin - are employed to evaluate the performance of K-Means as a preprocessing technique. Results from cross-validation demonstrate significant improvements in accuracy, precision, recall, and F1 scores when employing K-Means clustering compared to models without preprocessing. The findings highlight the efficacy of K-Means clustering in enhancing the discriminative power of MLP classification models by organizing data into clusters based on similarity. These results have practical implications, underlining the importance of appropriate preprocessing techniques in improving classification performance. Future research could explore additional preprocessing methods and their impact on classification accuracy across diverse datasets, advancing the field of machine learning and its applications
利用 K-Means 聚类提高多层感知器性能
机器学习在识别数据中的模式方面发挥着至关重要的作用,其中分类是一项突出的应用。本研究调查了多层感知器(MLP)分类模型的使用情况,并探索了预处理技术,特别是 K-Means 聚类,以提高模型性能。过拟合是 MLP 模型面临的常见挑战,本研究通过应用 K-Means 聚类技术来解决这一问题,从而简化数据准备工作并提高分类准确性。研究首先概述了 MLP 模型中的过拟合问题,强调了缓解这一问题的重要性。研究回顾了解决过拟合问题的各种技术,包括正则化、剔除、提前停止、数据增强和集合方法。此外,还强调了 K-Means 聚类在提高模型性能方面的补充作用。使用 K-Means 聚类进行预处理的目的是降低数据复杂性,防止 MLP 模型过度拟合。研究采用了三个数据集--虹膜、葡萄酒和威斯康星州乳腺癌--来评估 K-Means 作为预处理技术的性能。交叉验证的结果表明,与未进行预处理的模型相比,采用 K-Means 聚类技术的准确度、精确度、召回率和 F1 分数都有显著提高。研究结果突出表明,K-Means 聚类技术能根据相似性将数据组织成群,从而提高 MLP 分类模型的判别能力。这些结果具有实际意义,强调了适当的预处理技术对提高分类性能的重要性。未来的研究可以探索更多的预处理方法及其对不同数据集分类准确性的影响,从而推动机器学习领域及其应用的发展。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信