Sefa Aras , Elif Aras , Eyüp Gedikli , Hamdi Tolga Kahraman
{"title":"多目标CNN优化:自动化模型设计的鲁棒框架","authors":"Sefa Aras , Elif Aras , Eyüp Gedikli , Hamdi Tolga Kahraman","doi":"10.1016/j.ins.2025.122468","DOIUrl":null,"url":null,"abstract":"<div><div>Convolutional neural networks (CNNs) have demonstrated high performance in classifying image data. However, CNNs require expert-level hyperparameter tuning and involve substantial computational complexity, which hinders their effective deployment in real-time and IoT systems. Research on CNNs indicates that optimal hyperparameter configurations can enhance inference speed while improving classification accuracy. We developed a novel approach that uses multi-objective optimization to design CNN models automatically. Our method tunes hyperparameters to balance classification accuracy and inference speed. We define classification performance and inference speed as objectives and balance them using a Pareto-optimal strategy. Unlike traditional approaches, MoCNN systematically explores Pareto-optimal trade-offs between classification performance and computational efficiency, enabling a fully automated architecture search without manual intervention. In this study, we select NSGA-II as our preferred MOEA while ensuring the framework remains flexible enough to accommodate other evolutionary strategies. Experimental evaluations on benchmark datasets (CIFAR-10, CIFAR-100, and FRUITS-360) demonstrate that MoCNN reduces inference time by up to 72.02% on average and improves classification accuracy by 6.72% compared to manually tuned CNN architectures. By eliminating the need for heuristic hyperparameter selection, MoCNN enhances scalability and is particularly well suited for real-time, mobile AI, and edge-computing applications. Our results show that MoCNN outperforms state-of-the-art optimization frameworks in both computational efficiency and predictive performance, highlighting its potential for deployment in scenarios where accuracy and speed are critical.</div></div>","PeriodicalId":51063,"journal":{"name":"Information Sciences","volume":"719 ","pages":"Article 122468"},"PeriodicalIF":8.1000,"publicationDate":"2025-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Multi-objective CNN optimization: A robust framework for automated model design\",\"authors\":\"Sefa Aras , Elif Aras , Eyüp Gedikli , Hamdi Tolga Kahraman\",\"doi\":\"10.1016/j.ins.2025.122468\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Convolutional neural networks (CNNs) have demonstrated high performance in classifying image data. However, CNNs require expert-level hyperparameter tuning and involve substantial computational complexity, which hinders their effective deployment in real-time and IoT systems. Research on CNNs indicates that optimal hyperparameter configurations can enhance inference speed while improving classification accuracy. We developed a novel approach that uses multi-objective optimization to design CNN models automatically. Our method tunes hyperparameters to balance classification accuracy and inference speed. We define classification performance and inference speed as objectives and balance them using a Pareto-optimal strategy. Unlike traditional approaches, MoCNN systematically explores Pareto-optimal trade-offs between classification performance and computational efficiency, enabling a fully automated architecture search without manual intervention. In this study, we select NSGA-II as our preferred MOEA while ensuring the framework remains flexible enough to accommodate other evolutionary strategies. Experimental evaluations on benchmark datasets (CIFAR-10, CIFAR-100, and FRUITS-360) demonstrate that MoCNN reduces inference time by up to 72.02% on average and improves classification accuracy by 6.72% compared to manually tuned CNN architectures. By eliminating the need for heuristic hyperparameter selection, MoCNN enhances scalability and is particularly well suited for real-time, mobile AI, and edge-computing applications. Our results show that MoCNN outperforms state-of-the-art optimization frameworks in both computational efficiency and predictive performance, highlighting its potential for deployment in scenarios where accuracy and speed are critical.</div></div>\",\"PeriodicalId\":51063,\"journal\":{\"name\":\"Information Sciences\",\"volume\":\"719 \",\"pages\":\"Article 122468\"},\"PeriodicalIF\":8.1000,\"publicationDate\":\"2025-07-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Information Sciences\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0020025525006000\",\"RegionNum\":1,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"0\",\"JCRName\":\"COMPUTER SCIENCE, INFORMATION SYSTEMS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Information Sciences","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0020025525006000","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"0","JCRName":"COMPUTER SCIENCE, INFORMATION SYSTEMS","Score":null,"Total":0}
Multi-objective CNN optimization: A robust framework for automated model design
Convolutional neural networks (CNNs) have demonstrated high performance in classifying image data. However, CNNs require expert-level hyperparameter tuning and involve substantial computational complexity, which hinders their effective deployment in real-time and IoT systems. Research on CNNs indicates that optimal hyperparameter configurations can enhance inference speed while improving classification accuracy. We developed a novel approach that uses multi-objective optimization to design CNN models automatically. Our method tunes hyperparameters to balance classification accuracy and inference speed. We define classification performance and inference speed as objectives and balance them using a Pareto-optimal strategy. Unlike traditional approaches, MoCNN systematically explores Pareto-optimal trade-offs between classification performance and computational efficiency, enabling a fully automated architecture search without manual intervention. In this study, we select NSGA-II as our preferred MOEA while ensuring the framework remains flexible enough to accommodate other evolutionary strategies. Experimental evaluations on benchmark datasets (CIFAR-10, CIFAR-100, and FRUITS-360) demonstrate that MoCNN reduces inference time by up to 72.02% on average and improves classification accuracy by 6.72% compared to manually tuned CNN architectures. By eliminating the need for heuristic hyperparameter selection, MoCNN enhances scalability and is particularly well suited for real-time, mobile AI, and edge-computing applications. Our results show that MoCNN outperforms state-of-the-art optimization frameworks in both computational efficiency and predictive performance, highlighting its potential for deployment in scenarios where accuracy and speed are critical.
期刊介绍:
Informatics and Computer Science Intelligent Systems Applications is an esteemed international journal that focuses on publishing original and creative research findings in the field of information sciences. We also feature a limited number of timely tutorial and surveying contributions.
Our journal aims to cater to a diverse audience, including researchers, developers, managers, strategic planners, graduate students, and anyone interested in staying up-to-date with cutting-edge research in information science, knowledge engineering, and intelligent systems. While readers are expected to share a common interest in information science, they come from varying backgrounds such as engineering, mathematics, statistics, physics, computer science, cell biology, molecular biology, management science, cognitive science, neurobiology, behavioral sciences, and biochemistry.