{"title":"Toward low-complexity neural networks for failure management in optical networks","authors":"Lareb Zar Khan;Joao Pedro;Omran Ayoub;Nelson Costa;Andrea Sgambelluri;Lorenzo De Marinis;Antonio Napoli;Nicola Sambo","doi":"10.1364/JOCN.550933","DOIUrl":null,"url":null,"abstract":"Machine learning (ML) continues to show its potential and efficacy in automating network management tasks, such as failure management. However, as ML deployment considerations broaden, aspects that go beyond predictive performance, such as a model’s computational complexity (CC), start to gain significance, as higher CC incurs higher costs and energy consumption. Balancing high predictive performance with reduced CC is an important aspect, and therefore, it needs more investigation, especially in the context of optical networks. In this work, we focus on the problem of reducing the CC of ML models, specifically neural networks (NNs), for the use case of failure identification in optical networks. We propose an approach that exploits the relative activity of neurons in NNs to reduce their size (and hence, their CC). Our proposed approach, referred to as iterative neural removal (INR), iteratively computes neurons’ activity and removes neurons with no activity until reaching a predefined stopping condition. We also propose another approach, referred to as guided knowledge distillation (GKD), that combines INR with knowledge distillation (KD), a known technique for compression of NNs. GKD inherently determines the size of the compressed NN without requiring any manual suboptimal selection or other time-consuming optimization strategies, as in traditional KD. To quantify the effectiveness of INR and GKD, we evaluate their performance against pruning (i.e., a well-known NN compression technique) in terms of impact on predictive performance and reduction in CC and memory footprint. For the considered scenario, experimental results on testbed data show that INR and GKD are more effective than pruning in reducing CC and memory footprint.","PeriodicalId":50103,"journal":{"name":"Journal of Optical Communications and Networking","volume":"17 7","pages":"555-563"},"PeriodicalIF":4.0000,"publicationDate":"2025-06-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Optical Communications and Networking","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/11027919/","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, HARDWARE & ARCHITECTURE","Score":null,"Total":0}
引用次数: 0
Abstract
Machine learning (ML) continues to show its potential and efficacy in automating network management tasks, such as failure management. However, as ML deployment considerations broaden, aspects that go beyond predictive performance, such as a model’s computational complexity (CC), start to gain significance, as higher CC incurs higher costs and energy consumption. Balancing high predictive performance with reduced CC is an important aspect, and therefore, it needs more investigation, especially in the context of optical networks. In this work, we focus on the problem of reducing the CC of ML models, specifically neural networks (NNs), for the use case of failure identification in optical networks. We propose an approach that exploits the relative activity of neurons in NNs to reduce their size (and hence, their CC). Our proposed approach, referred to as iterative neural removal (INR), iteratively computes neurons’ activity and removes neurons with no activity until reaching a predefined stopping condition. We also propose another approach, referred to as guided knowledge distillation (GKD), that combines INR with knowledge distillation (KD), a known technique for compression of NNs. GKD inherently determines the size of the compressed NN without requiring any manual suboptimal selection or other time-consuming optimization strategies, as in traditional KD. To quantify the effectiveness of INR and GKD, we evaluate their performance against pruning (i.e., a well-known NN compression technique) in terms of impact on predictive performance and reduction in CC and memory footprint. For the considered scenario, experimental results on testbed data show that INR and GKD are more effective than pruning in reducing CC and memory footprint.
期刊介绍:
The scope of the Journal includes advances in the state-of-the-art of optical networking science, technology, and engineering. Both theoretical contributions (including new techniques, concepts, analyses, and economic studies) and practical contributions (including optical networking experiments, prototypes, and new applications) are encouraged. Subareas of interest include the architecture and design of optical networks, optical network survivability and security, software-defined optical networking, elastic optical networks, data and control plane advances, network management related innovation, and optical access networks. Enabling technologies and their applications are suitable topics only if the results are shown to directly impact optical networking beyond simple point-to-point networks.