{"title":"Model Efficiency Improvement with Simplified Architectures","authors":"Jing-Ming Guo, Ting-Yu Chang, Chun-Wei Huang","doi":"10.1109/ICCE59016.2024.10444337","DOIUrl":null,"url":null,"abstract":"In previous lightweight methods, whether in the design of lightweight model architectures or in model compression and acceleration techniques, most design approaches focused on reducing the number of parameters as the primary objective. However, reducing the number of parameters does not always guarantee improved acceleration. Sometimes, the two factors are complementary, and during the process of reducing parameters, there is a simultaneous decrease in accuracy. Therefore, achieving a balance between efficiency and accuracy has always been an important issue in lightweight methods. This paper employs a different approach than previous model compression and acceleration methods. Instead of continuously reducing the number of parameters to achieve faster model inference, it optimizes the algorithm of the model inference stage and the overall complexity of the architecture. This approach ensures that the model does not incur additional inference time costs while maintaining the original architectural performance, thereby improving model inference speed without sacrificing accuracy.","PeriodicalId":518694,"journal":{"name":"2024 IEEE International Conference on Consumer Electronics (ICCE)","volume":"66 6","pages":"1-4"},"PeriodicalIF":0.0000,"publicationDate":"2024-01-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2024 IEEE International Conference on Consumer Electronics (ICCE)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICCE59016.2024.10444337","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
In previous lightweight methods, whether in the design of lightweight model architectures or in model compression and acceleration techniques, most design approaches focused on reducing the number of parameters as the primary objective. However, reducing the number of parameters does not always guarantee improved acceleration. Sometimes, the two factors are complementary, and during the process of reducing parameters, there is a simultaneous decrease in accuracy. Therefore, achieving a balance between efficiency and accuracy has always been an important issue in lightweight methods. This paper employs a different approach than previous model compression and acceleration methods. Instead of continuously reducing the number of parameters to achieve faster model inference, it optimizes the algorithm of the model inference stage and the overall complexity of the architecture. This approach ensures that the model does not incur additional inference time costs while maintaining the original architectural performance, thereby improving model inference speed without sacrificing accuracy.