基于tda的低位权神经网络性能分析

IF 0.8 Q4 ROBOTICS
Yugo Ogio, Naoki Tsubone, Yuki Minami, Masato Ishikawa
{"title":"基于tda的低位权神经网络性能分析","authors":"Yugo Ogio,&nbsp;Naoki Tsubone,&nbsp;Yuki Minami,&nbsp;Masato Ishikawa","doi":"10.1007/s10015-025-01005-5","DOIUrl":null,"url":null,"abstract":"<div><p>Advances in neural network (NN) models and learning methods have resulted in breakthroughs in various fields. A larger NN model is more difficult to install on a computer with limited computing resources. One method for compressing NN models is to quantize the weights, in which the connection weights of the NNs are approximated with low-bit precision. The existing quantization methods for NN models can be categorized into two approaches: quantization-aware training (QAT) and post-training quantization (PTQ). In this study, we focused on the performance degradation of NN models using PTQ. This paper proposes a method for visually evaluating the performance of quantized NNs using topological data analysis (TDA). Subjecting the structure of NNs to TDA allows the performance of quantized NNs to be assessed without experiments or simulations. We developed a TDA-based evaluation method for NNs with low-bit weights by referring to previous research on a TDA-based evaluation method for NNs with high-bit weights. We also tested the TDA-based method using the MNIST dataset. Finally, we compared the performance of the quantized NNs generated by static and dynamic quantization through a visual demonstration.</p></div>","PeriodicalId":46050,"journal":{"name":"Artificial Life and Robotics","volume":"30 2","pages":"332 - 341"},"PeriodicalIF":0.8000,"publicationDate":"2025-02-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"A TDA-based performance analysis for neural networks with low-bit weights\",\"authors\":\"Yugo Ogio,&nbsp;Naoki Tsubone,&nbsp;Yuki Minami,&nbsp;Masato Ishikawa\",\"doi\":\"10.1007/s10015-025-01005-5\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><p>Advances in neural network (NN) models and learning methods have resulted in breakthroughs in various fields. A larger NN model is more difficult to install on a computer with limited computing resources. One method for compressing NN models is to quantize the weights, in which the connection weights of the NNs are approximated with low-bit precision. The existing quantization methods for NN models can be categorized into two approaches: quantization-aware training (QAT) and post-training quantization (PTQ). In this study, we focused on the performance degradation of NN models using PTQ. This paper proposes a method for visually evaluating the performance of quantized NNs using topological data analysis (TDA). Subjecting the structure of NNs to TDA allows the performance of quantized NNs to be assessed without experiments or simulations. We developed a TDA-based evaluation method for NNs with low-bit weights by referring to previous research on a TDA-based evaluation method for NNs with high-bit weights. We also tested the TDA-based method using the MNIST dataset. Finally, we compared the performance of the quantized NNs generated by static and dynamic quantization through a visual demonstration.</p></div>\",\"PeriodicalId\":46050,\"journal\":{\"name\":\"Artificial Life and Robotics\",\"volume\":\"30 2\",\"pages\":\"332 - 341\"},\"PeriodicalIF\":0.8000,\"publicationDate\":\"2025-02-12\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Artificial Life and Robotics\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://link.springer.com/article/10.1007/s10015-025-01005-5\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q4\",\"JCRName\":\"ROBOTICS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Artificial Life and Robotics","FirstCategoryId":"1085","ListUrlMain":"https://link.springer.com/article/10.1007/s10015-025-01005-5","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q4","JCRName":"ROBOTICS","Score":null,"Total":0}
引用次数: 0

摘要

神经网络(NN)模型和学习方法的进步导致了各个领域的突破。在计算资源有限的计算机上,更大的神经网络模型更难安装。一种压缩神经网络模型的方法是量化权重,其中神经网络的连接权重以低比特精度近似。现有的神经网络模型量化方法可分为量化感知训练(QAT)和训练后量化(PTQ)两种。在这项研究中,我们关注的是使用PTQ的神经网络模型的性能退化。本文提出了一种利用拓扑数据分析(TDA)直观评价量化神经网络性能的方法。将神经网络的结构置于TDA下,可以在没有实验或模拟的情况下评估量化神经网络的性能。参考前人关于基于tda的高比特权重神经网络评价方法的研究,我们开发了一种基于tda的低比特权重神经网络评价方法。我们还使用MNIST数据集测试了基于tda的方法。最后,我们通过视觉演示比较了静态和动态量化生成的量化神经网络的性能。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
A TDA-based performance analysis for neural networks with low-bit weights

Advances in neural network (NN) models and learning methods have resulted in breakthroughs in various fields. A larger NN model is more difficult to install on a computer with limited computing resources. One method for compressing NN models is to quantize the weights, in which the connection weights of the NNs are approximated with low-bit precision. The existing quantization methods for NN models can be categorized into two approaches: quantization-aware training (QAT) and post-training quantization (PTQ). In this study, we focused on the performance degradation of NN models using PTQ. This paper proposes a method for visually evaluating the performance of quantized NNs using topological data analysis (TDA). Subjecting the structure of NNs to TDA allows the performance of quantized NNs to be assessed without experiments or simulations. We developed a TDA-based evaluation method for NNs with low-bit weights by referring to previous research on a TDA-based evaluation method for NNs with high-bit weights. We also tested the TDA-based method using the MNIST dataset. Finally, we compared the performance of the quantized NNs generated by static and dynamic quantization through a visual demonstration.

求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
CiteScore
2.00
自引率
22.20%
发文量
101
期刊介绍: Artificial Life and Robotics is an international journal publishing original technical papers and authoritative state-of-the-art reviews on the development of new technologies concerning artificial life and robotics, especially computer-based simulation and hardware for the twenty-first century. This journal covers a broad multidisciplinary field, including areas such as artificial brain research, artificial intelligence, artificial life, artificial living, artificial mind research, brain science, chaos, cognitive science, complexity, computer graphics, evolutionary computations, fuzzy control, genetic algorithms, innovative computations, intelligent control and modelling, micromachines, micro-robot world cup soccer tournament, mobile vehicles, neural networks, neurocomputers, neurocomputing technologies and applications, robotics, robus virtual engineering, and virtual reality. Hardware-oriented submissions are particularly welcome. Publishing body: International Symposium on Artificial Life and RoboticsEditor-in-Chiei: Hiroshi Tanaka Hatanaka R Apartment 101, Hatanaka 8-7A, Ooaza-Hatanaka, Oita city, Oita, Japan 870-0856 ©International Symposium on Artificial Life and Robotics
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信