A QKeras Neural Network Zoo for Deeply Quantized Imaging

F. Loro, D. Pau, V. Tomaselli
{"title":"A QKeras Neural Network Zoo for Deeply Quantized Imaging","authors":"F. Loro, D. Pau, V. Tomaselli","doi":"10.1109/rtsi50628.2021.9597341","DOIUrl":null,"url":null,"abstract":"Neural network zoos are quite common in the literature and are particularly useful for demonstrating the potential of any deep learning framework by providing examples of its use to the Artificial Intelligence community. Unfortunately most of them uses FP32 (32bits floating point) or INT8 (8bits integer) precision for activation and weights. Communities such as TinyML are paying more and more attention to memory and energy-saving to achieve mW and below power consumptions and therefore to Deeply Quantized Neural Networks (DQNNs). Two frameworks: QKeras and Larq, are gaining momentum for defining and training DQNNs. To best of our knowledge, the only available zoo for DQNN is the Larq framework. In this work we developed a new QKeras zoo and comparing the accuracy with the available Larq zoo. To avoid costly re-training, we show how to re-use the weights from Larq zoo. We also developed the zoo with ten networks and matched the performance of the Larq zoo for seven out of ten networks. Our work will be made publicly available through a GitHub repository.","PeriodicalId":294628,"journal":{"name":"2021 IEEE 6th International Forum on Research and Technology for Society and Industry (RTSI)","volume":"22 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-09-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 IEEE 6th International Forum on Research and Technology for Society and Industry (RTSI)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/rtsi50628.2021.9597341","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 2

Abstract

Neural network zoos are quite common in the literature and are particularly useful for demonstrating the potential of any deep learning framework by providing examples of its use to the Artificial Intelligence community. Unfortunately most of them uses FP32 (32bits floating point) or INT8 (8bits integer) precision for activation and weights. Communities such as TinyML are paying more and more attention to memory and energy-saving to achieve mW and below power consumptions and therefore to Deeply Quantized Neural Networks (DQNNs). Two frameworks: QKeras and Larq, are gaining momentum for defining and training DQNNs. To best of our knowledge, the only available zoo for DQNN is the Larq framework. In this work we developed a new QKeras zoo and comparing the accuracy with the available Larq zoo. To avoid costly re-training, we show how to re-use the weights from Larq zoo. We also developed the zoo with ten networks and matched the performance of the Larq zoo for seven out of ten networks. Our work will be made publicly available through a GitHub repository.
深度量化成像的QKeras神经网络动物园
神经网络动物园在文献中相当常见,并且通过向人工智能社区提供示例来展示任何深度学习框架的潜力特别有用。不幸的是,它们中的大多数使用FP32(32位浮点数)或INT8(8位整数)精度进行激活和权重。TinyML等社区越来越关注内存和节能,以实现mW及以下的功耗,因此深度量化神经网络(DQNNs)。两个框架:QKeras和Larq,正在获得定义和训练dqnn的动力。据我们所知,DQNN唯一可用的动物园是Larq框架。在这项工作中,我们开发了一个新的QKeras动物园,并与现有的Larq动物园进行了精度比较。为了避免昂贵的重新训练,我们展示了如何重用来自Larq动物园的权重。我们还开发了包含10个网络的动物园,并且在10个网络中有7个网络的性能与Larq动物园相当。我们的工作将通过GitHub存储库公开发布。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信