Accelerating Convolutional Neural Network Training for Colon Histopathology Images by Customizing Deep Learning Framework

Toto Haryanto, H. Suhartanto, A. M. Arymurthy, K. Kusmardi
{"title":"Accelerating Convolutional Neural Network Training for Colon Histopathology Images by Customizing Deep Learning Framework","authors":"Toto Haryanto, H. Suhartanto, A. M. Arymurthy, K. Kusmardi","doi":"10.2991/aisr.k.200424.063","DOIUrl":null,"url":null,"abstract":"Cancer diagnose based on the histopathology images is still have some challenges. Convolutional Neural Network (CNN) is one of deep learning architecture that has widely used in medical image processing especially for cancer detection. The high resolution of images and complexity of CNN architecture causes cost-intensive in the training process. One way of reducing the training processes time is by introducing parallel processing. Graphics Processing Unit (GPU) is a graphics card which has many processors and has been widely used to speed-up the process. However, the problem in GPU is the limitation of memory size. Therefore, this study proposes alternative ways to utilize the GPU memory in the training of CNN architecture. Theano is one of middle-level framework for deep application. GPU memory is a critical task in training activity and will affect to the number of batch-size. Customizing memory allocation in Theano can be conducted by utilizing library called ‘cnmem’. For training CNN architecture, we use NVIDIA GTX-980 that accelerated by customizing CUDA memory allocation from ‘cnmem’ library located in ‘theanorc’ file. In the experiment, the parameter of cnmem are chosen between 0 (not apply cnmem) or 1 (apply cnmem). We use image variation from 32x32, 64x64, 128x128, 180x180 and 200x200 pixels. In the training, a number of batch-size is selected experimentally from 10, 20, 50, 100 and 150 images. Our experiments show that enabling cnmem with the value of 1 will increase the speed-up. The 200x200 images show the most significant efficiency of GPU performance when training CNN. Speed-up is measured by comparing training time of GTX-980 with CPU core i7 machine from 16, 8, 4, 2 cores and the single-core. The highest speed-up GTX-980 obtained with enabling cnmem are 4.49, 5.00, 7.58, 11,97 and 16.19 compare to 16, 8, 4, 2 and 1 core processor respectively","PeriodicalId":147693,"journal":{"name":"Proceedings of the Sriwijaya International Conference on Information Technology and Its Applications (SICONIAN 2019)","volume":"22 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the Sriwijaya International Conference on Information Technology and Its Applications (SICONIAN 2019)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.2991/aisr.k.200424.063","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Cancer diagnose based on the histopathology images is still have some challenges. Convolutional Neural Network (CNN) is one of deep learning architecture that has widely used in medical image processing especially for cancer detection. The high resolution of images and complexity of CNN architecture causes cost-intensive in the training process. One way of reducing the training processes time is by introducing parallel processing. Graphics Processing Unit (GPU) is a graphics card which has many processors and has been widely used to speed-up the process. However, the problem in GPU is the limitation of memory size. Therefore, this study proposes alternative ways to utilize the GPU memory in the training of CNN architecture. Theano is one of middle-level framework for deep application. GPU memory is a critical task in training activity and will affect to the number of batch-size. Customizing memory allocation in Theano can be conducted by utilizing library called ‘cnmem’. For training CNN architecture, we use NVIDIA GTX-980 that accelerated by customizing CUDA memory allocation from ‘cnmem’ library located in ‘theanorc’ file. In the experiment, the parameter of cnmem are chosen between 0 (not apply cnmem) or 1 (apply cnmem). We use image variation from 32x32, 64x64, 128x128, 180x180 and 200x200 pixels. In the training, a number of batch-size is selected experimentally from 10, 20, 50, 100 and 150 images. Our experiments show that enabling cnmem with the value of 1 will increase the speed-up. The 200x200 images show the most significant efficiency of GPU performance when training CNN. Speed-up is measured by comparing training time of GTX-980 with CPU core i7 machine from 16, 8, 4, 2 cores and the single-core. The highest speed-up GTX-980 obtained with enabling cnmem are 4.49, 5.00, 7.58, 11,97 and 16.19 compare to 16, 8, 4, 2 and 1 core processor respectively
自定义深度学习框架加速结肠组织病理学图像卷积神经网络训练
基于组织病理图像的肿瘤诊断仍存在一些挑战。卷积神经网络(Convolutional Neural Network, CNN)是一种深度学习架构,在医学图像处理特别是癌症检测中得到了广泛的应用。由于图像的高分辨率和CNN架构的复杂性,使得训练过程中的成本非常高。减少训练过程时间的一种方法是引入并行处理。图形处理单元(Graphics Processing Unit, GPU)是一种具有多个处理器的图形卡,被广泛用于加快处理速度。然而,GPU的问题是内存大小的限制。因此,本研究提出了在CNN架构训练中利用GPU内存的替代方法。该框架是面向深度应用的中层框架之一。GPU内存是训练活动中的一个关键任务,它会影响到批处理的数量。在Theano中,可以通过使用名为“cnmem”的库来定制内存分配。对于训练CNN架构,我们使用NVIDIA GTX-980,通过从位于theanorc文件中的“cnmem”库中定制CUDA内存分配来加速。在实验中,cnmem的参数在0(不应用cnmem)和1(应用cnmem)之间选择。我们使用32x32、64x64、128x128、180x180和200x200像素的图像变化。在训练中,从10张、20张、50张、100张和150张图像中实验选择了若干批大小。我们的实验表明,启用值为1的cnmem将增加加速。在训练CNN时,200x200的图像显示出最显著的GPU性能效率。通过比较GTX-980与16核、8核、4核、2核和单核的CPU i7机器的训练时间来衡量加速。与16核、8核、4核、2核和1核处理器相比,启用cnmem后GTX-980的最高加速分别为4.49、5.00、7.58、11、97和16.19
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信