基于CUDA和消息传递接口的多GPU机器学习计算

Bhagirath, Neetu Mittal, Sushil Kumar
{"title":"基于CUDA和消息传递接口的多GPU机器学习计算","authors":"Bhagirath, Neetu Mittal, Sushil Kumar","doi":"10.1109/PEEIC47157.2019.8976714","DOIUrl":null,"url":null,"abstract":"In this paper, we provide our efforts to implement machine learning modeling on commodity hardware such as general purpose graphical processing unit (GPU) and multiple GPU's connected with message passing interface (MPI). We consider risk models that involve a large number of iterations to come up with a probability of defaults for any credit account. This is computed based on the Markov Chain analysis. We discuss data structures and efficient implementation of machine learning models on the GPU platform. Idea is to leverage the power of fast GPU RAM and thousands of GPU core for fasten the execution process and reduce overall time. When we increase the number of GPU in our experiment, it also increases the programming complexity and increase the number of I/O which leads to increase overall turnaround time. We benchmarked the scalability and performance of our implementation with respect to size of the data. Performing model computations on huge amount o.f data is a compute intensive and costly task. We purpose four combinations of CPU, GPU and MPI for machine learning modeling. Experiment on real data show that to training machine leaning model on single GPU outperform as compare to CPu, Multiple GPU and GPU connected with MPI","PeriodicalId":203504,"journal":{"name":"2019 2nd International Conference on Power Energy, Environment and Intelligent Control (PEEIC)","volume":"15 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"4","resultStr":"{\"title\":\"Machine Learning Computation on Multiple GPU's using CUDA and Message Passing Interface\",\"authors\":\"Bhagirath, Neetu Mittal, Sushil Kumar\",\"doi\":\"10.1109/PEEIC47157.2019.8976714\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"In this paper, we provide our efforts to implement machine learning modeling on commodity hardware such as general purpose graphical processing unit (GPU) and multiple GPU's connected with message passing interface (MPI). We consider risk models that involve a large number of iterations to come up with a probability of defaults for any credit account. This is computed based on the Markov Chain analysis. We discuss data structures and efficient implementation of machine learning models on the GPU platform. Idea is to leverage the power of fast GPU RAM and thousands of GPU core for fasten the execution process and reduce overall time. When we increase the number of GPU in our experiment, it also increases the programming complexity and increase the number of I/O which leads to increase overall turnaround time. We benchmarked the scalability and performance of our implementation with respect to size of the data. Performing model computations on huge amount o.f data is a compute intensive and costly task. We purpose four combinations of CPU, GPU and MPI for machine learning modeling. Experiment on real data show that to training machine leaning model on single GPU outperform as compare to CPu, Multiple GPU and GPU connected with MPI\",\"PeriodicalId\":203504,\"journal\":{\"name\":\"2019 2nd International Conference on Power Energy, Environment and Intelligent Control (PEEIC)\",\"volume\":\"15 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2019-10-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"4\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2019 2nd International Conference on Power Energy, Environment and Intelligent Control (PEEIC)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/PEEIC47157.2019.8976714\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2019 2nd International Conference on Power Energy, Environment and Intelligent Control (PEEIC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/PEEIC47157.2019.8976714","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 4

摘要

在本文中,我们提供了在商品硬件上实现机器学习建模的努力,例如通用图形处理单元(GPU)和与消息传递接口(MPI)连接的多个GPU。我们考虑涉及大量迭代的风险模型,以得出任何信用帐户违约的概率。这是基于马尔可夫链分析计算的。我们讨论了数据结构和机器学习模型在GPU平台上的有效实现。想法是利用快速GPU RAM和数千个GPU核心的力量来加快执行过程并减少总体时间。当我们在实验中增加GPU的数量时,它也增加了编程的复杂性,增加了I/O的数量,从而增加了整体的周转时间。我们根据数据大小对实现的可伸缩性和性能进行基准测试。对大量数据进行模型计算是一项计算密集型且成本高昂的任务。我们将CPU、GPU和MPI的四种组合用于机器学习建模。实际数据实验表明,在单GPU上训练机器学习模型优于CPu、多GPU和与MPI连接的GPU
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Machine Learning Computation on Multiple GPU's using CUDA and Message Passing Interface
In this paper, we provide our efforts to implement machine learning modeling on commodity hardware such as general purpose graphical processing unit (GPU) and multiple GPU's connected with message passing interface (MPI). We consider risk models that involve a large number of iterations to come up with a probability of defaults for any credit account. This is computed based on the Markov Chain analysis. We discuss data structures and efficient implementation of machine learning models on the GPU platform. Idea is to leverage the power of fast GPU RAM and thousands of GPU core for fasten the execution process and reduce overall time. When we increase the number of GPU in our experiment, it also increases the programming complexity and increase the number of I/O which leads to increase overall turnaround time. We benchmarked the scalability and performance of our implementation with respect to size of the data. Performing model computations on huge amount o.f data is a compute intensive and costly task. We purpose four combinations of CPU, GPU and MPI for machine learning modeling. Experiment on real data show that to training machine leaning model on single GPU outperform as compare to CPu, Multiple GPU and GPU connected with MPI
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信