用于训练深度神经网络的可扩展gpu支持框架

Bonaventura Del Monte, R. Prodan
{"title":"用于训练深度神经网络的可扩展gpu支持框架","authors":"Bonaventura Del Monte, R. Prodan","doi":"10.1109/ICGHPC.2016.7508071","DOIUrl":null,"url":null,"abstract":"In the last fifteen years, Big Data created a new generation of data analysis problems, which does not only involve the problems themselves but also the way these data are handled. Since managing terabytes of data without a proper infrastructure is unfeasible, a smart way to process these data is also necessary. A solution to this aspect deals with the creation of general algorithms that learn from observations. In this context, Deep Learning promises general, powerful, and fast machine learning algorithms, moving them one step closer to artificial intelligence. Nevertheless, fitting a deep learning model may require an huge amount of time, thus, the need of scalable infrastructures for processing large scale data sets has become ever more meaningful. In this paper, we present a framework for training these deep neural networks using heterogeneous computing resources of either grid or cloud infrastructures. The framework lets the end-users define the deep architecture they need for processing their own Big Data, while dealing with the execution of the learning algorithms on a distributed set of nodes (through Apache Flink) as well as with offloading the computation on multiple Graphics Processing Units.","PeriodicalId":268630,"journal":{"name":"2016 2nd International Conference on Green High Performance Computing (ICGHPC)","volume":"31 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2016-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"4","resultStr":"{\"title\":\"A scalable GPU-enabled framework for training deep neural networks\",\"authors\":\"Bonaventura Del Monte, R. Prodan\",\"doi\":\"10.1109/ICGHPC.2016.7508071\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"In the last fifteen years, Big Data created a new generation of data analysis problems, which does not only involve the problems themselves but also the way these data are handled. Since managing terabytes of data without a proper infrastructure is unfeasible, a smart way to process these data is also necessary. A solution to this aspect deals with the creation of general algorithms that learn from observations. In this context, Deep Learning promises general, powerful, and fast machine learning algorithms, moving them one step closer to artificial intelligence. Nevertheless, fitting a deep learning model may require an huge amount of time, thus, the need of scalable infrastructures for processing large scale data sets has become ever more meaningful. In this paper, we present a framework for training these deep neural networks using heterogeneous computing resources of either grid or cloud infrastructures. The framework lets the end-users define the deep architecture they need for processing their own Big Data, while dealing with the execution of the learning algorithms on a distributed set of nodes (through Apache Flink) as well as with offloading the computation on multiple Graphics Processing Units.\",\"PeriodicalId\":268630,\"journal\":{\"name\":\"2016 2nd International Conference on Green High Performance Computing (ICGHPC)\",\"volume\":\"31 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2016-02-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"4\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2016 2nd International Conference on Green High Performance Computing (ICGHPC)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICGHPC.2016.7508071\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2016 2nd International Conference on Green High Performance Computing (ICGHPC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICGHPC.2016.7508071","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 4

摘要

在过去的15年里,大数据创造了新一代的数据分析问题,这些问题不仅涉及问题本身,还涉及处理这些数据的方式。由于在没有适当基础设施的情况下管理tb级数据是不可行的,因此还需要一种处理这些数据的智能方法。这方面的解决方案涉及创建从观察中学习的通用算法。在这种情况下,深度学习承诺通用、强大、快速的机器学习算法,使它们更接近人工智能。然而,拟合深度学习模型可能需要大量的时间,因此,对处理大规模数据集的可扩展基础设施的需求变得越来越有意义。在本文中,我们提出了一个使用网格或云基础设施的异构计算资源来训练这些深度神经网络的框架。该框架允许最终用户定义处理自己的大数据所需的深度架构,同时在分布式节点集(通过Apache Flink)上处理学习算法的执行,以及在多个图形处理单元上卸载计算。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
A scalable GPU-enabled framework for training deep neural networks
In the last fifteen years, Big Data created a new generation of data analysis problems, which does not only involve the problems themselves but also the way these data are handled. Since managing terabytes of data without a proper infrastructure is unfeasible, a smart way to process these data is also necessary. A solution to this aspect deals with the creation of general algorithms that learn from observations. In this context, Deep Learning promises general, powerful, and fast machine learning algorithms, moving them one step closer to artificial intelligence. Nevertheless, fitting a deep learning model may require an huge amount of time, thus, the need of scalable infrastructures for processing large scale data sets has become ever more meaningful. In this paper, we present a framework for training these deep neural networks using heterogeneous computing resources of either grid or cloud infrastructures. The framework lets the end-users define the deep architecture they need for processing their own Big Data, while dealing with the execution of the learning algorithms on a distributed set of nodes (through Apache Flink) as well as with offloading the computation on multiple Graphics Processing Units.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信