{"title":"用于训练深度神经网络的可扩展gpu支持框架","authors":"Bonaventura Del Monte, R. Prodan","doi":"10.1109/ICGHPC.2016.7508071","DOIUrl":null,"url":null,"abstract":"In the last fifteen years, Big Data created a new generation of data analysis problems, which does not only involve the problems themselves but also the way these data are handled. Since managing terabytes of data without a proper infrastructure is unfeasible, a smart way to process these data is also necessary. A solution to this aspect deals with the creation of general algorithms that learn from observations. In this context, Deep Learning promises general, powerful, and fast machine learning algorithms, moving them one step closer to artificial intelligence. Nevertheless, fitting a deep learning model may require an huge amount of time, thus, the need of scalable infrastructures for processing large scale data sets has become ever more meaningful. In this paper, we present a framework for training these deep neural networks using heterogeneous computing resources of either grid or cloud infrastructures. The framework lets the end-users define the deep architecture they need for processing their own Big Data, while dealing with the execution of the learning algorithms on a distributed set of nodes (through Apache Flink) as well as with offloading the computation on multiple Graphics Processing Units.","PeriodicalId":268630,"journal":{"name":"2016 2nd International Conference on Green High Performance Computing (ICGHPC)","volume":"31 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2016-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"4","resultStr":"{\"title\":\"A scalable GPU-enabled framework for training deep neural networks\",\"authors\":\"Bonaventura Del Monte, R. Prodan\",\"doi\":\"10.1109/ICGHPC.2016.7508071\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"In the last fifteen years, Big Data created a new generation of data analysis problems, which does not only involve the problems themselves but also the way these data are handled. Since managing terabytes of data without a proper infrastructure is unfeasible, a smart way to process these data is also necessary. A solution to this aspect deals with the creation of general algorithms that learn from observations. In this context, Deep Learning promises general, powerful, and fast machine learning algorithms, moving them one step closer to artificial intelligence. Nevertheless, fitting a deep learning model may require an huge amount of time, thus, the need of scalable infrastructures for processing large scale data sets has become ever more meaningful. In this paper, we present a framework for training these deep neural networks using heterogeneous computing resources of either grid or cloud infrastructures. The framework lets the end-users define the deep architecture they need for processing their own Big Data, while dealing with the execution of the learning algorithms on a distributed set of nodes (through Apache Flink) as well as with offloading the computation on multiple Graphics Processing Units.\",\"PeriodicalId\":268630,\"journal\":{\"name\":\"2016 2nd International Conference on Green High Performance Computing (ICGHPC)\",\"volume\":\"31 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2016-02-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"4\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2016 2nd International Conference on Green High Performance Computing (ICGHPC)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICGHPC.2016.7508071\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2016 2nd International Conference on Green High Performance Computing (ICGHPC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICGHPC.2016.7508071","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
A scalable GPU-enabled framework for training deep neural networks
In the last fifteen years, Big Data created a new generation of data analysis problems, which does not only involve the problems themselves but also the way these data are handled. Since managing terabytes of data without a proper infrastructure is unfeasible, a smart way to process these data is also necessary. A solution to this aspect deals with the creation of general algorithms that learn from observations. In this context, Deep Learning promises general, powerful, and fast machine learning algorithms, moving them one step closer to artificial intelligence. Nevertheless, fitting a deep learning model may require an huge amount of time, thus, the need of scalable infrastructures for processing large scale data sets has become ever more meaningful. In this paper, we present a framework for training these deep neural networks using heterogeneous computing resources of either grid or cloud infrastructures. The framework lets the end-users define the deep architecture they need for processing their own Big Data, while dealing with the execution of the learning algorithms on a distributed set of nodes (through Apache Flink) as well as with offloading the computation on multiple Graphics Processing Units.