{"title":"Efficient parallel Stochastic Gradient Descent for matrix factorization using GPU","authors":"Mohamed A. Nassar, Layla A. A. El-Sayed, Y. Taha","doi":"10.1109/ICITST.2016.7856668","DOIUrl":null,"url":null,"abstract":"Matrix factorization is an advanced and efficient technique for recommender systems. Recently, Stochastic Gradient Descent (SGD) method is considered to be one of the most popular techniques for matrix factorization. SGD is a sequential algorithm, which is difficult to be parallelized for large-scale problems. Nowadays, researches focus on efficiently parallelizing SGD. In this research, we propose an efficient parallel SGD method, ESGD, for GPU. ESGD is more efficient than recent parallel methods because it utilizes GPU, reducing non-coalesced access of global memory and achieving load balance of threads. In addition, ESGD does not require any sorting and/or data shuffling as preprocessing phase. Although platform used for ESGD implementation is old, ESGD demonstrates 12.5× speedup over state-of-the-art GPU method, BSGD.","PeriodicalId":258740,"journal":{"name":"2016 11th International Conference for Internet Technology and Secured Transactions (ICITST)","volume":"32 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2016-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2016 11th International Conference for Internet Technology and Secured Transactions (ICITST)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICITST.2016.7856668","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1
Abstract
Matrix factorization is an advanced and efficient technique for recommender systems. Recently, Stochastic Gradient Descent (SGD) method is considered to be one of the most popular techniques for matrix factorization. SGD is a sequential algorithm, which is difficult to be parallelized for large-scale problems. Nowadays, researches focus on efficiently parallelizing SGD. In this research, we propose an efficient parallel SGD method, ESGD, for GPU. ESGD is more efficient than recent parallel methods because it utilizes GPU, reducing non-coalesced access of global memory and achieving load balance of threads. In addition, ESGD does not require any sorting and/or data shuffling as preprocessing phase. Although platform used for ESGD implementation is old, ESGD demonstrates 12.5× speedup over state-of-the-art GPU method, BSGD.