{"title":"私有云环境下基于MapReduce的数据迁移算法","authors":"A. Pandey, R. Thulasiram, A. Thavaneswaran","doi":"10.5121/CSIT.2019.90916","DOIUrl":null,"url":null,"abstract":"When a resource in a data center reaches its end-of-life, instead of investing in upgrading, it is possibly the time to decommission such a resource and migrate workloads to other resources in the data center. Data migration between different cloud servers is risky due to the possibility of data loss. The current studies in the literature do not optimize the data before migration, which could avoid data loss. MapReduce is a software framework for distributed processing of large data sets with reduced overhead of migrating data. For this study, we design a MapReduce based algorithm and introduce a few metrics to test and evaluate our proposed framework. We deploy an architecture for creating an Apache Hadoop environment for our experiments. We show that our algorithm for data migration works efficiently for text, image, audio and video files with minimum data loss and scale well for large files as well.","PeriodicalId":248929,"journal":{"name":"9th International Conference on Computer Science, Engineering and Applications (CCSEA 2019)","volume":"45 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-07-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":"{\"title\":\"A MapReduce based Algorithm for Data Migration in a Private Cloud Environment\",\"authors\":\"A. Pandey, R. Thulasiram, A. Thavaneswaran\",\"doi\":\"10.5121/CSIT.2019.90916\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"When a resource in a data center reaches its end-of-life, instead of investing in upgrading, it is possibly the time to decommission such a resource and migrate workloads to other resources in the data center. Data migration between different cloud servers is risky due to the possibility of data loss. The current studies in the literature do not optimize the data before migration, which could avoid data loss. MapReduce is a software framework for distributed processing of large data sets with reduced overhead of migrating data. For this study, we design a MapReduce based algorithm and introduce a few metrics to test and evaluate our proposed framework. We deploy an architecture for creating an Apache Hadoop environment for our experiments. We show that our algorithm for data migration works efficiently for text, image, audio and video files with minimum data loss and scale well for large files as well.\",\"PeriodicalId\":248929,\"journal\":{\"name\":\"9th International Conference on Computer Science, Engineering and Applications (CCSEA 2019)\",\"volume\":\"45 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2019-07-13\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"9th International Conference on Computer Science, Engineering and Applications (CCSEA 2019)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.5121/CSIT.2019.90916\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"9th International Conference on Computer Science, Engineering and Applications (CCSEA 2019)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.5121/CSIT.2019.90916","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
A MapReduce based Algorithm for Data Migration in a Private Cloud Environment
When a resource in a data center reaches its end-of-life, instead of investing in upgrading, it is possibly the time to decommission such a resource and migrate workloads to other resources in the data center. Data migration between different cloud servers is risky due to the possibility of data loss. The current studies in the literature do not optimize the data before migration, which could avoid data loss. MapReduce is a software framework for distributed processing of large data sets with reduced overhead of migrating data. For this study, we design a MapReduce based algorithm and introduce a few metrics to test and evaluate our proposed framework. We deploy an architecture for creating an Apache Hadoop environment for our experiments. We show that our algorithm for data migration works efficiently for text, image, audio and video files with minimum data loss and scale well for large files as well.