{"title":"在各种硬件架构上利用适应性迁移学习应用 GPT-2 模型的效率分析","authors":"Dejan Dodić, Dušan Regodić","doi":"10.61837/mbuir020124174d","DOIUrl":null,"url":null,"abstract":"This paper conducts an analysis of the efficiency in implementing the GPT-2 model, one of the advanced artificial intelligence models for text generation, through adapted transfer learning, focusing particularly on the utilization of various GPU architectures. The primary goal of this research is to examine the impact of adapted transfer learning on the performance of the GPT-2 model exclusively on various GPU architectures, assessing how different GPU strengths enhance or influence the model's efficiency. The work relies on an experimental method to evaluate and compare the model's performance in terms of accuracy, processing speed, and energy efficiency on each of the tested platforms. Special attention is given to analysing how different characteristics of hardware architectures, such as processing power and memory capacity, affect the efficiency of the transfer learning process. This study provides important insights into the potential for optimizing the GPT-2 model for specific hardware platforms, which is crucial for its application in a wide range of real-world scenarios. The results of this research offer valuable information for researchers in the fields of artificial intelligence and machine learning, providing a foundation for further development and improvement of AI technologies.","PeriodicalId":514816,"journal":{"name":"MB University International Review","volume":"21 13","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-07-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"ANALYSIS OF THE EFFICIENCY OF GPT-2 MODEL APPLICATION WITH ADAPTED TRANSFER LEARNING ON VARIOUS HARDWARE ARCHITECTURES\",\"authors\":\"Dejan Dodić, Dušan Regodić\",\"doi\":\"10.61837/mbuir020124174d\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"This paper conducts an analysis of the efficiency in implementing the GPT-2 model, one of the advanced artificial intelligence models for text generation, through adapted transfer learning, focusing particularly on the utilization of various GPU architectures. The primary goal of this research is to examine the impact of adapted transfer learning on the performance of the GPT-2 model exclusively on various GPU architectures, assessing how different GPU strengths enhance or influence the model's efficiency. The work relies on an experimental method to evaluate and compare the model's performance in terms of accuracy, processing speed, and energy efficiency on each of the tested platforms. Special attention is given to analysing how different characteristics of hardware architectures, such as processing power and memory capacity, affect the efficiency of the transfer learning process. This study provides important insights into the potential for optimizing the GPT-2 model for specific hardware platforms, which is crucial for its application in a wide range of real-world scenarios. The results of this research offer valuable information for researchers in the fields of artificial intelligence and machine learning, providing a foundation for further development and improvement of AI technologies.\",\"PeriodicalId\":514816,\"journal\":{\"name\":\"MB University International Review\",\"volume\":\"21 13\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-07-25\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"MB University International Review\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.61837/mbuir020124174d\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"MB University International Review","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.61837/mbuir020124174d","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
ANALYSIS OF THE EFFICIENCY OF GPT-2 MODEL APPLICATION WITH ADAPTED TRANSFER LEARNING ON VARIOUS HARDWARE ARCHITECTURES
This paper conducts an analysis of the efficiency in implementing the GPT-2 model, one of the advanced artificial intelligence models for text generation, through adapted transfer learning, focusing particularly on the utilization of various GPU architectures. The primary goal of this research is to examine the impact of adapted transfer learning on the performance of the GPT-2 model exclusively on various GPU architectures, assessing how different GPU strengths enhance or influence the model's efficiency. The work relies on an experimental method to evaluate and compare the model's performance in terms of accuracy, processing speed, and energy efficiency on each of the tested platforms. Special attention is given to analysing how different characteristics of hardware architectures, such as processing power and memory capacity, affect the efficiency of the transfer learning process. This study provides important insights into the potential for optimizing the GPT-2 model for specific hardware platforms, which is crucial for its application in a wide range of real-world scenarios. The results of this research offer valuable information for researchers in the fields of artificial intelligence and machine learning, providing a foundation for further development and improvement of AI technologies.