{"title":"Computer-generated Humour Based on GPT-2","authors":"Yuchen Su","doi":"10.1109/ICDSCA56264.2022.9987901","DOIUrl":null,"url":null,"abstract":"Humour generation has always been a huge challenge in the area of computational humour. In this paper, we explore how to generate jokes based on keywords by fine-tuning GPT-2 pre-training model based on keywords task and comparing them with the LSTM-based encoder-decoder model. We trained the model by the short jokes of Conan O'Brien and the puns dataset of Yang et al. with the help of Pos-Tagger. Then, we evaluate the humour by using human evaluation and the similarity with keywords and jokes by using automatic evaluation. In terms of the final average score, the performance of our model is better than the LSTM-based encoder-decoder model.","PeriodicalId":416983,"journal":{"name":"2022 IEEE 2nd International Conference on Data Science and Computer Application (ICDSCA)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-10-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 IEEE 2nd International Conference on Data Science and Computer Application (ICDSCA)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICDSCA56264.2022.9987901","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Humour generation has always been a huge challenge in the area of computational humour. In this paper, we explore how to generate jokes based on keywords by fine-tuning GPT-2 pre-training model based on keywords task and comparing them with the LSTM-based encoder-decoder model. We trained the model by the short jokes of Conan O'Brien and the puns dataset of Yang et al. with the help of Pos-Tagger. Then, we evaluate the humour by using human evaluation and the similarity with keywords and jokes by using automatic evaluation. In terms of the final average score, the performance of our model is better than the LSTM-based encoder-decoder model.