{"title":"Multimodal Sarcasm Detection (MSD) in Videos using Deep Learning Models","authors":"Ananya Pandey, D. Vishwakarma","doi":"10.1109/APSIT58554.2023.10201731","DOIUrl":null,"url":null,"abstract":"Every day, individuals all over the globe use video-sharing websites to broadcast their thoughts, experiences, and recommendations to the world. There has been a huge rise in educational interest in the area of sarcasm detection for these opinionated videos. Although sarcasm has proven effective for written text, it remains an unexplored area of research when applied to video and other forms of multimodal data. Numerous verbal and nonverbal signs, such as a shift in voice, an overemphasis in a phrase, a stretched pronunciation, or a stiff-looking face, are often used to convey sarcasm. The majority of current research on sarcasm detection has heavily relied on textual content. Hence, in this paper, we suggest that the use of multiple modalities can help with more accurate sarcasm identification. The results of our studies on one of the multimodal sarcasm detection (MSD) benchmark datasets “MUSTARD” reveal that our strategy outperforms other algorithms in sarcasm classification.","PeriodicalId":170044,"journal":{"name":"2023 International Conference in Advances in Power, Signal, and Information Technology (APSIT)","volume":"103 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-06-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2023 International Conference in Advances in Power, Signal, and Information Technology (APSIT)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/APSIT58554.2023.10201731","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Every day, individuals all over the globe use video-sharing websites to broadcast their thoughts, experiences, and recommendations to the world. There has been a huge rise in educational interest in the area of sarcasm detection for these opinionated videos. Although sarcasm has proven effective for written text, it remains an unexplored area of research when applied to video and other forms of multimodal data. Numerous verbal and nonverbal signs, such as a shift in voice, an overemphasis in a phrase, a stretched pronunciation, or a stiff-looking face, are often used to convey sarcasm. The majority of current research on sarcasm detection has heavily relied on textual content. Hence, in this paper, we suggest that the use of multiple modalities can help with more accurate sarcasm identification. The results of our studies on one of the multimodal sarcasm detection (MSD) benchmark datasets “MUSTARD” reveal that our strategy outperforms other algorithms in sarcasm classification.