{"title":"Deep multimodal fusion for video game age rating classification","authors":"Caner BALIM","doi":"10.1016/j.entcom.2025.100980","DOIUrl":null,"url":null,"abstract":"<div><div>Video games appeal to a wide range of ages, from children to adults. As a result, reliable age rating systems like the Entertainment Software Rating Board (ESRB) and Pan European Game Information (PEGI) are essential for guarding younger gamers from improper content. These organizations rate games based on content submitted by video game developers. This paper proposes a multimodal deep learning framework that predicts age ratings by analyzing both video game cover images and textual descriptions. A dataset of 39,212 games was constructed using publicly available information, including ESRB and PEGI labels. Both individual models based on visual or textual features and fusion models that combine these modalities using simple concatenation and Deep Canonical Correlation Analysis (DCCA) were employed to perform the classification task. Experimental results indicate that the simple concatenation model achieves the highest accuracy compared to the individual modalities and the DCCA-based approach, reaching 0.678 for ESRB prediction and 0.584 for PEGI prediction. The findings highlight that using only visual information has limitations, and that textual descriptions play an important role in determining the appropriate age rating for a game. This study shows that future research can benefit from using additional content like gameplay videos and audio.</div></div>","PeriodicalId":55997,"journal":{"name":"Entertainment Computing","volume":"55 ","pages":"Article 100980"},"PeriodicalIF":2.8000,"publicationDate":"2025-06-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Entertainment Computing","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1875952125000606","RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"COMPUTER SCIENCE, CYBERNETICS","Score":null,"Total":0}
引用次数: 0
Abstract
Video games appeal to a wide range of ages, from children to adults. As a result, reliable age rating systems like the Entertainment Software Rating Board (ESRB) and Pan European Game Information (PEGI) are essential for guarding younger gamers from improper content. These organizations rate games based on content submitted by video game developers. This paper proposes a multimodal deep learning framework that predicts age ratings by analyzing both video game cover images and textual descriptions. A dataset of 39,212 games was constructed using publicly available information, including ESRB and PEGI labels. Both individual models based on visual or textual features and fusion models that combine these modalities using simple concatenation and Deep Canonical Correlation Analysis (DCCA) were employed to perform the classification task. Experimental results indicate that the simple concatenation model achieves the highest accuracy compared to the individual modalities and the DCCA-based approach, reaching 0.678 for ESRB prediction and 0.584 for PEGI prediction. The findings highlight that using only visual information has limitations, and that textual descriptions play an important role in determining the appropriate age rating for a game. This study shows that future research can benefit from using additional content like gameplay videos and audio.
期刊介绍:
Entertainment Computing publishes original, peer-reviewed research articles and serves as a forum for stimulating and disseminating innovative research ideas, emerging technologies, empirical investigations, state-of-the-art methods and tools in all aspects of digital entertainment, new media, entertainment computing, gaming, robotics, toys and applications among researchers, engineers, social scientists, artists and practitioners. Theoretical, technical, empirical, survey articles and case studies are all appropriate to the journal.