{"title":"基于视觉的作曲家分类音乐数据表示","authors":"S. Deepaisarn, Suphachok Buaruk, Sirawit Chokphantavee, Sorawit Chokphantavee, Phuriphan Prathipasen, Virach Sornlertlamvanich","doi":"10.1109/iSAI-NLP56921.2022.9960254","DOIUrl":null,"url":null,"abstract":"Automated classification for musical genres and composers is an artificial intelligence research challenge insofar as music lacks a rigidly defined structure and may result in varied interpretations by individuals. This research collected acoustic features from a sizable musical database to create an image dataset for formulating a classification model. Each image was constructed by combining pitch, temporal index length, and additional incorporated features of velocity, onset, duration, and a combination of the three. Incorporated features underwent Sigmoid scaling, creating a novel visual-based music representation. A deep learning framework, fast.ai, was used as the primary classification instrument for generated images. The results were that using velocity solely as an incorporated feature provides optimal performance, with an F1-score of 0.85 using the ResN$e$t34 model. These findings offer preliminary insight into composer classification for heightening understanding of music composer signature characterizations.","PeriodicalId":399019,"journal":{"name":"2022 17th International Joint Symposium on Artificial Intelligence and Natural Language Processing (iSAI-NLP)","volume":"35 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":"{\"title\":\"Visual-based Musical Data Representation for Composer Classification\",\"authors\":\"S. Deepaisarn, Suphachok Buaruk, Sirawit Chokphantavee, Sorawit Chokphantavee, Phuriphan Prathipasen, Virach Sornlertlamvanich\",\"doi\":\"10.1109/iSAI-NLP56921.2022.9960254\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Automated classification for musical genres and composers is an artificial intelligence research challenge insofar as music lacks a rigidly defined structure and may result in varied interpretations by individuals. This research collected acoustic features from a sizable musical database to create an image dataset for formulating a classification model. Each image was constructed by combining pitch, temporal index length, and additional incorporated features of velocity, onset, duration, and a combination of the three. Incorporated features underwent Sigmoid scaling, creating a novel visual-based music representation. A deep learning framework, fast.ai, was used as the primary classification instrument for generated images. The results were that using velocity solely as an incorporated feature provides optimal performance, with an F1-score of 0.85 using the ResN$e$t34 model. These findings offer preliminary insight into composer classification for heightening understanding of music composer signature characterizations.\",\"PeriodicalId\":399019,\"journal\":{\"name\":\"2022 17th International Joint Symposium on Artificial Intelligence and Natural Language Processing (iSAI-NLP)\",\"volume\":\"35 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2022-11-05\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2022 17th International Joint Symposium on Artificial Intelligence and Natural Language Processing (iSAI-NLP)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/iSAI-NLP56921.2022.9960254\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 17th International Joint Symposium on Artificial Intelligence and Natural Language Processing (iSAI-NLP)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/iSAI-NLP56921.2022.9960254","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Visual-based Musical Data Representation for Composer Classification
Automated classification for musical genres and composers is an artificial intelligence research challenge insofar as music lacks a rigidly defined structure and may result in varied interpretations by individuals. This research collected acoustic features from a sizable musical database to create an image dataset for formulating a classification model. Each image was constructed by combining pitch, temporal index length, and additional incorporated features of velocity, onset, duration, and a combination of the three. Incorporated features underwent Sigmoid scaling, creating a novel visual-based music representation. A deep learning framework, fast.ai, was used as the primary classification instrument for generated images. The results were that using velocity solely as an incorporated feature provides optimal performance, with an F1-score of 0.85 using the ResN$e$t34 model. These findings offer preliminary insight into composer classification for heightening understanding of music composer signature characterizations.