{"title":"Multimedia information retrieval: music and audio","authors":"M. Schedl, E. Gómez, Masataka Goto","doi":"10.1145/2502081.2502237","DOIUrl":null,"url":null,"abstract":"Music is an omnipresent topic in our daily lives, as almost everyone enjoys listening to his or her favorite tunes. Music information retrieval (MIR) is a research field that aims – among other things – at automatically extracting semantically meaningful information from various representations of music entities, such as a digital audio file, a band’s web page, a song’s lyrics, or a tweet about a microblogger’s current listening activity. A key approach in MIR is to describe music via computational features, which can be categorized into: music content, music context, and user context. The music content refers to features extracted from the audio signal, while information about musical entities not encoded in the signal (e.g., image of an artist or political background of a song) are referred to as music context. The user context, in contrast, includes environmental aspects as well as physical and mental activities of the music listener. MIR research has been seeing a paradigm shift over the last couple of years, as an increasing number of recent approaches and commercial technologies combine content-based techniques (focusing on the audio signal) with multimedia context data mined, e.g. from web sources and with user context information.","PeriodicalId":20448,"journal":{"name":"Proceedings of the 21st ACM international conference on Multimedia","volume":"13 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2013-10-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"6","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 21st ACM international conference on Multimedia","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/2502081.2502237","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 6
Abstract
Music is an omnipresent topic in our daily lives, as almost everyone enjoys listening to his or her favorite tunes. Music information retrieval (MIR) is a research field that aims – among other things – at automatically extracting semantically meaningful information from various representations of music entities, such as a digital audio file, a band’s web page, a song’s lyrics, or a tweet about a microblogger’s current listening activity. A key approach in MIR is to describe music via computational features, which can be categorized into: music content, music context, and user context. The music content refers to features extracted from the audio signal, while information about musical entities not encoded in the signal (e.g., image of an artist or political background of a song) are referred to as music context. The user context, in contrast, includes environmental aspects as well as physical and mental activities of the music listener. MIR research has been seeing a paradigm shift over the last couple of years, as an increasing number of recent approaches and commercial technologies combine content-based techniques (focusing on the audio signal) with multimedia context data mined, e.g. from web sources and with user context information.
音乐在我们的日常生活中是一个无处不在的话题,因为几乎每个人都喜欢听他或她最喜欢的曲调。音乐信息检索(Music information retrieval, MIR)是一个研究领域,其目的之一是自动从音乐实体的各种表示中提取语义上有意义的信息,例如数字音频文件、乐队的网页、歌曲的歌词,或者关于微博用户当前收听活动的tweet。MIR中的一个关键方法是通过计算特征来描述音乐,这些特征可以分为:音乐内容、音乐上下文和用户上下文。音乐内容是指从音频信号中提取的特征,而未在信号中编码的关于音乐实体的信息(例如,艺术家的图像或歌曲的政治背景)称为音乐上下文。相比之下,用户环境包括环境方面以及音乐听众的身体和心理活动。在过去的几年里,随着越来越多的最新方法和商业技术将基于内容的技术(专注于音频信号)与多媒体上下文数据(例如,从web来源和用户上下文信息)相结合,MIR研究已经看到了范式的转变。