M.I.N.U.E.T.: Procedural Musical Accompaniment for Textual Narratives

Mehak Maniktala, Chris Miller, Aaron Margolese-Malin, A. Jhala, Chris Martens
{"title":"M.I.N.U.E.T.: Procedural Musical Accompaniment for Textual Narratives","authors":"Mehak Maniktala, Chris Miller, Aaron Margolese-Malin, A. Jhala, Chris Martens","doi":"10.1145/3402942.3409602","DOIUrl":null,"url":null,"abstract":"Extensive research has been conducted on using procedural music generation in real-time applications such as accompaniment to musicians, visual narratives, and games. However, less attention has been paid to the enhancement of textual narratives through music. In this paper, we present Mood Into Note Using Extracted Text (MINUET), a novel system that can procedurally generate music for textual narrative segments using sentiment analysis. Textual analysis of the flow and sentiment derived from the text is used as input to condition accompanying music. Music generation systems have addressed variations through changes in sentiment. By using an ensemble predictor model to classify sentences as belonging to particular emotions, MINUET generates text-accompanying music with the goal of enhancing a reader’s experience beyond the limits of the author’s words. Music is played via the JMusic library and a set of Markov chains specific to each emotion with mood classifications evaluated via stratified 10-fold cross validation. The development of MINUET affords the reflection and analysis of features that affect the quality of generated musical accompaniment for text. It also serves as a sandbox for further evaluating sentiment-based systems on both text and music generation sides in a coherent experience of an implemented and extendable experiential artifact.","PeriodicalId":421754,"journal":{"name":"Proceedings of the 15th International Conference on the Foundations of Digital Games","volume":"4 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-09-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 15th International Conference on the Foundations of Digital Games","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3402942.3409602","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 2

Abstract

Extensive research has been conducted on using procedural music generation in real-time applications such as accompaniment to musicians, visual narratives, and games. However, less attention has been paid to the enhancement of textual narratives through music. In this paper, we present Mood Into Note Using Extracted Text (MINUET), a novel system that can procedurally generate music for textual narrative segments using sentiment analysis. Textual analysis of the flow and sentiment derived from the text is used as input to condition accompanying music. Music generation systems have addressed variations through changes in sentiment. By using an ensemble predictor model to classify sentences as belonging to particular emotions, MINUET generates text-accompanying music with the goal of enhancing a reader’s experience beyond the limits of the author’s words. Music is played via the JMusic library and a set of Markov chains specific to each emotion with mood classifications evaluated via stratified 10-fold cross validation. The development of MINUET affords the reflection and analysis of features that affect the quality of generated musical accompaniment for text. It also serves as a sandbox for further evaluating sentiment-based systems on both text and music generation sides in a coherent experience of an implemented and extendable experiential artifact.
文本叙事的程序音乐伴奏
在实时应用(如音乐家伴奏、视觉叙事和游戏)中使用程序音乐生成已经进行了广泛的研究。然而,很少有人关注通过音乐来增强文本叙事。在本文中,我们提出了使用提取文本(MINUET)将情绪转化为音符的新系统,该系统可以使用情感分析程序地为文本叙事片段生成音乐。从文本中提取的流和情感的文本分析被用作条件伴奏音乐的输入。音乐生成系统通过情绪的变化来处理变化。通过使用集成预测模型将句子分类为属于特定情感的句子,MINUET生成文本伴奏音乐,其目标是增强读者的体验,超越作者的文字限制。音乐是通过JMusic库和一组特定于每种情绪的马尔可夫链播放的,通过分层的10倍交叉验证来评估情绪分类。MINUET的发展提供了对影响文本生成音乐伴奏质量的特征的反思和分析。它还可以作为沙盒,用于进一步评估文本和音乐生成方面的基于情感的系统,以实现和可扩展的体验工件的连贯体验。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信