Music visualization based on the MIDI specifications for multidimensional polyphonic expression

Z. Xie, Yu Chen
{"title":"Music visualization based on the MIDI specifications for multidimensional polyphonic expression","authors":"Z. Xie, Yu Chen","doi":"10.1109/ICSAI.2017.8248478","DOIUrl":null,"url":null,"abstract":"Music visualization is a branch of information visualization. First, the advantages of MIDI specifications for multidimensional polyphonic expression in “channel per note” were analyzed. Second, the new music visualization mapping model was proposed, which was based on the structural similarity of human and computer recognition in sound and image. The model corresponded to pitch bend, control changes, velocity to hue, saturation, and value respectively. Third, the scheme of music visualization was proposed. The base image was obtained by improving the brightness of original image by 20%–30%; then the influence image was divided into three parts as the left, the middle and the right solid color area, which corresponded to the bass voice, the alto voice and the treble voice respectively; next, music information from different areas were mapped to calculate the HSV values of solid color and then the HSV values were converted to the RGB values; finally, the RGB values of these three parts were glued together and combined with the base image in multiply mode to obtain the final image. Fourth, taking advantage of Cycling'74 Max/Jitter 7.3.4 and optimizing image combination calculation, five experiments were completed under the circumstances of bass voice, alto voice, treble voice, two notes shown in the same area and two notes shown in two different areas. Finally, by calculation of Jit.fpsgui, the final frame rate of the image was around 80fps if RGB values were glued every 16 milliseconds. The research shows that this method can meet real-time needs of music visualization and it can be used for live shows.","PeriodicalId":285726,"journal":{"name":"2017 4th International Conference on Systems and Informatics (ICSAI)","volume":"50 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2017-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2017 4th International Conference on Systems and Informatics (ICSAI)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICSAI.2017.8248478","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Music visualization is a branch of information visualization. First, the advantages of MIDI specifications for multidimensional polyphonic expression in “channel per note” were analyzed. Second, the new music visualization mapping model was proposed, which was based on the structural similarity of human and computer recognition in sound and image. The model corresponded to pitch bend, control changes, velocity to hue, saturation, and value respectively. Third, the scheme of music visualization was proposed. The base image was obtained by improving the brightness of original image by 20%–30%; then the influence image was divided into three parts as the left, the middle and the right solid color area, which corresponded to the bass voice, the alto voice and the treble voice respectively; next, music information from different areas were mapped to calculate the HSV values of solid color and then the HSV values were converted to the RGB values; finally, the RGB values of these three parts were glued together and combined with the base image in multiply mode to obtain the final image. Fourth, taking advantage of Cycling'74 Max/Jitter 7.3.4 and optimizing image combination calculation, five experiments were completed under the circumstances of bass voice, alto voice, treble voice, two notes shown in the same area and two notes shown in two different areas. Finally, by calculation of Jit.fpsgui, the final frame rate of the image was around 80fps if RGB values were glued every 16 milliseconds. The research shows that this method can meet real-time needs of music visualization and it can be used for live shows.
基于MIDI规范的多维复调表达音乐可视化
音乐可视化是信息可视化的一个分支。首先,分析了MIDI规范在“每音通道”多维复调表达中的优势。其次,提出了基于人与计算机识别在声音和图像上的结构相似性的音乐可视化映射模型。该模型分别对应于俯仰弯曲、控制变化、速度、色调、饱和度和值。第三,提出了音乐可视化方案。将原始图像的亮度提高20% ~ 30%得到基础图像;然后将影响图像分为左、中、右三部分纯色区,分别对应低音、中音和高音;然后,对不同区域的音乐信息进行映射,计算纯色的HSV值,并将HSV值转换为RGB值;最后,将这三部分的RGB值粘合在一起,与底图进行乘法组合,得到最终的图像。第四,利用Cycling’74 Max/Jitter 7.3.4,优化图像组合计算,在低音、中音、高音、两个音符显示在同一区域和两个音符显示在两个不同区域的情况下完成5个实验。最后,通过计算Jit。在fpsgui中,如果RGB值每16毫秒粘合一次,那么图像的最终帧速率约为80fps。研究表明,该方法能够满足音乐可视化的实时性需求,可用于现场演出。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信