{"title":"Music visualization based on the MIDI specifications for multidimensional polyphonic expression","authors":"Z. Xie, Yu Chen","doi":"10.1109/ICSAI.2017.8248478","DOIUrl":null,"url":null,"abstract":"Music visualization is a branch of information visualization. First, the advantages of MIDI specifications for multidimensional polyphonic expression in “channel per note” were analyzed. Second, the new music visualization mapping model was proposed, which was based on the structural similarity of human and computer recognition in sound and image. The model corresponded to pitch bend, control changes, velocity to hue, saturation, and value respectively. Third, the scheme of music visualization was proposed. The base image was obtained by improving the brightness of original image by 20%–30%; then the influence image was divided into three parts as the left, the middle and the right solid color area, which corresponded to the bass voice, the alto voice and the treble voice respectively; next, music information from different areas were mapped to calculate the HSV values of solid color and then the HSV values were converted to the RGB values; finally, the RGB values of these three parts were glued together and combined with the base image in multiply mode to obtain the final image. Fourth, taking advantage of Cycling'74 Max/Jitter 7.3.4 and optimizing image combination calculation, five experiments were completed under the circumstances of bass voice, alto voice, treble voice, two notes shown in the same area and two notes shown in two different areas. Finally, by calculation of Jit.fpsgui, the final frame rate of the image was around 80fps if RGB values were glued every 16 milliseconds. The research shows that this method can meet real-time needs of music visualization and it can be used for live shows.","PeriodicalId":285726,"journal":{"name":"2017 4th International Conference on Systems and Informatics (ICSAI)","volume":"50 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2017-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2017 4th International Conference on Systems and Informatics (ICSAI)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICSAI.2017.8248478","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Music visualization is a branch of information visualization. First, the advantages of MIDI specifications for multidimensional polyphonic expression in “channel per note” were analyzed. Second, the new music visualization mapping model was proposed, which was based on the structural similarity of human and computer recognition in sound and image. The model corresponded to pitch bend, control changes, velocity to hue, saturation, and value respectively. Third, the scheme of music visualization was proposed. The base image was obtained by improving the brightness of original image by 20%–30%; then the influence image was divided into three parts as the left, the middle and the right solid color area, which corresponded to the bass voice, the alto voice and the treble voice respectively; next, music information from different areas were mapped to calculate the HSV values of solid color and then the HSV values were converted to the RGB values; finally, the RGB values of these three parts were glued together and combined with the base image in multiply mode to obtain the final image. Fourth, taking advantage of Cycling'74 Max/Jitter 7.3.4 and optimizing image combination calculation, five experiments were completed under the circumstances of bass voice, alto voice, treble voice, two notes shown in the same area and two notes shown in two different areas. Finally, by calculation of Jit.fpsgui, the final frame rate of the image was around 80fps if RGB values were glued every 16 milliseconds. The research shows that this method can meet real-time needs of music visualization and it can be used for live shows.