Reiji Suzuki, Hao Zhao, Shinji Sumitani, Shiho Matsubayashi, Takaya Arita, K. Nakadai, H. Okuno
{"title":"利用机器人试听技术可视化鸟类发声的定向音景","authors":"Reiji Suzuki, Hao Zhao, Shinji Sumitani, Shiho Matsubayashi, Takaya Arita, K. Nakadai, H. Okuno","doi":"10.1109/IEEECONF49454.2021.9382639","DOIUrl":null,"url":null,"abstract":"A visualisation of the soundscape dynamics is one of the important topics in ecoacousics. However, existing approaches mainly focused on the soundscape in the frequency domain while the soundscape in the directional or spatial domain is also essential to better understand animal vocalizations. This paper proposes and discusses novel applications of robot audition techniques to visualize soundscape dynamics in the directional or spatial domain by using the directional information of sound sources obtained from the robot audition software HARK (Honda Research Institute Japan Audition for Robots with Kyoto University) and the software for birdsong localization HARKBird. First, we create a false-color spectrogram that visualizes directional soundscapes in which the color of the spectrogram reflects the direction of arrival of separated sounds. We also visualize the distribution of directional soundscapes by combining the entropy of the likelihood of sound existence (MUSIC spectrum) and a latent space embedding method (UMAP). We applied these techniques to a 5 min recording with 6 individuals of Zebra Finch in order to show that the extracted visual information can reflect acoustic structures among the group of bird individuals in the directional domain.","PeriodicalId":395378,"journal":{"name":"2021 IEEE/SICE International Symposium on System Integration (SII)","volume":"2016 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-01-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Visualizing Directional Soundscapes of Bird Vocalizations Using Robot Audition Techniques\",\"authors\":\"Reiji Suzuki, Hao Zhao, Shinji Sumitani, Shiho Matsubayashi, Takaya Arita, K. Nakadai, H. Okuno\",\"doi\":\"10.1109/IEEECONF49454.2021.9382639\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"A visualisation of the soundscape dynamics is one of the important topics in ecoacousics. However, existing approaches mainly focused on the soundscape in the frequency domain while the soundscape in the directional or spatial domain is also essential to better understand animal vocalizations. This paper proposes and discusses novel applications of robot audition techniques to visualize soundscape dynamics in the directional or spatial domain by using the directional information of sound sources obtained from the robot audition software HARK (Honda Research Institute Japan Audition for Robots with Kyoto University) and the software for birdsong localization HARKBird. First, we create a false-color spectrogram that visualizes directional soundscapes in which the color of the spectrogram reflects the direction of arrival of separated sounds. We also visualize the distribution of directional soundscapes by combining the entropy of the likelihood of sound existence (MUSIC spectrum) and a latent space embedding method (UMAP). We applied these techniques to a 5 min recording with 6 individuals of Zebra Finch in order to show that the extracted visual information can reflect acoustic structures among the group of bird individuals in the directional domain.\",\"PeriodicalId\":395378,\"journal\":{\"name\":\"2021 IEEE/SICE International Symposium on System Integration (SII)\",\"volume\":\"2016 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-01-11\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2021 IEEE/SICE International Symposium on System Integration (SII)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/IEEECONF49454.2021.9382639\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 IEEE/SICE International Symposium on System Integration (SII)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/IEEECONF49454.2021.9382639","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Visualizing Directional Soundscapes of Bird Vocalizations Using Robot Audition Techniques
A visualisation of the soundscape dynamics is one of the important topics in ecoacousics. However, existing approaches mainly focused on the soundscape in the frequency domain while the soundscape in the directional or spatial domain is also essential to better understand animal vocalizations. This paper proposes and discusses novel applications of robot audition techniques to visualize soundscape dynamics in the directional or spatial domain by using the directional information of sound sources obtained from the robot audition software HARK (Honda Research Institute Japan Audition for Robots with Kyoto University) and the software for birdsong localization HARKBird. First, we create a false-color spectrogram that visualizes directional soundscapes in which the color of the spectrogram reflects the direction of arrival of separated sounds. We also visualize the distribution of directional soundscapes by combining the entropy of the likelihood of sound existence (MUSIC spectrum) and a latent space embedding method (UMAP). We applied these techniques to a 5 min recording with 6 individuals of Zebra Finch in order to show that the extracted visual information can reflect acoustic structures among the group of bird individuals in the directional domain.