{"title":"基于无人机的灾难受害者搜索中人声重点成像的创新颜色图","authors":"Tomokichi Furusawa, C. Premachandra","doi":"10.1109/TENSYMP55890.2023.10223627","DOIUrl":null,"url":null,"abstract":"Unmanned aerial vehicles (UAVs) are being utilized for damage assessment in natural disasters and for search and rescue operations. Currently, the search for victims primarily relies on analyzing images captured by cameras mounted on UAVs. However, this approach has limitations when it comes to locating victims who are not within the camera's field of view. As a result, sound-based search methods are being considered. In this method, a voice message is transmitted to the disaster area through a speaker mounted on a UAV, and the presence of victims is confirmed by detecting their response using the onboard microphone of the UAV. Nevertheless, the UAV's microphone captures both the sound of the victim and the propeller rotation, posing a significant challenge in extracting the victim's voice from this combined audio. To address this issue, we propose a solution that involves generating spectrogram images of the sound mixture and the propeller sound, and extracting the human sound by subtracting them. We found that the conventional colormap was ineffective in emphasizing the human sound in the spectrogram images. To overcome this limitation, this paper introduces a new colormap based on the normal distribution. This colormap enhances human voices while attenuating propeller sounds by adjusting the mean and variance. Through the results of our experiments, we confirm that the proposed colormap effectively reduces propeller sound interference in the sound mixing and simultaneously emphasizes the voice of a disaster victim. By utilizing the proposed colormap, it becomes possible to visualize the victim's voice from the audio mixture acquired by the UAV's onboard microphone, enabling the identification of the victim.","PeriodicalId":314726,"journal":{"name":"2023 IEEE Region 10 Symposium (TENSYMP)","volume":"81 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-09-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Innovative Colormap for Emphatic Imaging of Human Voice for UAV-Based Disaster Victim Search\",\"authors\":\"Tomokichi Furusawa, C. Premachandra\",\"doi\":\"10.1109/TENSYMP55890.2023.10223627\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Unmanned aerial vehicles (UAVs) are being utilized for damage assessment in natural disasters and for search and rescue operations. Currently, the search for victims primarily relies on analyzing images captured by cameras mounted on UAVs. However, this approach has limitations when it comes to locating victims who are not within the camera's field of view. As a result, sound-based search methods are being considered. In this method, a voice message is transmitted to the disaster area through a speaker mounted on a UAV, and the presence of victims is confirmed by detecting their response using the onboard microphone of the UAV. Nevertheless, the UAV's microphone captures both the sound of the victim and the propeller rotation, posing a significant challenge in extracting the victim's voice from this combined audio. To address this issue, we propose a solution that involves generating spectrogram images of the sound mixture and the propeller sound, and extracting the human sound by subtracting them. We found that the conventional colormap was ineffective in emphasizing the human sound in the spectrogram images. To overcome this limitation, this paper introduces a new colormap based on the normal distribution. This colormap enhances human voices while attenuating propeller sounds by adjusting the mean and variance. Through the results of our experiments, we confirm that the proposed colormap effectively reduces propeller sound interference in the sound mixing and simultaneously emphasizes the voice of a disaster victim. By utilizing the proposed colormap, it becomes possible to visualize the victim's voice from the audio mixture acquired by the UAV's onboard microphone, enabling the identification of the victim.\",\"PeriodicalId\":314726,\"journal\":{\"name\":\"2023 IEEE Region 10 Symposium (TENSYMP)\",\"volume\":\"81 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2023-09-06\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2023 IEEE Region 10 Symposium (TENSYMP)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/TENSYMP55890.2023.10223627\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2023 IEEE Region 10 Symposium (TENSYMP)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/TENSYMP55890.2023.10223627","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Innovative Colormap for Emphatic Imaging of Human Voice for UAV-Based Disaster Victim Search
Unmanned aerial vehicles (UAVs) are being utilized for damage assessment in natural disasters and for search and rescue operations. Currently, the search for victims primarily relies on analyzing images captured by cameras mounted on UAVs. However, this approach has limitations when it comes to locating victims who are not within the camera's field of view. As a result, sound-based search methods are being considered. In this method, a voice message is transmitted to the disaster area through a speaker mounted on a UAV, and the presence of victims is confirmed by detecting their response using the onboard microphone of the UAV. Nevertheless, the UAV's microphone captures both the sound of the victim and the propeller rotation, posing a significant challenge in extracting the victim's voice from this combined audio. To address this issue, we propose a solution that involves generating spectrogram images of the sound mixture and the propeller sound, and extracting the human sound by subtracting them. We found that the conventional colormap was ineffective in emphasizing the human sound in the spectrogram images. To overcome this limitation, this paper introduces a new colormap based on the normal distribution. This colormap enhances human voices while attenuating propeller sounds by adjusting the mean and variance. Through the results of our experiments, we confirm that the proposed colormap effectively reduces propeller sound interference in the sound mixing and simultaneously emphasizes the voice of a disaster victim. By utilizing the proposed colormap, it becomes possible to visualize the victim's voice from the audio mixture acquired by the UAV's onboard microphone, enabling the identification of the victim.