Mau-Luen Tham, Y. Wong, Ban-Hoe Kwan, Y. Owada, M. Sein, Yoong Choon Chang
{"title":"基于多任务学习的联合灾害分类与受害者检测","authors":"Mau-Luen Tham, Y. Wong, Ban-Hoe Kwan, Y. Owada, M. Sein, Yoong Choon Chang","doi":"10.1109/uemcon53757.2021.9666576","DOIUrl":null,"url":null,"abstract":"Recent advances in deep learning and computer vision have transformed surveillance into an important application for smart disaster monitoring systems. Based on the detected number of victims and activity of disasters, emergency response unit can dispatch manpower more efficiently, which could save more lives. However, most of existing disaster detection methods fall into the class of single-task learning, which can either detect victim or classify disaster. In contrast, this paper proposes a YOLO-based multi-task model which performs the aforementioned tasks simultaneously. This is accomplished by attaching a disaster classification head model to the backbone of a victim detection model. The head model is inherited from the MobileNetv2 architecture, and we precisely select the backbone feature map layer to which the head model is attached. For the victim detection, results reveal that the solution achieves up to 0.6938 and 20.31 in terms of average precision and frame per second, respectively. Whereas for the disaster classification, the algorithm is comparable with most deep learning models that are specifically trained for single task. This shows that our solution is flexible and robust enough to handle both victim detection and disaster classification.","PeriodicalId":127072,"journal":{"name":"2021 IEEE 12th Annual Ubiquitous Computing, Electronics & Mobile Communication Conference (UEMCON)","volume":"26 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":"{\"title\":\"Joint Disaster Classification and Victim Detection using Multi-Task Learning\",\"authors\":\"Mau-Luen Tham, Y. Wong, Ban-Hoe Kwan, Y. Owada, M. Sein, Yoong Choon Chang\",\"doi\":\"10.1109/uemcon53757.2021.9666576\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Recent advances in deep learning and computer vision have transformed surveillance into an important application for smart disaster monitoring systems. Based on the detected number of victims and activity of disasters, emergency response unit can dispatch manpower more efficiently, which could save more lives. However, most of existing disaster detection methods fall into the class of single-task learning, which can either detect victim or classify disaster. In contrast, this paper proposes a YOLO-based multi-task model which performs the aforementioned tasks simultaneously. This is accomplished by attaching a disaster classification head model to the backbone of a victim detection model. The head model is inherited from the MobileNetv2 architecture, and we precisely select the backbone feature map layer to which the head model is attached. For the victim detection, results reveal that the solution achieves up to 0.6938 and 20.31 in terms of average precision and frame per second, respectively. Whereas for the disaster classification, the algorithm is comparable with most deep learning models that are specifically trained for single task. This shows that our solution is flexible and robust enough to handle both victim detection and disaster classification.\",\"PeriodicalId\":127072,\"journal\":{\"name\":\"2021 IEEE 12th Annual Ubiquitous Computing, Electronics & Mobile Communication Conference (UEMCON)\",\"volume\":\"26 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-12-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"3\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2021 IEEE 12th Annual Ubiquitous Computing, Electronics & Mobile Communication Conference (UEMCON)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/uemcon53757.2021.9666576\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 IEEE 12th Annual Ubiquitous Computing, Electronics & Mobile Communication Conference (UEMCON)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/uemcon53757.2021.9666576","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Joint Disaster Classification and Victim Detection using Multi-Task Learning
Recent advances in deep learning and computer vision have transformed surveillance into an important application for smart disaster monitoring systems. Based on the detected number of victims and activity of disasters, emergency response unit can dispatch manpower more efficiently, which could save more lives. However, most of existing disaster detection methods fall into the class of single-task learning, which can either detect victim or classify disaster. In contrast, this paper proposes a YOLO-based multi-task model which performs the aforementioned tasks simultaneously. This is accomplished by attaching a disaster classification head model to the backbone of a victim detection model. The head model is inherited from the MobileNetv2 architecture, and we precisely select the backbone feature map layer to which the head model is attached. For the victim detection, results reveal that the solution achieves up to 0.6938 and 20.31 in terms of average precision and frame per second, respectively. Whereas for the disaster classification, the algorithm is comparable with most deep learning models that are specifically trained for single task. This shows that our solution is flexible and robust enough to handle both victim detection and disaster classification.