{"title":"基于CNN的火灾检测与定位的多实例学习","authors":"M. Aktas, Ali Bayramcavus, Toygar Akgun","doi":"10.1109/AVSS.2019.8909842","DOIUrl":null,"url":null,"abstract":"Motivated by the state-of-the-art performance achieved by convolutional neural networks (CNN) in visual detection and classification tasks, CNNs have recently been applied to the visual fire detection problem. In this work, we extend the existing CNN based approaches to fire detection in video sequences by incorporating Multiple Instance Learning (MIL). MIL relaxes the requirement of having accurate locations of fire patches in video frames, which are needed for patch level CNN training. Instead, only frame level labels indicating the presence of fire somewhere in a video frame are needed, substantially alleviating the annotation and training efforts. The resulting approach is tested on a new fire dataset obtained by extending some of the previously used fire datasets with video sequences collected from the web. Experimental results show that the proposed method improves fire detection performance upto 2.5%, while providing patch level localization without requiring patch level annotations.","PeriodicalId":243194,"journal":{"name":"2019 16th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS)","volume":"43 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"6","resultStr":"{\"title\":\"Multiple Instance Learning for CNN Based Fire Detection and Localization\",\"authors\":\"M. Aktas, Ali Bayramcavus, Toygar Akgun\",\"doi\":\"10.1109/AVSS.2019.8909842\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Motivated by the state-of-the-art performance achieved by convolutional neural networks (CNN) in visual detection and classification tasks, CNNs have recently been applied to the visual fire detection problem. In this work, we extend the existing CNN based approaches to fire detection in video sequences by incorporating Multiple Instance Learning (MIL). MIL relaxes the requirement of having accurate locations of fire patches in video frames, which are needed for patch level CNN training. Instead, only frame level labels indicating the presence of fire somewhere in a video frame are needed, substantially alleviating the annotation and training efforts. The resulting approach is tested on a new fire dataset obtained by extending some of the previously used fire datasets with video sequences collected from the web. Experimental results show that the proposed method improves fire detection performance upto 2.5%, while providing patch level localization without requiring patch level annotations.\",\"PeriodicalId\":243194,\"journal\":{\"name\":\"2019 16th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS)\",\"volume\":\"43 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2019-09-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"6\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2019 16th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/AVSS.2019.8909842\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2019 16th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/AVSS.2019.8909842","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Multiple Instance Learning for CNN Based Fire Detection and Localization
Motivated by the state-of-the-art performance achieved by convolutional neural networks (CNN) in visual detection and classification tasks, CNNs have recently been applied to the visual fire detection problem. In this work, we extend the existing CNN based approaches to fire detection in video sequences by incorporating Multiple Instance Learning (MIL). MIL relaxes the requirement of having accurate locations of fire patches in video frames, which are needed for patch level CNN training. Instead, only frame level labels indicating the presence of fire somewhere in a video frame are needed, substantially alleviating the annotation and training efforts. The resulting approach is tested on a new fire dataset obtained by extending some of the previously used fire datasets with video sequences collected from the web. Experimental results show that the proposed method improves fire detection performance upto 2.5%, while providing patch level localization without requiring patch level annotations.