{"title":"快速视频阴影检测的时空融合网络","authors":"Jun-Hong Lin, Liansheng Wang","doi":"10.1145/3574131.3574455","DOIUrl":null,"url":null,"abstract":"Existing video shadow detectors often need postprocessing or additional input to perform better, thereby degrading their video shadow detection speed. In this work, we present a novel spatial-temporal fusion network (STF-Net), which can efficiently detect shadows in videos with real-time speed (30FPS) and postprocessing-free. Our STF-Net is based solely on an attention-based spatial-temporal fusion block, equipping with recurrence and CNNs entirely. Experimental results on ViSha validation dataset show that our network exceeds state-of-the-art methods quantitatively and qualitatively.","PeriodicalId":111802,"journal":{"name":"Proceedings of the 18th ACM SIGGRAPH International Conference on Virtual-Reality Continuum and its Applications in Industry","volume":"242 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-12-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Spatial-temporal Fusion Network for Fast Video Shadow Detection\",\"authors\":\"Jun-Hong Lin, Liansheng Wang\",\"doi\":\"10.1145/3574131.3574455\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Existing video shadow detectors often need postprocessing or additional input to perform better, thereby degrading their video shadow detection speed. In this work, we present a novel spatial-temporal fusion network (STF-Net), which can efficiently detect shadows in videos with real-time speed (30FPS) and postprocessing-free. Our STF-Net is based solely on an attention-based spatial-temporal fusion block, equipping with recurrence and CNNs entirely. Experimental results on ViSha validation dataset show that our network exceeds state-of-the-art methods quantitatively and qualitatively.\",\"PeriodicalId\":111802,\"journal\":{\"name\":\"Proceedings of the 18th ACM SIGGRAPH International Conference on Virtual-Reality Continuum and its Applications in Industry\",\"volume\":\"242 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2022-12-27\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Proceedings of the 18th ACM SIGGRAPH International Conference on Virtual-Reality Continuum and its Applications in Industry\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1145/3574131.3574455\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 18th ACM SIGGRAPH International Conference on Virtual-Reality Continuum and its Applications in Industry","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3574131.3574455","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Spatial-temporal Fusion Network for Fast Video Shadow Detection
Existing video shadow detectors often need postprocessing or additional input to perform better, thereby degrading their video shadow detection speed. In this work, we present a novel spatial-temporal fusion network (STF-Net), which can efficiently detect shadows in videos with real-time speed (30FPS) and postprocessing-free. Our STF-Net is based solely on an attention-based spatial-temporal fusion block, equipping with recurrence and CNNs entirely. Experimental results on ViSha validation dataset show that our network exceeds state-of-the-art methods quantitatively and qualitatively.