{"title":"A Framework for Queryable Video Analysis: A Case-Study on Transport Modelling","authors":"Mark Bugeja, A. Dingli, M. Attard, D. Seychell","doi":"10.1145/3349622.3355448","DOIUrl":null,"url":null,"abstract":"Analysing video data requires the use of different models trained to retrieve or process data for a particular task. In this paper, we introduce an approach to represent the visual context within a video as queryable information. Through the use of computer vision techniques, we can detect and classify objects. Our system processes these classifications in order to construct a queryable data-set referred to as the real world model. The advantage of this approach is that through the formalisation of the information, we can create generic queries to retrieve information. This approach allows for processing to be done on edge devices such as embedded cameras while only transmitting detected information reducing the transmission bandwidth as well as infrastructural costs. The final recognition data is processed on the cloud. The analysed case study works on video traffic representation - an experiment around the transport domain. We evaluate and validate our approach by posing several queries to the system that generates information on the traffic situation, such as car counting and traffic flow. The results show that our approach can add context to classifications with a high degree of accuracy in some of the cases, achieving 95% car counting accuracy during the day. Fine tuning approaches are also discussed with reference to the video traffic representation case while keeping to the same proposed methodology.","PeriodicalId":334983,"journal":{"name":"Proceedings of the 1st ACM Workshop on Emerging Smart Technologies and Infrastructures for Smart Mobility and Sustainability","volume":"5 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-10-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 1st ACM Workshop on Emerging Smart Technologies and Infrastructures for Smart Mobility and Sustainability","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3349622.3355448","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1
Abstract
Analysing video data requires the use of different models trained to retrieve or process data for a particular task. In this paper, we introduce an approach to represent the visual context within a video as queryable information. Through the use of computer vision techniques, we can detect and classify objects. Our system processes these classifications in order to construct a queryable data-set referred to as the real world model. The advantage of this approach is that through the formalisation of the information, we can create generic queries to retrieve information. This approach allows for processing to be done on edge devices such as embedded cameras while only transmitting detected information reducing the transmission bandwidth as well as infrastructural costs. The final recognition data is processed on the cloud. The analysed case study works on video traffic representation - an experiment around the transport domain. We evaluate and validate our approach by posing several queries to the system that generates information on the traffic situation, such as car counting and traffic flow. The results show that our approach can add context to classifications with a high degree of accuracy in some of the cases, achieving 95% car counting accuracy during the day. Fine tuning approaches are also discussed with reference to the video traffic representation case while keeping to the same proposed methodology.