A Framework for Queryable Video Analysis: A Case-Study on Transport Modelling

Mark Bugeja, A. Dingli, M. Attard, D. Seychell
{"title":"A Framework for Queryable Video Analysis: A Case-Study on Transport Modelling","authors":"Mark Bugeja, A. Dingli, M. Attard, D. Seychell","doi":"10.1145/3349622.3355448","DOIUrl":null,"url":null,"abstract":"Analysing video data requires the use of different models trained to retrieve or process data for a particular task. In this paper, we introduce an approach to represent the visual context within a video as queryable information. Through the use of computer vision techniques, we can detect and classify objects. Our system processes these classifications in order to construct a queryable data-set referred to as the real world model. The advantage of this approach is that through the formalisation of the information, we can create generic queries to retrieve information. This approach allows for processing to be done on edge devices such as embedded cameras while only transmitting detected information reducing the transmission bandwidth as well as infrastructural costs. The final recognition data is processed on the cloud. The analysed case study works on video traffic representation - an experiment around the transport domain. We evaluate and validate our approach by posing several queries to the system that generates information on the traffic situation, such as car counting and traffic flow. The results show that our approach can add context to classifications with a high degree of accuracy in some of the cases, achieving 95% car counting accuracy during the day. Fine tuning approaches are also discussed with reference to the video traffic representation case while keeping to the same proposed methodology.","PeriodicalId":334983,"journal":{"name":"Proceedings of the 1st ACM Workshop on Emerging Smart Technologies and Infrastructures for Smart Mobility and Sustainability","volume":"5 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-10-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 1st ACM Workshop on Emerging Smart Technologies and Infrastructures for Smart Mobility and Sustainability","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3349622.3355448","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1

Abstract

Analysing video data requires the use of different models trained to retrieve or process data for a particular task. In this paper, we introduce an approach to represent the visual context within a video as queryable information. Through the use of computer vision techniques, we can detect and classify objects. Our system processes these classifications in order to construct a queryable data-set referred to as the real world model. The advantage of this approach is that through the formalisation of the information, we can create generic queries to retrieve information. This approach allows for processing to be done on edge devices such as embedded cameras while only transmitting detected information reducing the transmission bandwidth as well as infrastructural costs. The final recognition data is processed on the cloud. The analysed case study works on video traffic representation - an experiment around the transport domain. We evaluate and validate our approach by posing several queries to the system that generates information on the traffic situation, such as car counting and traffic flow. The results show that our approach can add context to classifications with a high degree of accuracy in some of the cases, achieving 95% car counting accuracy during the day. Fine tuning approaches are also discussed with reference to the video traffic representation case while keeping to the same proposed methodology.
可查询视频分析框架:交通建模案例研究
分析视频数据需要使用经过训练的不同模型来检索或处理特定任务的数据。在本文中,我们介绍了一种将视频中的视觉上下文表示为可查询信息的方法。通过使用计算机视觉技术,我们可以对物体进行检测和分类。我们的系统处理这些分类是为了构造一个可查询的数据集,即现实世界模型。这种方法的优点是,通过信息的形式化,我们可以创建通用查询来检索信息。这种方法允许在边缘设备(如嵌入式摄像机)上进行处理,同时只传输检测到的信息,从而减少了传输带宽和基础设施成本。最终的识别数据在云端处理。分析的案例研究是关于视频流量表示的——一个围绕传输领域的实验。我们通过向系统提出几个查询来评估和验证我们的方法,这些查询生成有关交通状况的信息,例如车辆计数和交通流量。结果表明,我们的方法可以在某些情况下以很高的准确率为分类添加上下文,白天的汽车计数准确率达到95%。在保持所提出的方法不变的情况下,还参考视频流量表示情况讨论了微调方法。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信