{"title":"From PC2BIM: Automatic Model generation from Indoor Point Cloud","authors":"Danielle Tchuinkou Kwadjo, Erman Nghonda Tchinda, C. Bobda, Nareph Menadjou, Cedrique Fotsing, Nafissetou Nziengam, Nafissetou Nziengam","doi":"10.1145/3349801.3349825","DOIUrl":"https://doi.org/10.1145/3349801.3349825","url":null,"abstract":"In this paper, we present a system to automatically generate BIMs1 model from indoor point cloud. In contrary to previous works, our approach is able to take as input a point cloud with the minimum of information namely the points of coordinates (x, y, z) and produce excellent results. We first detect major flat surfaces such a walls, floor, and ceiling which are the bedrocks of our structure. Then, we present a novel 2D matrix template representation of walls which ease the operations like room layout and openings detection in polynomial time. Finally, we generate the BIM model rich with spatial and semantic information about the physical structures. A series of experiments performed show the efficiency and the precision of our approach.","PeriodicalId":299138,"journal":{"name":"Proceedings of the 13th International Conference on Distributed Smart Cameras","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-09-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114720446","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
A. Truong, W. Philips, Junzhi Guan, N. Deligiannis, L. Abrahamyan
{"title":"Automatic Extrinsic Calibration of Camera Networks Based on Pedestrians","authors":"A. Truong, W. Philips, Junzhi Guan, N. Deligiannis, L. Abrahamyan","doi":"10.1145/3349801.3349802","DOIUrl":"https://doi.org/10.1145/3349801.3349802","url":null,"abstract":"Extrinsic camera calibration is essential for any computer vision tasks in a camera network. Usually, researchers place calibration objects in the scene to calibrate the cameras. However, when installing cameras in the field, this approach can be costly and impractical, especially when recalibration is needed. This paper proposes a novel accurate and fully automatic extrinsic calibration framework for camera networks with partially overlapping views. It is based on the analysis of pedestrian tracks without other calibration objects. Compared to the state of the art, the new method is fully automatic and robust. Our method detects human poses in the camera images and then models walking persons as vertical sticks. We propose a brute-force method to determine the pedestrian correspondences in multiple camera images. This information along with 3D estimated locations of the head and feet of the pedestrians are then used to compute the camera extrinsic matrices. We verified the robustness of the method in different camera setups and for both single pedestrian and multiple walking people. The results show that the proposed method can obtain the triangulation error of a few centimeters. Typically, it requires 40 seconds of collecting data from walking people to reach this accuracy in controlled environments and a few minutes for uncontrolled environments. As well as compute relative extrinsic parameters connecting the coordinate systems of cameras in a pairwise fashion automatically. Our proposed method could perform well in various situations such as multi-person, occlusions, or even at real intersections on the street.","PeriodicalId":299138,"journal":{"name":"Proceedings of the 13th International Conference on Distributed Smart Cameras","volume":"29 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-09-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130274177","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Dynamic Obstacle Detection in Traffic Environments","authors":"G. Erabati, Helder Araújo","doi":"10.1145/3349801.3357134","DOIUrl":"https://doi.org/10.1145/3349801.3357134","url":null,"abstract":"The research on autonomous vehicles has grown increasingly with the advent of neural networks. Dynamic obstacle detection is a fundamental step for self-driving vehicles in traffic environments. This paper presents a comparison of state-of-art object detection techniques like Faster R-CNN, YOLO and SSD with 2D image data. The algorithms for detection in driving, must be reliable, robust and should have a real time performance. The three methods are trained and tested on PASCAL VOC 2007 and 2012 datasets and both qualitative and quantitative results are presented. SSD model can be seen as a tradeoff for speed and small object detection. A novel method for object detection using 3D data (RGB and depth) is proposed. The proposed model incorporates two stage architecture modality for RGB and depth processing and later fused hierarchically. The model will be trained and tested on RGBD dataset in the future.","PeriodicalId":299138,"journal":{"name":"Proceedings of the 13th International Conference on Distributed Smart Cameras","volume":"148 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-09-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132259201","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Blockchain-grade privacy protection in surveillance systems","authors":"Andrea Zanotto, N. Conci, A. Montresor","doi":"10.1145/3349801.3357132","DOIUrl":"https://doi.org/10.1145/3349801.3357132","url":null,"abstract":"We propose a solution to increase the privacy of people recorded with security cameras without decreasing the details stored in the videos. We strongly believe that CCTV recordings are a necessary and precious source of information to be analyzed when a crime or other unfortunate events happen; for this reason, we would like to have powerful surveillance systems that are able to hide the identity of the recorded people while allowing subsequent recovery of the data. We use face morphing algorithms in order to transform the faces in such a way that the protected video keeps the original likeness but does not leak sensitive face information. We store the transformed information in a decentralized way and we adopt smart contracts on permissioned blockchains to guarantee that in order to retrieve the data, a collection of trusted authorities must give their consent.","PeriodicalId":299138,"journal":{"name":"Proceedings of the 13th International Conference on Distributed Smart Cameras","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-09-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122049883","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Drone patrolling with reinforcement learning","authors":"C. Piciarelli, G. Foresti","doi":"10.1145/3349801.3349805","DOIUrl":"https://doi.org/10.1145/3349801.3349805","url":null,"abstract":"When a camera-equipped drone is assigned a patrolling task, it typically follows a pre-defined path that evenly covers the whole environment. In this paper instead we consider the problem of finding an ideal path under the assumption that not all the areas have the same coverage requirements. We thus propose a reinforcement learning approach that, given a relevance map representing coverage requirements, autonomously chooses the best drone actions to optimize the coverage.","PeriodicalId":299138,"journal":{"name":"Proceedings of the 13th International Conference on Distributed Smart Cameras","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-09-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129631238","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Rocco Pietrini, Davide Manco, M. Paolanti, V. Placidi, E. Frontoni, P. Zingaretti
{"title":"An IoT Edge-Fog-Cloud Architecture for Vision Based Planogram Integrity","authors":"Rocco Pietrini, Davide Manco, M. Paolanti, V. Placidi, E. Frontoni, P. Zingaretti","doi":"10.1145/3349801.3349807","DOIUrl":"https://doi.org/10.1145/3349801.3349807","url":null,"abstract":"A planogram is a detailed visual map of the products in the shelves of a retail store. The planogram wants to provide the best location of products on the shelves, with the aim to improve the customer experience and satisfaction, to increase sales and profits and to better manage the products on the shelves. However, an important aspect is not only to design the best planogram, but mostly to maintain the right position, orientation and, quantity of the products in the shelves according to the accepted planogram. For this purpose, we propose a fog computing architecture consisted of edge nodes which are low-cost cameras able to transmit in wireless mode shelf photos to local fog nodes in the same store. These last examine the images coming from the edges and sends results to the cloud for further data aggregation and analysis. The experimental results derive from a real environment considering a two-month observation period.","PeriodicalId":299138,"journal":{"name":"Proceedings of the 13th International Conference on Distributed Smart Cameras","volume":"40 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-09-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122653283","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A Distributed Approach to 3D Reconstruction in Marker Motion Capture Systems","authors":"A. Cenedese, Luca Varotto","doi":"10.1145/3349801.3349818","DOIUrl":"https://doi.org/10.1145/3349801.3349818","url":null,"abstract":"Optical motion capture systems have attracted much interest over the past years, due to their advantages with respect to non-optical counterparts. Moreover, with the technological advances on camera devices, computer graphics and computational methodologies, it becomes technically and economically feasible to consider motion capture systems made of large networks of cameras with embedded communication and processing units on board (i.e., smart cameras). In this case, the approaches relying on the classical 3D reconstruction methods would become inefficient, since their nature is intrinsically centralized. For this reason, we propose a distributed 3D reconstruction algorithm, which relies on a specific organization of cameras to remarkably speed up the scene reconstruction task. Indeed, numerical and experimental results show that the proposed computational scheme overcomes classical centralized solutions, in terms of reconstruction speed. Furthermore, the high processing speed does not compromise the estimation accuracy, since the algorithm is designed to be robust to occlusions and noise.","PeriodicalId":299138,"journal":{"name":"Proceedings of the 13th International Conference on Distributed Smart Cameras","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-09-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130787574","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"PLFCN","authors":"Shuo Liu, Wenrui Ding, Hongguang Li, Chunlei Liu","doi":"10.1145/3349801.3349819","DOIUrl":"https://doi.org/10.1145/3349801.3349819","url":null,"abstract":"In the field of remote sensing, the semantic segmentation network for orthophotos has received widely attention. However, it is usually impossible to achieve high accuracy and high efficiency at the same time. In this paper, we propose a novel pyramid loss reinforced fully convolutional network (PLFCN) to address this issue. By introducing deep pyramid supervisions, the network explores multi-scale spatial context information to improve performance of semantic segmentation. And the auxiliary pyramid loss structure can be ignored during testing, so that the network can inference as fast as FCN. The main contributions of this paper are as follows: 1) auxiliary pyramid loss structure is proposed to enhance the performance of FCN by multiscale and deep supervisions; 2) the advantages of multi scale structures and auxiliary loss is combined to improve the performance and maintain the efficiency at the same time. The results show that the semantic segmentation performance is significantly improved, while achieves the high effeciency as FCN.","PeriodicalId":299138,"journal":{"name":"Proceedings of the 13th International Conference on Distributed Smart Cameras","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-09-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115556811","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"EdgeNet: Balancing Accuracy and Performance for Edge-based Convolutional Neural Network Object Detectors","authors":"George Plastiras, C. Kyrkou, T. Theocharides","doi":"10.1145/3349801.3349809","DOIUrl":"https://doi.org/10.1145/3349801.3349809","url":null,"abstract":"Visual intelligence at the edge is becoming a growing necessity for low latency applications and situations where real-time decision is vital. Object detection, the first step in visual data analytics, has enjoyed significant improvements in terms of state-of-the-art accuracy due to the emergence of Convolutional Neural Networks (CNNs) and Deep Learning. However, such complex paradigms intrude increasing computational demands and hence prevent their deployment on resource-constrained devices. In this work, we propose a hierarchical framework that enables to detect objects in high-resolution video frames, and maintain the accuracy of state-of-the-art CNN-based object detectors while outperforming existing works in terms of processing speed when targeting a low-power embedded processor using an intelligent data reduction mechanism. Moreover, a use-case for pedestrian detection from Unmanned-Areal-Vehicle (UAV) is presented showing the impact that the proposed approach has on sensitivity, average processing time and power consumption when is implemented on different platforms. Using the proposed selection process our framework manages to reduce the processed data by 100x leading to under 4W power consumption on different edge devices.","PeriodicalId":299138,"journal":{"name":"Proceedings of the 13th International Conference on Distributed Smart Cameras","volume":"121 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-09-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122669967","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Deep Super-Resolution Network for Single Image Super-Resolution with Realistic Degradations","authors":"Rao Muhammad Umer, G. Foresti, C. Micheloni","doi":"10.1145/3349801.3349823","DOIUrl":"https://doi.org/10.1145/3349801.3349823","url":null,"abstract":"Single Image Super-Resolution (SISR) aims to generate a high-resolution (HR) image of a given low-resolution (LR) image. The most of existing convolutional neural network (CNN) based SISR methods usually take an assumption that a LR image is only bicubicly down sampled version of an HR image. However, the true degradation (i.e. the LR image is a bicubicly downsampled, blurred and noisy version of an HR image) of a LR image goes beyond the widely used bicubic assumption, which makes the SISR problem highly ill-posed nature of inverse problems. To address this issue, we propose a deep SISR network that works for blur kernels of different sizes, and different noise levels in an unified residual CNN-based denoiser network, which significantly improves a practical CNN-based super-resolver for real applications. Extensive experimental results on synthetic LR datasets and real images demonstrate that our proposed method not only can produce better results on more realistic degradation but also computational efficient to practical SISR applications.","PeriodicalId":299138,"journal":{"name":"Proceedings of the 13th International Conference on Distributed Smart Cameras","volume":"281 6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-09-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127479569","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}