{"title":"Microwave Imaging of Reinforced Concrete and Design of a Broadband Antenna","authors":"Shengxiang Qi, Hui Xu, J. Ren","doi":"10.1145/3177404.3177424","DOIUrl":"https://doi.org/10.1145/3177404.3177424","url":null,"abstract":"The location imaging of reinforced concrete in building structure is studied. An imaging system for detecting concrete cross sections with ultra-wideband stepping frequency conversion radar is designed. According to the requirement of imaging system, a wideband double ridged horn antenna is used as the sensor of the imaging system by using electromagnetic simulation software HFSS and the average of voltage standing-wave ratio (VSWR) is less than 2.5. The transmitted signal is changed in step frequency, and the echo signals of different groups at different positions are collected. The internal structure of concrete is reconstructed by using the ω-k algorithm in the synthetic aperture microwave imaging method. A diameter of 1cm of steel in the concrete at different depth is reconstructed successfully by using this imaging system.","PeriodicalId":133378,"journal":{"name":"Proceedings of the International Conference on Video and Image Processing","volume":"34 2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-12-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124961770","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"License Plate Detection Based on Convolutional Neural Network: Support Vector Machine (CNN-SVM)","authors":"Gamma Kosala, A. Harjoko, S. Hartati","doi":"10.1145/3177404.3177436","DOIUrl":"https://doi.org/10.1145/3177404.3177436","url":null,"abstract":"Automatic License Plate Recognition (ALPR) implementation can be used in many applications, such as road traffic monitoring, automatic toll payments, and parking management. License plate detection is the first and very critical stage in the ALPR system. Locating the license plate in the image becomes more difficult in the complex backgrounds such as the highways. This research develops the plate detection method in a complex environment in two stages: plate candidate extraction, and plate area selection. We use Sobel operator for vertical edge detection, closing morphological operation, and Connected Component Analysis (CCA) for contour detection in plate candidate extraction stage. Plate area selection is implemented by using Convolutional Neural Network -- Support Vector Machine (CNN - SVM). CNN acts as feature extraction method whereas SVM as a classifier. Compared to some other machine learning architecture, CNN-SVM reached the highest accuracy by 93%.","PeriodicalId":133378,"journal":{"name":"Proceedings of the International Conference on Video and Image Processing","volume":"31 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-12-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128242952","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Robust Video Super-resolution Using Low-rank Matrix Completion","authors":"Chenyu Liu, Xianlin Zhang, Yang Liu, Xueming Li","doi":"10.1145/3177404.3177423","DOIUrl":"https://doi.org/10.1145/3177404.3177423","url":null,"abstract":"In this paper, we propose a robust super-resolution method using low rank matrix completion for videos with local motions and local deformations. It is based on the multi-frame low rank matrix completion super-resolution (MCSR) framework proposed by Chen. Nonlocal multi-scale similar patches are extracted in registration instead of optical flow for complex motions. By rearranging patches extracted from low resolution frames, super-resolution problem is converted to matrix completion. Low resolution patches is represented as observed entries in a low-rank matrix. We adopt alternating direction method of multipliers (ADMM) to minimize nuclear norm and introduce a weighted fusion method to acquire final high resolution patches. Experimental results showed that the proposed method outperformed MCSR on videos with complex motions.","PeriodicalId":133378,"journal":{"name":"Proceedings of the International Conference on Video and Image Processing","volume":"1069 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-12-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116291470","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Extracting Generic Features of Artistic Style via Deep Convolutional Neural Network","authors":"Lili Kong, Jiancheng Lv, Mao Li, Hanwang Zhang","doi":"10.1145/3177404.3177421","DOIUrl":"https://doi.org/10.1145/3177404.3177421","url":null,"abstract":"While most existing works on image art style transformation generally focus on the transformation given a specific style image as input, in this paper, we consider it given a set of images of a generic style, e.g., images of Vincent van Gogh during 1889 to 1890. Compared to the specific style from only one input style image, our generic style transformation is able to remove the artifact generated from the single image such as specific objects and scenes. To this end, we propose a method to extract generic style features from a set of fine paintings. Generic style features describe these fine paintings from the global perspective, integrate features of brush strokes, color and pose contrast, scale information and orientation etc. We first obtain feature representation from these fine paintings using deep convolutional neural network (CNN), and then select generic representation from obtained representation. Finally, migrate visualized generic style features to input content image. Experimental results verify the efficiency and power of our method.","PeriodicalId":133378,"journal":{"name":"Proceedings of the International Conference on Video and Image Processing","volume":"120 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-12-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116374160","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A Graph-based Approach for Rectangle Detection Using Corners","authors":"Shengkai Wu, L. Gou, Hui Xiong, Xiaoping Li","doi":"10.1145/3177404.3177417","DOIUrl":"https://doi.org/10.1145/3177404.3177417","url":null,"abstract":"In this paper, a new rectangle detection approach is proposed. It is a bottom-up approach that contains four stages: line segment extraction, corner detection, corner-relation-graph generation and rectangle detection. Graph structure is used to construct the relations between corners and simplify the problem of rectangle detection. In addition, the approach can be extended to detect any polygons. Experiments on bin detection, traffic sign detection and license plate detection prove that the approach is robust.","PeriodicalId":133378,"journal":{"name":"Proceedings of the International Conference on Video and Image Processing","volume":"275 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-12-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122083834","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
A. Mohamed, Effha Binti Soter, Anand Singh, Nur Intan Raihana Ruhaiyem
{"title":"Gesture Based Help Identification for Hospital & ElderlyCare Using Dynamic Time Warping: A Systematic Study","authors":"A. Mohamed, Effha Binti Soter, Anand Singh, Nur Intan Raihana Ruhaiyem","doi":"10.1145/3177404.3177426","DOIUrl":"https://doi.org/10.1145/3177404.3177426","url":null,"abstract":"Recent advancement of depth imaging sensor technology can provide opportunities, especially in healthcare sector to improve the quality of life of the hospital patients and elderly care institutes by facilitating certain tasks without any assistant and enabling supportive ecosphere. This paper presents a novel hand gesture system via motion capture for appliance control in the hospital environment. A selection of hand gestures were prepared for the subject to perform with a set of questionnaires to learn the user acceptance towards this implementation with the adoption of dynamic time warping (DTW) algorithm for identifying the gestures. The results shown that gestures are easily performed by the patient to control specific tasks in a controlled environment. Most participants agreed that this system is easy to use and learn. The aim of this study is to evaluate the user experience on the proposed gesture setup and study the acceptance of the proposed hand gestures by the subjects.","PeriodicalId":133378,"journal":{"name":"Proceedings of the International Conference on Video and Image Processing","volume":"104 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-12-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124745701","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Cut and Paste: Generate Artificial Labels for Object Detection","authors":"Jianghao Rao, Jianlin Zhang","doi":"10.1145/3177404.3177440","DOIUrl":"https://doi.org/10.1145/3177404.3177440","url":null,"abstract":"In the domain of object detection, region proposal, feature extraction, recognition and the localization are the main three tasks. The end-to-end detection models by integrating the three parts together to simplify the structure of network and accelerate the process of training and detection. While the issues of illumination change, object deformation and scale change undermine the performance of detection methods largely. To promote the object detection accuracy rate and boost the detection speed simultaneously, we propose a new method of data augmentation. Different from the traditional methods, our method can increase the training data largely and be free from overfitting to some extent. With the new method, the abstraction ability of models improves a lot, the model has better performance to multiscale objects detection, and also has a stronger distinguishing ability in complex background.","PeriodicalId":133378,"journal":{"name":"Proceedings of the International Conference on Video and Image Processing","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-12-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122616365","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Nunik Pratiwi, M. R. Widyanto, T. Basaruddin, D. Liliana
{"title":"Nonlinear Fuzzy Robust PCA on Shape Modelling of Active Appearance Model for Facial Expression Recognition","authors":"Nunik Pratiwi, M. R. Widyanto, T. Basaruddin, D. Liliana","doi":"10.1145/3177404.3177444","DOIUrl":"https://doi.org/10.1145/3177404.3177444","url":null,"abstract":"Automatic facial expression recognition is one of the potential research area in the field of computer vison. It aims to improve the ability of machine to capture social signals in human. Automatic facial expression recognition is still a challenge. We proposed method using contrast limited adaptive histogram equalization (CLAHE) for pre-processing stage then performed feature extraction using active appearance model (AAM) based on nonlinear fuzzy robust principal component analysis (NFRPCA). The feature extraction results will be classified with support vector machine (SVM). Feature points generated AAM based on NFRPCA more adaptive compared to AAM based PCA. Our proposed method's the average accuracy rate reached 96,87% and 93,94% for six and seven basic emotions respectively.","PeriodicalId":133378,"journal":{"name":"Proceedings of the International Conference on Video and Image Processing","volume":"75 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-12-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126932645","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Fast transform mode decision for HEVC screen content coding","authors":"Dokyung Lee, Jecheng Jeong","doi":"10.1145/3177404.3177429","DOIUrl":"https://doi.org/10.1145/3177404.3177429","url":null,"abstract":"A screen content coding (SCC) standard based on high efficiency video coding (HEVC) was finalized by joint collaborative team on video coding (JCT-VC). The coding efficiency of the standard has been improved by adopting new technologies, such as intra block copy, palette mode, and adaptive motion vector resolution. However, the encoding time is significantly increased. Also, a transform skip mode is selected more frequently because traditional transform techniques for video compression cannot efficiently compress screen content videos. Therefore, we propose an early determination method for transform process in the HEVC-SCC. Residual variance is employed as a measure of complexity. Using statistics of residual variance and an online learning system, we can adaptively determine a threshold to detect the transform skip mode. Experimental results demonstrate that the proposed algorithm successfully reduces encoding time with a small coding efficiency loss.","PeriodicalId":133378,"journal":{"name":"Proceedings of the International Conference on Video and Image Processing","volume":"19 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-12-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126772065","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Exposing Video Forgeries by Detecting Misaligned Double Compression","authors":"Shan Bian, Weiqi Luo, Jiwu Huang","doi":"10.1145/3177404.3177420","DOIUrl":"https://doi.org/10.1145/3177404.3177420","url":null,"abstract":"Powerful and fast video editing software tools have made video tampering an easy work. In video forgeries, re-compression is one of the inevitable post-processing, so it is usually explored in video forensics works. However, most works only consider aligned double compression---the order of frames remains unchanged before and after re-compression. In some cases, misaligned double compression could happen as well in video forgeries, such as video splicing, etc. In this paper, we propose an effective method to detect misaligned double compression. Based on extensive experiments and analysis, we found that H.264/AVC compression generates different strengths of blurring artifacts in different types of frames. After misaligned re-compression, such blurring artifacts introduced by the previous compression would be preserved in the re-compressed frames. Based on this observation, we propose a compact yet effective feature vector (MBAS) to expose video forgeries. The experiments evaluated on standard test sequences with a variety of encoding parameters have shown the effectiveness of the proposed method.","PeriodicalId":133378,"journal":{"name":"Proceedings of the International Conference on Video and Image Processing","volume":"190 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-12-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114855553","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}