{"title":"Dense real-time stereo matching using memory efficient semi-global-matching variant based on FPGAs","authors":"M. Buder","doi":"10.1117/12.921147","DOIUrl":"https://doi.org/10.1117/12.921147","url":null,"abstract":"This paper presents a stereo image matching system that takes advantage of a global image matching method. The system \u0000is designed to provide depth information for mobile robotic applications. Typical tasks of the proposed system are to assist \u0000in obstacle avoidance, SLAM and path planning. Mobile robots pose strong requirements about size, energy consumption, \u0000reliability and output quality of the image matching subsystem. Current available systems either rely on active sensors or \u0000on local stereo image matching algorithms. The first are only suitable in controlled environments while the second suffer \u0000from low quality depth-maps. Top ranking quality results are only achieved by an iterative approach using global image \u0000matching and color segmentation techniques which are computationally demanding and therefore difficult to be executed \u0000in realtime. Attempts were made to still reach realtime performance with global methods by simplifying the routines. The \u0000depth maps are at the end almost comparable to local methods. An equally named semi-global algorithm was proposed \u0000earlier that shows both very good image matching results and relatively simple operations. A memory efficient variant of \u0000the Semi-Global-Matching algorithm is reviewed and adopted for an implementation based on reconfigurable hardware. \u0000The implementation is suitable for realtime execution in the field of robotics. It will be shown that the modified version of \u0000the efficient Semi-Global-Matching method is delivering equivalent result compared to the original algorithm based on the \u0000Middlebury dataset. \u0000The system has proven to be capable of processing VGA sized images with a disparity resolution of 64 pixel at \u000033 frames per second based on low cost to mid-range hardware. In case the focus is shifted to a higher image resolution, \u00001024×1024-sized stereo frames may be processed with the same hardware at 10 fps. The disparity resolution settings \u0000stay unchanged. A mobile system that covers preprocessing, matching and interfacing operations is also presented.","PeriodicalId":369288,"journal":{"name":"Real-Time Image and Video Processing","volume":"46 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124013581","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
J. Portillo-Portillo, F. Funes, A. Rosales-Silva, V. Ponomaryov
{"title":"Movement detection using an order statistics algorithm","authors":"J. Portillo-Portillo, F. Funes, A. Rosales-Silva, V. Ponomaryov","doi":"10.1117/12.924143","DOIUrl":"https://doi.org/10.1117/12.924143","url":null,"abstract":"ABSTRACT In this paper, we present a novel algorithm to motion detection in video sequences. The proposed algorithm is based in the use of the median of the absolute deviations from the me dian (MAD) as a measure of statistical dispersion of pixels in a video sequence to provide the robustness needed to detect motion in a frame of video sequence. By using the MAD, the proposed algorithm is able to detect small or big objects, the size of the detected objects depend of the size of kernel used in the analysis of the video sequence. Experimental results in the human motion detection are presented showing that the proposed algorithm can be used in security applications. Keywords: Median of the absolute deviations from the median, motion detection, video sequences 1. INTRODUCTION Video surveillance has been a popular security tool for years [1]. They are used for safety and security in public environments and in the private sector. The science and technological applications of advanced video surveillance systems have progressed tremendously in recent years due to widened research in areas including transport networks, elder care, traffic monitoring, traffic flow analysis, endangered species conservation, home nursing, human activity understanding, observation of people and vehicles within a busy environment, etc. [2]. The design of an automatic video surveillance system requir es several critical functionalities such as, motion detection and tracking, classification, behavior monitoring, activity analysis, and identification [2-6]. Motion detection is one of the current research areas in the design of video surveillanc e systems. It provides segmentation of the video streams into foreground and background components in order to extract the desired moving objects and it is a critical preprocess for several computer vision applications, including object-based video encoding, human motion analysis, and humanmachine interactions [2]. Detecting humans in frames of video sequences is a useful application of computer vision. Loose and textured clothing, occlusion and scene clutter make the human detection a difficult problem due bottom-up segmentation and grouping do not always work [7]. Several methods to motion detection can be implemented in video surveillance systems [1-7]. The common and easy way to develop such a system is to compute the absolute value of the difference between the current frame and a reference frame. After this, a threshold is imposed in the difference matrix. The pixels whose value are higher than a threshold are supposed to be part of an object moving into the scene, and those that are smaller belong to the background [4]. The value of the threshold to be chosen has a critical part in th is kind of systems, this value must be applied to the whole image. If the value of the mentioned threshold is small, the false alarm probability can increase, mainly due to the noise picked up by the camera, and the noise gene rated by a non-static","PeriodicalId":369288,"journal":{"name":"Real-Time Image and Video Processing","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128634335","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
M. Loktev, G. Vdovin, O. Soloviev, S. Kryukov, S. Savenko
{"title":"Adaptive optics combined with computer post-processing for horizontal turbulent imaging","authors":"M. Loktev, G. Vdovin, O. Soloviev, S. Kryukov, S. Savenko","doi":"10.1117/12.927853","DOIUrl":"https://doi.org/10.1117/12.927853","url":null,"abstract":"We describe an approach that oers an almost real time image enhancement through turbulent and wavy me- \u0000dia. The approach consists in a combination of optimization-based adaptive optics with digital multi-frame \u0000post-processing. Applications in astronomical and terrestrial imaging { where the image features are initially \u0000unresolved due to loss of contrast, blur, vibrations and image wander { have been illustrated by experimental \u0000results. A new software from Flexible Optical BV is presented","PeriodicalId":369288,"journal":{"name":"Real-Time Image and Video Processing","volume":"28 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128956147","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Abdul Waheed Malik, Benny Thörnberg, Xiaozhou Meng, Muhammad Imran
{"title":"Real-time machine vision system using FPGA and soft-core processor","authors":"Abdul Waheed Malik, Benny Thörnberg, Xiaozhou Meng, Muhammad Imran","doi":"10.1117/12.927854","DOIUrl":"https://doi.org/10.1117/12.927854","url":null,"abstract":"This paper presents a machine vision system for real-time computation of distance and angle of a camera from reference points in the environment. Image pre-processing, component labeling and featur ...","PeriodicalId":369288,"journal":{"name":"Real-Time Image and Video Processing","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"120954008","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Invariant methods for real-time object recognition and image understanding","authors":"Peter F. Stiller","doi":"10.1117/12.922857","DOIUrl":"https://doi.org/10.1117/12.922857","url":null,"abstract":"In this paper we discuss certain recently developed invariant geometric techniques that can be used for fast \u0000object recognition or fast image understanding. The results make use of techniques from algebraic geometry \u0000that allow one to relate the geometric invariants of a feature set in 3D to similar invariants in 2D or 1D. The \u0000methods apply equally well to optical images or radar images. In addition to the \"object/image\" equations \u0000relating these invariants, we also discuss certain invariant metrics and show why they provide a more natural \u0000and robust test for matching object features to image features. Additional aspects of the work as it applies to \u0000shape reconstruction and shape statistics will also be explored.","PeriodicalId":369288,"journal":{"name":"Real-Time Image and Video Processing","volume":"64 4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116556991","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Benchmarking real-time HEVC streaming","authors":"James Nightingale, Qi Wang, C. Grecos","doi":"10.1117/12.921406","DOIUrl":"https://doi.org/10.1117/12.921406","url":null,"abstract":"Work towards the standardisation of High Efficiency Video Coding (HEVC), the next generation video coding scheme, is currently gaining pace. HEVC offers the prospect of a 50% improvement in compression over the current H.264 Advanced Video Coding standard (H.264/AVC). Thus far, work on HEVC has concentrated on improvements to the coding efficiency and has not yet addressed transmission in networks other than to mandate byte stream compliance with Annex B of H.264/AVC. For practical networked HEVC applications a number of essential building blocks have yet to be defined. In this work, we design and prototype a real-time HEVC streaming system and empirically evaluate its performance, in particular we consider the robustness of the current Test Model under Consideration (TMuC HM4.0) for HEVC to packet loss caused by a reduction in available bandwidth both in terms of decoder resilience and degradation in perceptual video quality. A NAL unit packetisation and streaming framework for HEVC encoded video streams is designed, implemented and empirically tested in a number of streaming environments including wired, wireless, single path and multiple path network scenarios. As a first step the HEVC decoder’s error resilience is tested under a comprehensive set of packet loss conditions and a simple error concealment method for HEVC is implemented. Similarly to H.264 encoded streams, the size and distribution of NAL units within an HEVC stream and the nature of the NAL unit dependencies influences the packetisation and streaming strategies which may be employed for such streams. The relationships between HEVC encoding mode and the quality of the received video are shown under a wide range of bandwidth constraints. HEVC streaming is evaluated in both single and multipath network configuration scenarios. Through the use of extensive experimentation, we establish a comprehensive set of benchmarks for HEVC streaming in loss prone network environments. We show the visual quality reduction in terms of PSNR which results from a reduction in available bandwidth. To the best of our knowledge, this is the first time that such a fully functional streaming system for HEVC, together with the benchmark evaluation results, has been reported. This study will open up more timely research opportunities in this cutting edge area.","PeriodicalId":369288,"journal":{"name":"Real-Time Image and Video Processing","volume":"32 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125958659","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Video-based realtime IMU-camera calibration for robot navigation","authors":"Arne Petersen, R. Koch","doi":"10.1117/12.924066","DOIUrl":"https://doi.org/10.1117/12.924066","url":null,"abstract":"This paper introduces a new method for fast calibration of inertial measurement units (IMU) with cameras being rigidly \u0000coupled. That is, the relative rotation and translation between the IMU and the camera is estimated, allowing for the \u0000transfer of IMU data to the cameras coordinate frame. Moreover, the IMUs nuisance parameters (biases and scales) and \u0000the horizontal alignment of the initial camera frame are determined. Since an iterated Kalman Filter is used for estimation, \u0000information on the estimations precision is also available. Such calibrations are crucial for IMU-aided visual robot \u0000navigation, i.e. SLAM, since wrong calibrations cause biases and drifts in the estimated position and orientation. As the \u0000estimation is performed in realtime, the calibration can be done using a freehand movement and the estimated parameters \u0000can be validated just in time. This provides the opportunity of optimizing the used trajectory online, increasing the quality \u0000and minimizing the time effort for calibration. Except for a marker pattern, used for visual tracking, no additional hardware \u0000is required. \u0000As will be shown, the system is capable of estimating the calibration within a short period of time. Depending on \u0000the requested precision trajectories of 30 seconds to a few minutes are sufficient. This allows for calibrating the system \u0000at startup. By this, deviations in the calibration due to transport and storage can be compensated. The estimation quality \u0000and consistency are evaluated in dependency of the traveled trajectories and the amount of IMU-camera displacement and \u0000rotation misalignment. It is analyzed, how different types of visual markers, i.e. 2- and 3-dimensional patterns, effect the \u0000estimation. Moreover, the method is applied to mono and stereo vision systems, providing information on the applicability \u0000to robot systems. The algorithm is implemented using a modular software framework, such that it can be adopted to altered \u0000conditions easily.","PeriodicalId":369288,"journal":{"name":"Real-Time Image and Video Processing","volume":"34 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127070509","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Real-time visual communication to aid disaster recovery in a multi-segment hybrid wireless networking system","authors":"Tawfik Al Hadhrami, Qi Wang, C. Grecos","doi":"10.1117/12.924257","DOIUrl":"https://doi.org/10.1117/12.924257","url":null,"abstract":"When natural disasters or other large-scale incidents occur, obtaining accurate and timely information on the developing \u0000situation is vital to effective disaster recovery operations. \u0000High-quality video streams and high-resolution images, if \u0000available in real time, would provide an invaluable source of current situation reports to the incident management team. \u0000Meanwhile, a disaster often causes significant damage to the communications infrastructure. Therefore, another essential \u0000requirement for disaster management is the ability to rapidly deploy a flexible incident area communication network. \u0000Such a network would facilitate the transmission of real-time video streams and still images from the disrupted area to \u0000remote command and control locations. \u0000In this paper, a comprehensive end-to-end video/image transmission system between an incident area and a remote \u0000control centre is proposed and implemented, and its performance is experimentally investigated. In this study a hybrid \u0000multi-segment communication network is designed that seamlessly integrates terrestrial wireless mesh networks \u0000(WMNs), distributed wireless visual sensor networks, an airborne platform with video camera balloons, and a Digital \u0000Video Broadcasting- Satellite (DVB-S) system. \u0000By carefully integrating all of these rapidly deployable, interworking and collaborative networking technologies, we can \u0000fully exploit the joint benefits provided by WMNs, WSNs, balloon camera networks and DVB-S for real-time video \u0000streaming and image delivery in emergency situations among the disaster hit area, the remote control centre and the \u0000rescue teams in the field. The whole proposed system is implemented in a proven simulator. Through extensive \u0000simulations, the real-time visual communication performance of this integrated system has been numerically evaluated, \u0000towards a more in-depth understanding in supporting high-quality visual communications in such a demanding context.","PeriodicalId":369288,"journal":{"name":"Real-Time Image and Video Processing","volume":"19 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125527754","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Real-time video streaming in mobile cloud over heterogeneous wireless networks","authors":"Saleh Abdallah-Saleh, Qi Wang, C. Grecos","doi":"10.1117/12.924258","DOIUrl":"https://doi.org/10.1117/12.924258","url":null,"abstract":"Recently, the concept of Mobile Cloud Computing (MCC) has been proposed to offload the resource requirements in \u0000computational capabilities, storage and security from mobile devices into the cloud. Internet video applications such as \u0000real-time streaming are expected to be ubiquitously deployed and supported over the cloud for mobile users, who \u0000typically encounter a range of wireless networks of diverse radio access technologies during their roaming. However, \u0000real-time video streaming for mobile cloud users across heterogeneous wireless networks presents multiple challenges. \u0000The network-layer quality of service (QoS) provision to support high-quality mobile video delivery in this demanding \u0000scenario remains an open research question, and this in turn affects the application-level visual quality and impedes \u0000mobile users' perceived quality of experience (QoE). \u0000In this paper, we devise a framework to support real-time video streaming in this new mobile video networking paradigm \u0000and evaluate the performance of the proposed framework empirically through a lab-based yet realistic testing platform. \u0000One particular issue we focus on is the effect of users' mobility on the QoS of video streaming over the cloud. We design \u0000and implement a hybrid platform comprising of a test-bed and an emulator, on which our concept of mobile cloud \u0000computing, video streaming and heterogeneous wireless networks are implemented and integrated to allow the testing of \u0000our framework. As representative heterogeneous wireless networks, the popular WLAN (Wi-Fi) and MAN (WiMAX) \u0000networks are incorporated in order to evaluate effects of handovers between these different radio access technologies. \u0000The H.264/AVC (Advanced Video Coding) standard is employed for real-time video streaming from a server to mobile \u0000users (client nodes) in the networks. Mobility support is introduced to enable continuous streaming experience for a \u0000mobile user across the heterogeneous wireless network. Real-time video stream packets are captured for analytical \u0000purposes on the mobile user node. Experimental results are obtained and analysed. Future work is identified towards \u0000further improvement of the current design and implementation. \u0000With this new mobile video networking concept and paradigm implemented and evaluated, results and observations \u0000obtained from this study would form the basis of a more in-depth, comprehensive understanding of various challenges \u0000and opportunities in supporting high-quality real-time video streaming in mobile cloud over heterogeneous wireless \u0000networks.","PeriodicalId":369288,"journal":{"name":"Real-Time Image and Video Processing","volume":"61 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128875970","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Jakub Rosner, Hannes Fassold, M. Winter, P. Schallauer
{"title":"Real-time video breakup detection for multiple HD video streams on a single GPU","authors":"Jakub Rosner, Hannes Fassold, M. Winter, P. Schallauer","doi":"10.1117/12.921529","DOIUrl":"https://doi.org/10.1117/12.921529","url":null,"abstract":"An important task in film and video preservation is the quality assessment of the content to be archived or reused out of \u0000the archive. This task, if done manually, is a straining and time consuming process, so it is highly recommended to \u0000automate this process as far as possible. In this paper, we show how to port a previously proposed algorithm for detection \u0000of severe analog and digital video distortions (termed \"video breakup\"), efficiently to NVIDIA GPUs of the Fermi \u0000Architecture with CUDA. By parallizing of the algorithm massively in order to take usage of the hundreds of cores on a \u0000typical GPU and careful usage of GPU features like atomic functions, texture and shared memory, we achive a speedup \u0000of roughly 10-15 when comparing the GPU implementation with an highly optimized, multi-threaded CPU \u0000implementation. Thus our GPU algorithm is able to analyze nine Full HD (1920 × 1080) video streams or 40 standard \u0000definition (720 × 576) video streams in real-time on a single inexpensive Nvidia Geforce GTX 480 GPU. Additionally, \u0000we present the AV-Inspector application for video quality analysis where the video breakup algorithm has been \u0000integrated.","PeriodicalId":369288,"journal":{"name":"Real-Time Image and Video Processing","volume":"40 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127792172","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}