Farhad Pakdaman, M. Gabbouj, M. Hashemi, M. Ghanbari
{"title":"Fast Motion Estimation Algorithm with Efficient Memory Access for HEVC Hardware Encoders","authors":"Farhad Pakdaman, M. Gabbouj, M. Hashemi, M. Ghanbari","doi":"10.1109/EUVIP.2018.8611766","DOIUrl":"https://doi.org/10.1109/EUVIP.2018.8611766","url":null,"abstract":"The encoding process in the HEVC standard is several times more complex than the previous standards. Since motion estimation is responsible for most of this complexity, the new Test Zone (TZ) search is usually adopted as the fast search algorithm, to alleviate the complexity. However, the TZ search requires a high rate of access to the off-chip memory, which contributes heavily to the total consumed encoding power. In this paper we demonstrate that the process of finding the best starting search point in this algorithm, does not allow effective reduction of memory access in hardware encoders. As a solution, a new fast motion estimation algorithm is proposed which estimates a proper single starting search point, in addition to an adaptively reduced search range, based on available information from the coded neighboring blocks. The experimental results show that this algorithm on average can reduce the required memory access for ME by ~78% and reduce the integer ME time by ~70%, with only 1.1% Bjontegaard Delta (BD) Rate.","PeriodicalId":252212,"journal":{"name":"2018 7th European Workshop on Visual Information Processing (EUVIP)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115862131","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Nahid Sheikhi-Pour, S. Schwarz, V. Vadakital, M. Gabbouj
{"title":"Efficient 2D Video Coding of Volumetric Video Data","authors":"Nahid Sheikhi-Pour, S. Schwarz, V. Vadakital, M. Gabbouj","doi":"10.1109/EUVIP.2018.8611742","DOIUrl":"https://doi.org/10.1109/EUVIP.2018.8611742","url":null,"abstract":"Point clouds for immersive media technology have received substantial interest in the recent years. Such representation of 3D scenery provides freedom of movement for the viewer. But transmitting or storing such content requires large amount of data and it is not feasible on today's network technology. Thus, there is a necessity for having efficient compression algorithms. Recently, projection-based methods have been considered for compressing point cloud data. In these methods, the point cloud data are projected onto a 2D image plane in order to utilize the current 2D video coding standards for compressing such content. Such coding schemes provide significant improvement over the state-of-the-art methods in terms of coding efficiency. However, the projection-based point cloud compression requires special handling of boundaries and sparsity in the 2D projections. This paper addresses these issues by proposing two methods which improve the compression performance of both intra-frame and inter-frame coding for 2D video coding of volumetric data. The conducted experiments illustrated that the bitrate requirements are reduced by around 26% and 29% for geometry and color attributes, respectively compared to the case that the proposed algorithms are not applied.","PeriodicalId":252212,"journal":{"name":"2018 7th European Workshop on Visual Information Processing (EUVIP)","volume":"1986 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127805841","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Licinio Oliveira, Jaime S. Cardoso, A. Lourenço, Christer Ahlström
{"title":"Driver drowsiness detection: a comparison between intrusive and non-intrusive signal acquisition methods","authors":"Licinio Oliveira, Jaime S. Cardoso, A. Lourenço, Christer Ahlström","doi":"10.1109/EUVIP.2018.8611704","DOIUrl":"https://doi.org/10.1109/EUVIP.2018.8611704","url":null,"abstract":"Driver drowsiness is a major cause of road accidents, many of which result in fatalities. A solution to this problem is the inclusion of a drowsiness detector in vehicles to alert the driver if sleepiness is detected. To detect drowsiness, physiologic, behavioral (visual) and vehicle-based methods can be used, however, only measures that can be acquired non-intrusively are viable in a real life application. This work uses data from a real-road experiment with sleep deprived drivers to compare the performance of driver drowsiness detection using intrusive acquisition methods, namely electrooculogram (EOG), with camera-based, non-intrusive, methods. A hybrid strategy, combining the described methods with electrocardiogram (ECG) measures, is also evaluated. Overall, the obtained results show that drowsiness detection performance is similar using non-intrusive camera-based measures or intrusive EOG measures. The detection performance increases when combining two methods (ECG + visual) or (ECG + EOG).","PeriodicalId":252212,"journal":{"name":"2018 7th European Workshop on Visual Information Processing (EUVIP)","volume":"47 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131104345","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"JPEG based Compression of Digital Holograms","authors":"Nada Chamakhi, I. Bouzidi, A. O. Zaid, F. Dufaux","doi":"10.1109/EUVIP.2018.8611713","DOIUrl":"https://doi.org/10.1109/EUVIP.2018.8611713","url":null,"abstract":"Modern holography for 3D imaging allows to reconstruct all the parallaxes that are needed for a truly immersive visualisation. Nevertheless, it represents a huge amount of data which induces higher transmission and storage requirements. To gain more popularity and acceptance, digital holography demands development of efficient coding schemes that provide significant data compression at low computation cost. Another issue that needs to be tackled when designing holography coding algorithms is interoperability with commonly used formats. The upcoming JPEG Pleno standard aims to develop a standard framework for the representation and exchange of new imaging modalities, such as holographic imaging, while maintaining backward compatibility with legacy JPEG decoders. This paper presents a lossy compression method for holographic images that exhibits good coding performance while considering the computation cost and backward compatibility with legacy JPEG standard. To validate our findings, the results of our tests are shown and interpreted.","PeriodicalId":252212,"journal":{"name":"2018 7th European Workshop on Visual Information Processing (EUVIP)","volume":"367 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127583955","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Virtual Reality meets Degas: an immersive framework for art exploration and learning","authors":"F. Battisti, Chiara Di Stefano","doi":"10.1109/EUVIP.2018.8611753","DOIUrl":"https://doi.org/10.1109/EUVIP.2018.8611753","url":null,"abstract":"The recent increase in the spreading of Virtual Reality allows a large number of consumers to experience this technology in several fields, from gaming to learning. In this paper we propose a framework that allows the user to freely move and interact in a virtual museum implemented with the HTC Vive. The goals of this work are: to investigate the possibility of using this approach for providing the user with a personalized and educative exploration of an artistic content, and to measure the immersivity and usability of the proposed system by means of a subjective experiment.","PeriodicalId":252212,"journal":{"name":"2018 7th European Workshop on Visual Information Processing (EUVIP)","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134060938","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"THe Gaussian Hypothesis in Subjective Quality Evaluation for Stereoscopic and 2D Video Content","authors":"Rania Bensaied, M. Mitrea","doi":"10.1109/EUVIP.2018.8611747","DOIUrl":"https://doi.org/10.1109/EUVIP.2018.8611747","url":null,"abstract":"The present paper investigates two controversial issues related to subjective video quality assessment, namely the relationship between continuous and discrete scale evaluations and the impact of the Gaussian hypothesis in modeling the scores assigned by the human observers. The theoretical background is obtained by reconsidering a non-linear random variable transformation formula established in our previous work and by instantiating it for both Gaussian and non-Gaussian cases. The experimental results are obtained by considering a total of 240 human observers, evaluating both stereoscopic and 2D video quality content, each of which considered at both high quality and low quality (as a priori expressed by objective measures). The results demonstrate that the Gaussian hypothesis cannot be accepted and that the relative error in the mean opinion score estimation can be reduced by 2-4% when considering a non-Gaussian model.","PeriodicalId":252212,"journal":{"name":"2018 7th European Workshop on Visual Information Processing (EUVIP)","volume":"180 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134312667","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
G. Atkinson, Thomas J. Thornton, Demitri I. C. Peynado, J. Ernst
{"title":"High-precision polarization measurements and analysis for machine vision applications","authors":"G. Atkinson, Thomas J. Thornton, Demitri I. C. Peynado, J. Ernst","doi":"10.1109/EUVIP.2018.8611762","DOIUrl":"https://doi.org/10.1109/EUVIP.2018.8611762","url":null,"abstract":"Polarization is a source of information that is steadily attracting attention in the field of computer vision due to its ability to tap into information not readily available in standard colour or greyscale cameras. Unfortunately, most existing data capture methods tend to suffer from either poor signal-to-noise ratio or long capture times. Further, most existing literature relies on making heavy assumptions about the polarizing properties of surfaces, which limits their application. This paper aims to optimise image capture conditions for polarization data in order to maximise the signal-to-noise ratio. Using the discovered optimal settings, a variety of images of different scenes are captured illustrating a range of reflectance properties typically overlooked previously. Such phenomena include inter-reflections, combined specular-diffuse reflection and surface conductance. The output from the paper is a set of key requirements and considerations necessary to further advance the field of polarization vision.","PeriodicalId":252212,"journal":{"name":"2018 7th European Workshop on Visual Information Processing (EUVIP)","volume":"15 5","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114102606","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Building a labeled dataset for recognition of handball actions using mask R-CNN and STIPS","authors":"Marina Ivasic-Kos, M. Pobar","doi":"10.1109/EUVIP.2018.8611642","DOIUrl":"https://doi.org/10.1109/EUVIP.2018.8611642","url":null,"abstract":"Building successful machine learning models depends on large amounts of training data that often needs to be labelled manually. We propose a method to efficiently build an action recognition dataset in the handball domain, focusing on minimizing the manual labor required to label the individual players performing the chosen actions. The method uses existing deep learning object recognition methods for player detection and combines the obtained location information with a player activity measure based on spatio-temporal interest points to track players that are performing the currently relevant action, here called active players. The method was successfully used on a challenging dataset of real-world handball practice videos, where the leading active player was correctly tracked and labeled in 84 % of cases.","PeriodicalId":252212,"journal":{"name":"2018 7th European Workshop on Visual Information Processing (EUVIP)","volume":"2012 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128008812","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Optimization of Hybrid Optics with Multilevel Phase Mask for Improved Depth of Focus Broadband Imaging","authors":"V. Katkovnik, Mykola Ponomarenko, K. Egiazarian","doi":"10.1109/EUVIP.2018.8611767","DOIUrl":"https://doi.org/10.1109/EUVIP.2018.8611767","url":null,"abstract":"This paper is devoted to designing of hybrid refractive/diffractive optics for high-quality imaging with improved depth of focus (DoF). The novelty of the concept is in sharing of the optical power of the refractive lens between the lens and a multilevel phase mask (MLM) as a diffractive optical element (DOE). The efficiency of the design is confirmed by numerical results. Broadband multiwavelength test-images are exploited for the design and testing of the system. It is shown that the obtained hybrid optical system in combination with computational inverse imaging provides both a better image quality (due to the robustness of the designed optics to chromatic aberrations) and an extended depth of focus as compared with the refractive lens and the diffractive lensless design with MLM.","PeriodicalId":252212,"journal":{"name":"2018 7th European Workshop on Visual Information Processing (EUVIP)","volume":"54 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115283432","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
A. Degerli, S. Aslan, Mehmet Yamaç, B. Sankur, M. Gabbouj
{"title":"Compressively Sensed Image Recognition","authors":"A. Degerli, S. Aslan, Mehmet Yamaç, B. Sankur, M. Gabbouj","doi":"10.1109/EUVIP.2018.8611657","DOIUrl":"https://doi.org/10.1109/EUVIP.2018.8611657","url":null,"abstract":"Compressive Sensing (CS) theory asserts that sparse signal reconstruction is possible from a small number of linear measurements. Although CS enables low-cost linear sampling, it requires non-linear and costly reconstruction. Recent literature works show that compressive image classification is possible in CS domain without reconstruction of the signal. In this work, we introduce a DCT base method that extracts binary discriminative features directly from CS measurements. These CS measurements can be obtained by using (i) a random or a pseudorandom measurement matrix, or (ii) a measurement matrix whose elements are learned from the training data to optimize the given classification task. We further introduce feature fusion by concatenating Bag of Words (BoW) representation of our binary features with one of the two state-of-the-art CNN-based feature vectors. We show that our fused feature outperforms the state-of-the-art in both cases.","PeriodicalId":252212,"journal":{"name":"2018 7th European Workshop on Visual Information Processing (EUVIP)","volume":"115 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-10-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114069070","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}