{"title":"Comparison of Open and Free Video Compression Systems - A Performance Evaluation","authors":"Till Halbach","doi":"10.5220/0001809700740080","DOIUrl":"https://doi.org/10.5220/0001809700740080","url":null,"abstract":"This article gives a technical overview of two open and free video compression systems, Dirac and Theora I, and evaluates the rate distortion performance and visual quality of these systems regarding lossy and lossless compression, as well as intra-frame and inter-frame coding. The evaluation shows that there is a substantial performance gap of Theora and Dirac when compared to H.264and Motion JPEG2000-compliant reference systems. However, an algorithm subset of Dirac, Dirac Pro, achieves a performance comparable to that of Motion JPEG2000, and which can be less than one dB below the PSNR performance of H.264 with TV-size and HD video material. It is further shown that the reference implementations of the codecs of concern still have potential for efficiency improvements.","PeriodicalId":231479,"journal":{"name":"International Conference on Imaging Theory and Applications","volume":"36 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-09-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134288770","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"EPSNR for Objective Image Quality Measurements","authors":"Chulhee Lee, G. Seo, Sangwook Lee","doi":"10.5220/0001807800920095","DOIUrl":"https://doi.org/10.5220/0001807800920095","url":null,"abstract":"In this paper, we explore the possibility to apply a recently standardized method for objective video quality measurements to measure perceptual quality of still images. It has been known that the human visual system is more sensitive to edge degradation. We apply this standardized method to several image data sets which have subjective scores. The standardized method is compared with existing objective models for still images. Experimental results show that the standardized method shows better performance than the conventional PSNR and show similar performance compared to top-performance models.","PeriodicalId":231479,"journal":{"name":"International Conference on Imaging Theory and Applications","volume":"71 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-08-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127690820","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Magnetic Resonance Imaging of the Vocal Tract - Techniques and Applications","authors":"S. R. Ventura, D. Freitas, J. Tavares","doi":"10.5220/0001792901050110","DOIUrl":"https://doi.org/10.5220/0001792901050110","url":null,"abstract":"Magnetic resonance (MR) imaging has been used to analyse and evaluate the vocal tract shape through different techniques and with promising results in several fields. Our purpose is to demonstrate the relevance of MR and image processing for the vocal tract study. The extraction of contours of the air cavities allowed the set-up of a number of 3D reconstruction image stacks by means of the combination of orthogonally oriented sets of slices for each articulatory gesture, as a new approach to solve the expected spatial under sampling of the imaging process. In result these models give improved information for the visualization of morphologic and anatomical aspects and are useful for partial measurements of the vocal tract shape in different situations. Potential use can be found in Medical and therapeutic applications as well as in acoustic articulatory speech modelling. Magnetic Resonance (MR) improvements, in the past decades, allowed vocal tract imaging, making it currently one of the most promising tools in speech research. Speech is the most important instrument of human communication and interaction. Nevertheless, the knowledge about its production is far from being complete or even sufficient to describe the most relevant acoustic phenomena that are conditioned at morphological and dynamic levels. The anatomic and physiologic aspects of the vocal tract are claimed to be essential for a better understanding of this process. The quality and resolution of soft-tissues and the use of non-ionizing radiation are some of the most important advantages of MR imaging (Avila-García et al., 2004; Engwall, 2003). Several approaches have been used up to now for the study of the vocal tract based on MR images. Since the first study proposed by Baer et al. (1991), many MR techniques have been used (from static to dynamic studies, and more recently even done in real-time), starting by studies of vowel production (Badin et al., 1998; Demolin et al., 2000), followed by consonant production (Engwall, 2000b; Narayanan et al., 2004), for different languages such as French (Demolin et al., 1996; Serrurier & Badin, 2006), German (Behrends et al., 2001; Mády et al., 2001), and Japonese (Kitamura et al., 2005; Takemoto et al., 2003). The work presented in this paper, consisting basically in the static description of the vocal tract shape during sustained vowels and consonants and in the dynamic description of some syllables, is the first to report the application of MR imaging for the characterization of European Portuguese (EP). This study started in 2004, having attained a first series of results published in 2006 (Rua & Freitas, 2006). Our approach can be seen as a contribution to the wide area of articultory speech modeling, since it provides geometrical data to the acoustic modeling phase or research. In the articulatory speech research of EP a few studies of nasal vowels have been carried through, at the acoustic production and perceptual levels based on acous","PeriodicalId":231479,"journal":{"name":"International Conference on Imaging Theory and Applications","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-04-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130811304","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Robust Number Plate Recognition in Image Sequences","authors":"A. Zweng, M. Kampel","doi":"10.5220/0001801200560063","DOIUrl":"https://doi.org/10.5220/0001801200560063","url":null,"abstract":"License plate detection is done in three steps. The localization of the plate, the segmentation of the characters and the classification of the characters are the main steps to classify a license plate. Different algorithms for each of these steps are used depending on the area of usage. Corner detection or edge projection is used to localize the plate. Different algorithms are also available for character segmentation and character classification. A license plate is classified once for each car in images and in video streams, therefore it can happen that the single picture of the car is taken under bad lighting conditions or other bad conditions. In order to improve the recognition rate, it is not necessary to enhance character training or improve the localization and segmentation of the characters. In case of image sequences, temporal information of the existing license plate in consecutive frames can be used for statistical analysis to improve the recognition rate. In this paper an existing approach for a single classification of license plates and a new approach of license plate recognition in image sequences are presented. The motivation of using the information in image sequences and therefore classify one car multiple times is to have a more robust and converging classification where wrong single classifications can be suppressed.","PeriodicalId":231479,"journal":{"name":"International Conference on Imaging Theory and Applications","volume":"99 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129129558","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Automatic Recognition of Road Signs in Digital Images for GIS Update","authors":"A. Marçal, I. R. Goncalves","doi":"10.5220/0001790301290134","DOIUrl":"https://doi.org/10.5220/0001790301290134","url":null,"abstract":"A method for automatic recognition of road signs identified in digital video images is proposed. The method is based on features extracted from cumulative histograms and supervised classification. The training of the classifier is done with a small number of images (1 to 6) from each sign type. A practical experiment with 260 images and 26 different road sign was carried out. The average classification accuracy of the method with the standard settings was found to be 93.6%. The classification accuracy is improved to 96.2% by accepting the sign types ranked 1st and 2nd by the classifier, and to 97.4% by also accepting the sign type ranked 3rd . These results indicate that this can be a valuable tool to assist Geographic Information System (GIS) updating process based on Mobile Mapping System (MMS) data.","PeriodicalId":231479,"journal":{"name":"International Conference on Imaging Theory and Applications","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128687478","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Image Coding with Contourlet/Wavelet Transforms and SPIHT Algorithm - An Experimental Study","authors":"S. Nowak, P. Głomb","doi":"10.5220/0001767400370042","DOIUrl":"https://doi.org/10.5220/0001767400370042","url":null,"abstract":"We investigate the error resilience of the image coded with the wavelet/contourlet transform and the SPIHT algorithm. We experimentally verify the behaviour for two scenarios: partial decoding (as with scalable coding/transmission) and random sequence errors (as with transmission errors). Using a number of image quality metrics, we analyze the overall performance, as well as differences between each transform. We observe that error difference between transforms varies with length of decoded sequence.","PeriodicalId":231479,"journal":{"name":"International Conference on Imaging Theory and Applications","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-12-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130179622","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
J. A. Toque, Y. Sakatoku, Julia Anders, Y. Murayama, A. Ide-Ektessabi
{"title":"Multispectral Imaging - The Influence of Lighting Condition on Spectral Reflectance Reconstruction and Image Stitching of Traditional Japanese Paintings","authors":"J. A. Toque, Y. Sakatoku, Julia Anders, Y. Murayama, A. Ide-Ektessabi","doi":"10.5220/0001788000130020","DOIUrl":"https://doi.org/10.5220/0001788000130020","url":null,"abstract":"Illumination condition is one of the most important factors in imaging. Due to the relatively complex interaction occurring when an incident light is irradiated on the surface of an object, it has been a topic of researches and studies for quite a while now. In this study, its influence on the reconstruction of spectral reflectance and image stitching was explored. A traditional Japanese painting was used as the target. Spectral reflectance was estimated using pseudoinverse model from multispectral images captured with seven different filters with spectral features covering 380-850 nm wavelengths. It was observed that the accuracy of the estimation is dependent on the quality of multispectral images, which are greatly influenced by lighting conditions. High specular reflection on the target yielded large amount of estimation errors. In addition, the spectral feature of the filters was shown to be important. Data from at least four filters are necessary to get a satisfactory reconstruction. On the other hand, it was observed that in addition to specular reflection, the distribution of light highly affects image stitching. Image stitching is important especially when acquiring images of large objects. It was shown that multispectral images could be used for the analytical imaging of artworks.","PeriodicalId":231479,"journal":{"name":"International Conference on Imaging Theory and Applications","volume":"6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-12-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121312564","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Stereo Pair Matching of Archaeological Scenes using Phase Domain Methods","authors":"M. Alifragis, C. Tzafestas","doi":"10.5220/0001791500210028","DOIUrl":"https://doi.org/10.5220/0001791500210028","url":null,"abstract":"This paper conducts an experimental study on the application of some recent theories of image preprocessing and analysis in the frequency domain, particularly the phase congruency and monogenic filtering methods. Our goal was to examine the performance of such methods in a stereo matching problem setting, with photos of complicated scenes. Two objects were used: a scene of an ancient Greek temple of Acropolis and the outside scene of the gate of an ancient theatre. Due to the complex structure of the photographed object, classic techniques used for feature detection and matching give poor results. The phase-domain approach followed in this paper is based on the phase-congruency method for feature extraction, together with monogenic filtering and a new correlation measure in the frequency domain for image correspondence and stereo matching. Comparative results show that the three-dimensional models of the scene computed when applying these phase domain methods are much more detailed and consistent as compared to the models obtained when using classic approaches or the SIFT based techniques, which give poor depth representation and less accurate metric","PeriodicalId":231479,"journal":{"name":"International Conference on Imaging Theory and Applications","volume":"30 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116628471","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Active Contours with Optical Flow and Primitive Shape Priors for Echocardiographic Imagery","authors":"A. Hamou, M. El-Sakka","doi":"10.5220/0001804201110118","DOIUrl":"https://doi.org/10.5220/0001804201110118","url":null,"abstract":"Accurate delineation of object borders is highly desirable in echocardiography. Among other model-based techniques, active contours (or snakes) provide a unique and powerful approach to image analysis. In this work, we propose the use of a new external energy for a GVF snake, consisting of the optical flow data of moving heart structures (i.e. the perceived movement). This new external energy provides more information to the active contour model to combat noise in moving sequences. An automated primitive shape prior mechanism is also introduced, which further improves the results when dealing with especially noisy echocardiographic image cines. Results were compared with that of expert manual segmentations yielding promising sensitivities and system accuracies.","PeriodicalId":231479,"journal":{"name":"International Conference on Imaging Theory and Applications","volume":"120 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117169751","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Towards Computer Assisted Cardiac Catheterization - How 3D Visualization Supports It","authors":"K. Drechsler, C. O. Laura, G. Sakas","doi":"10.5220/0001771801590165","DOIUrl":"https://doi.org/10.5220/0001771801590165","url":null,"abstract":"Although cardiac catheterization procedures take place under x-ray guidance, the doctor is almost blind. Vessels are almost invisible until he injects a contrast agent and looking only at 2D x-ray images and reconstructing a 3D image in his head makes it error prone and tedious. Only experienced doctors are able to accomplish this procedure with the expected results. This paper describes our preliminary work and work in progress to support doctors during cardiac catheterizations using 3D visualization.","PeriodicalId":231479,"journal":{"name":"International Conference on Imaging Theory and Applications","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-11-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129388826","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}