{"title":"Calibration of an industrial vision system using an ellipsoid","authors":"J. Heather","doi":"10.1109/ICIP.2014.7025700","DOIUrl":"https://doi.org/10.1109/ICIP.2014.7025700","url":null,"abstract":"A robust multi-camera calibration algorithm developed for an industrial vision system is described. An ellipsoid with a simple surface pattern and accurately known geometry is used as a calibration target. Our algorithm automatically detects the presence of the ball on the conveyor and accurately determines the position of its outline and marker lines in each image frame using efficient image processing techniques. A fast least-squares minimization is then performed to determine the optimal camera and motion parameters. The method is fully automatic and requires no human interaction or guidance, helping to minimize machine setup and maintenance times. The calibration algorithm has been demonstrated on real image captures and performance is quantified using simulated image sequences.","PeriodicalId":6856,"journal":{"name":"2014 IEEE International Conference on Image Processing (ICIP)","volume":"363 1","pages":"3444-3448"},"PeriodicalIF":0.0,"publicationDate":"2014-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"76570603","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Junhui Hou, Lap-Pui Chau, Ying He, N. Magnenat-Thalmann
{"title":"Low-rank based compact representation of motion capture data","authors":"Junhui Hou, Lap-Pui Chau, Ying He, N. Magnenat-Thalmann","doi":"10.1109/ICIP.2014.7025296","DOIUrl":"https://doi.org/10.1109/ICIP.2014.7025296","url":null,"abstract":"In this paper, we propose a practical, elegant and effective scheme for compact mocap data representation. Guided by our analysis of the unique properties of mocap data, the input mocap sequence is optimally segmented into a set of subsequences. Then, we project the subsequences onto a pair of computational orthogonal matrices to explore strong low-rank characteristic within and among the subsequences. The experimental results show that the proposed scheme is much more effective for reducing the data size, compared with the existing techniques.","PeriodicalId":6856,"journal":{"name":"2014 IEEE International Conference on Image Processing (ICIP)","volume":"18 1","pages":"1480-1484"},"PeriodicalIF":0.0,"publicationDate":"2014-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"76102861","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Luis Kuhn Cuellar, Stefan Pfeffer, Yuxiang Chen, Friedrich Forster
{"title":"Automated detection of polysomes in cryoelectron tomography","authors":"Luis Kuhn Cuellar, Stefan Pfeffer, Yuxiang Chen, Friedrich Forster","doi":"10.1109/ICIP.2014.7025418","DOIUrl":"https://doi.org/10.1109/ICIP.2014.7025418","url":null,"abstract":"Ribosomes and messenger RNA assemble to polysomes during protein synthesis. Cryoelectron tomography enables detection and identification of large macromolecular complexes under physiological conditions making the method uniquely suitable to study the supercomplexes that govern translation of mRNA into proteins. Here, we describe a method for automated assignment of polysomes in cryoelectron tomograms using the positions and orientations of ribosomes, as localized by template matching on tomographic data, as input. On the basis of a training dataset of expert-curated polysomes in cryoelectron tomograms, we define the relative 3D arrangements of neighboring ribosomes in polysomes. This prior distribution is used in a probabilistic framework for polysome assignment: the localized ribosomes from a tomogram are represented as a graph of which the edge weights are defined by the prior distribution. A Markov Random Field is embedded on the graph structure, and a message-passing algorithm is used to infer a polysome-label for each ribosome, i.e., to cluster ribosomes into polysomes. The performance of the method is assessed based on simulated tomograms and experimental tomograms indicating that polysome detection is reliable for typical signal-to-noise ratios of cryoelectron tomograms.","PeriodicalId":6856,"journal":{"name":"2014 IEEE International Conference on Image Processing (ICIP)","volume":"47 1","pages":"2085-2089"},"PeriodicalIF":0.0,"publicationDate":"2014-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"76191050","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Weakly supervised pedestrian detector training by unsupervised prior learning and cue fusion in videos","authors":"K. K. Htike, David C. Hogg","doi":"10.1109/ICIP.2014.7025474","DOIUrl":"https://doi.org/10.1109/ICIP.2014.7025474","url":null,"abstract":"The growth in the amount of collected video data in the past decade necessitates automated video analysis for which pedestrian detection plays a key role. Training a pedestrian detector using supervised machine learning requires tedious manual annotation of pedestrians in the form of precise bounding boxes. In this paper, we propose a novel weakly supervised algorithm to train a pedestrian detector that only requires annotations of estimated centers of pedestrians instead of bounding boxes. Our algorithm makes use of a pedestrian prior learnt in an unsupervised way from the video and this prior is fused with the given weak supervision information in a principled manner. We show on publicly available datasets that our weakly supervised algorithm reduces the cost of manual annotation by over 4 times while achieving similar performance to a pedestrian detector trained with bounding box annotations.","PeriodicalId":6856,"journal":{"name":"2014 IEEE International Conference on Image Processing (ICIP)","volume":"120 1","pages":"2338-2342"},"PeriodicalIF":0.0,"publicationDate":"2014-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"87874760","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Hierarchical anchoring of motion fields for fully scalable video coding","authors":"Dominic Rüfenacht, R. Mathew, D. Taubman","doi":"10.1109/ICIP.2014.7025643","DOIUrl":"https://doi.org/10.1109/ICIP.2014.7025643","url":null,"abstract":"Traditional video codecs anchor motion fields in the frame that is to be predicted, which is natural in a non-scalable context. In this paper, we propose a hierarchical anchoring of motion fields at reference frames, which allows to “reuse” them at finer temporal levels - a very desirable property for temporal scalability. The main challenge using this approach is that the motion fields need to be warped to the target frames, leading to disocclusions and motion folding in the warped motion fields. We show how to resolve motion folding ambiguities that occur in the vicinity of moving object boundaries by using breakpoint fields that have recently been proposed for the scalable coding of motion. During the motion field warping process, we obtain disocclusion and folding maps on-the-fly, which are used to control the temporal update step of the Haar wavelet. Results on synthetic data show that the proposed hierarchical anchoring scheme outperforms the traditional way of anchoring motion fields.","PeriodicalId":6856,"journal":{"name":"2014 IEEE International Conference on Image Processing (ICIP)","volume":"28 1","pages":"3180-3184"},"PeriodicalIF":0.0,"publicationDate":"2014-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"87943895","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
T. Eftestøl, Frode Måløy, K. Engan, Lasya Priya Kotu, L. Woie, S. Ørn
{"title":"A texture-based probability mapping for localisation of clinically important cardiac segments in the myocardium in cardiac magnetic resonance images from myocardial infarction patients","authors":"T. Eftestøl, Frode Måløy, K. Engan, Lasya Priya Kotu, L. Woie, S. Ørn","doi":"10.1109/ICIP.2014.7025451","DOIUrl":"https://doi.org/10.1109/ICIP.2014.7025451","url":null,"abstract":"This paper presents a novel method for the identification of myocardial regions associated with increased risk of life threatening arrhythmia in patients with healed myocardial infarction assessed by late enhanced gadolinium magnetic resonance images. A probability mapping technique is used to create images where each pixel value corresponds to the probability of that pixel representing damaged myocardium. Cardiac segments are defined as the set of pixel positions associated with probability values between a lower and an upper threshold. From the corresponding pixels in the original images several features are calculated. The features studied here are the relative size and entropy values based on histograms with varying number of bins. Features calculated for a specific cardiac segment are compared between patients with high and low risk of arrhythmia. The results from comparing a large number of cardiac segments indicate that the entropy measure has a better localisation property compared to the relative size of the myocardial damage, and that the localisation is more focused for fewer number of bins in the entropy calculation.","PeriodicalId":6856,"journal":{"name":"2014 IEEE International Conference on Image Processing (ICIP)","volume":"1 1","pages":"2227-2231"},"PeriodicalIF":0.0,"publicationDate":"2014-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"87121589","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Enhanced spatio-temporal video copy detection by combining trajectory and spatial consistency","authors":"Savas Özkan, E. Esen, G. Akar","doi":"10.1109/ICIP.2014.7025511","DOIUrl":"https://doi.org/10.1109/ICIP.2014.7025511","url":null,"abstract":"The recent improvements on internet technologies and video coding techniques cause an increase in copyright infringements especially for video. Frequently, image-based approaches appear as an essential solution due to the fact that joint usage of quantization-based indexing and weak geometric consistency stages give a capability to compare duplicate videos quickly. However, exploiting purely spatial content ignores the temporal variation of video. In this work, we propose a system that combines the state-of-the-art quantization-based indexing scheme with a novel trajectory-based geometric consistency on spatio-temporal features. This combination improves duplicate video matching task significantly. Briefly, spatial mean and variance of the trajectories are incorporated to establish a weak geometric consistency among pair of frames. To show the success of the proposed method, content-based video copy detection field is selected and TRECVID 2009 dataset is utilized. The experimental results show that constituting trajectory-based consistency on corresponding feature pairs outperforms the performances of merely utilizing spatiotemporal signature and visual signature with enhanced weak geometric consistency.","PeriodicalId":6856,"journal":{"name":"2014 IEEE International Conference on Image Processing (ICIP)","volume":"50 1","pages":"2527-2531"},"PeriodicalIF":0.0,"publicationDate":"2014-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"87594605","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Degree of loop assessment in microvideo","authors":"Shumpei Sano, T. Yamasaki, K. Aizawa","doi":"10.1109/ICIP.2014.7026049","DOIUrl":"https://doi.org/10.1109/ICIP.2014.7026049","url":null,"abstract":"This paper presents a degree-of-loop assessment method for microvideo clips. Loop video is one of the popular features in microvideo, but there are so many non-loop video tagged with “loop” on microvideo services. This is because upload-ers or spammers also know that loop video is popular and they want to draw attention from viewers. In this paper, we statistically analyze the scene dynamics of the video by using color, optical flow, saliency maps, and evaluate the degree-of-loop. We have collected more than 1,000 video clips from Vine and subjectively evaluated their degree-of-loop. Experimental results show that our proposed algorithm can classify loop/non-loop video with 85.7% accuracy and categorize them into five degree-of-loop categories with 61.5% accuracy.","PeriodicalId":6856,"journal":{"name":"2014 IEEE International Conference on Image Processing (ICIP)","volume":"71 1","pages":"5182-5186"},"PeriodicalIF":0.0,"publicationDate":"2014-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"88100618","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
M. Tikekar, Chao-Tsung Huang, V. Sze, A. Chandrakasan
{"title":"Energy and area-efficient hardware implementation of HEVC inverse transform and dequantization","authors":"M. Tikekar, Chao-Tsung Huang, V. Sze, A. Chandrakasan","doi":"10.1109/ICIP.2014.7025421","DOIUrl":"https://doi.org/10.1109/ICIP.2014.7025421","url":null,"abstract":"High Efficiency Video Coding (HEVC) inverse transform for residual coding uses 2-D 4×4 to 32×32 transforms with higher precision as compared to H.264/AVC's 4×4 and 8×8 transforms resulting in an increased hardware complexity. In this paper, an energy and area-efficient VLSI architecture of an HEVC-compliant inverse transform and dequantization engine is presented. We implement a pipelining scheme to process all transform sizes at a minimum throughput of 2 pixel/cycle with zero-column skipping for improved throughput. We use data-gating in the 1-D Inverse Discrete Cosine Transform engine to improve energy-efficiency for smaller transform sizes. A high-density SRAM-based transpose memory is used for an area-efficient design. This design supports decoding of 4K Ultra-HD (3840×2160) video at 30 frame/sec. The inverse transform engine takes 98.1 kgate logic, 16.4 kbit SRAM and 10.82 pJ/pixel while the dequantization engine takes 27.7 kgate logic, 8.2 kbit SRAM and 1.10 pJ/pixel in 40 nm CMOS technology. Although larger transforms require more computation per coefficient, they typically contain a smaller proportion of non-zero coefficients. Due to this trade-off, larger transforms can be more energy-efficient.","PeriodicalId":6856,"journal":{"name":"2014 IEEE International Conference on Image Processing (ICIP)","volume":"21 1","pages":"2100-2104"},"PeriodicalIF":0.0,"publicationDate":"2014-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"86237186","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Emmanuel Soubies, L. Blanc-Féraud, S. Schaub, G. Aubert
{"title":"Sparse reconstruction from Multiple-Angle Total Internal Reflection fluorescence Microscopy","authors":"Emmanuel Soubies, L. Blanc-Féraud, S. Schaub, G. Aubert","doi":"10.1109/ICIP.2014.7025575","DOIUrl":"https://doi.org/10.1109/ICIP.2014.7025575","url":null,"abstract":"Super-resolution microscopy techniques allow to overstep the diffraction limit of conventional optics. Theses techniques are very promising since they give access to the visualisation of finer structures which is of fundamental importance in biology. In this paper we deal with Multiple-Angle Total Internal Reflection Microscopy (MA-TIRFM) which allows to reconstruct 3D sub-cellular structures of a single layer of ~ 300 nm behind the glass coverslip with a high axial resolution. The 3D volume reconstruction from a set of 2D measurements is an ill-posed inverse problem and a regularization is essential. Our aim in this work is to propose a new reconstruction method for sparse structures robust to Poisson noise and background fluorescence. The sparse property of the solution can be seen as a regularization using the `£° norm'. In order to solve this combinatorial problem, we propose a new algorithm based on smoothed `£° norm' allowing minimizing a non convex energy, composed of the Kullback-Leibler divergence data term and the £° regularization term, in a Graduated Non Convexity framework.","PeriodicalId":6856,"journal":{"name":"2014 IEEE International Conference on Image Processing (ICIP)","volume":"12 1","pages":"2844-2848"},"PeriodicalIF":0.0,"publicationDate":"2014-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"86555304","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}