{"title":"Efficient computation of 3D moments","authors":"Bingcheng Li, Songde Ma","doi":"10.1109/ICPR.1994.576218","DOIUrl":"https://doi.org/10.1109/ICPR.1994.576218","url":null,"abstract":"In this paper, we proposed a linear-transform-based method to compute 3D moments. Instead of computing the conventional moments directly, we compute the so-called LT moments which are related with the conventional moments by three linear transforms. By choosing appropriately these transforms, it is shown that the computational complexity can be considerably reduced.","PeriodicalId":312019,"journal":{"name":"Proceedings of 12th International Conference on Pattern Recognition","volume":"69 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1994-10-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121704503","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Reconstruction of multiple overlapping surfaces via standard regularization techniques","authors":"M. Shizawa","doi":"10.1109/ICPR.1994.576288","DOIUrl":"https://doi.org/10.1109/ICPR.1994.576288","url":null,"abstract":"A fundamental extension of the standard regularization technique is proposed for making data approximations using multivalued functions which are essential for solving the transparency problems in computational vision. Conventional standard regularization techniques can approximate scattered data by using a single-valued function which is smooth everywhere in the domain. However, to incorporate discontinuities of the functions, it is necessary to introduce the line process or an equivalent technique to break the coherence or smoothness of the approximating functions. Multilayer representations have been used in reconstruction of multiple overlapping surfaces. However this technique should incorporate auxiliary fields for segmenting given data. Furthermore, these two different approaches both have the difficulty implementing optimizations of their energy functionals since they always become nonquadratic, nonconvex minimization problems with respect to an unknown surface and auxiliary field parameters. This paper shows that by using a direct representation of multivalued functions, data approximation made using a multivalued function can be reduced to minimizations of a single quadratic convex functional. Therefore, since the Euler-Lagrange equation of the functional becomes linear in this case, it is possible to benefit from simple relaxation techniques of guaranteed convergence to the optimal solution.","PeriodicalId":312019,"journal":{"name":"Proceedings of 12th International Conference on Pattern Recognition","volume":"53 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1994-10-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126551378","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Using contextual information to set control parameters of a vision process","authors":"S. Houzelle, T. Strat, P. Fua, M. Fischler","doi":"10.1109/ICPR.1994.576460","DOIUrl":"https://doi.org/10.1109/ICPR.1994.576460","url":null,"abstract":"This paper presents a novel approach to supervised learning of image-understanding tactics based on the use of contextual information. We use a database to store past experiences. From this database and the context elements computed from the task specifications and the input data, we determine whether or not an algorithm is applicable, and which parameters are suitable for it. The database is continuously updated with information of success or failure of the system.","PeriodicalId":312019,"journal":{"name":"Proceedings of 12th International Conference on Pattern Recognition","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1994-10-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131968884","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Learning, recognition and navigation from a sequence of infrared images","authors":"Nicolas Milhaud, G. Medioni","doi":"10.1109/ICPR.1994.576457","DOIUrl":"https://doi.org/10.1109/ICPR.1994.576457","url":null,"abstract":"We address the problem in which an autonomous system equipped with a single infrared camera \"learns\" a designated rigid 3D scene so that, at another time, it can recognize it and guide itself relative to the reconstructed scene, starting from an approximately known viewpoint, to reach a give destination. This scenario is relevant to several domains, in particular military missions and robotic navigation. The goal of our system is to realise a real-time implementation (on special hardware). In this paper, we describe the software version of such an automated system and show results on real infrared images.","PeriodicalId":312019,"journal":{"name":"Proceedings of 12th International Conference on Pattern Recognition","volume":"271 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1994-10-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131551370","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Singularities of the visual motion field: 3D rotation or 3D translation","authors":"Richard P. Wildes","doi":"10.1109/ICPR.1994.576383","DOIUrl":"https://doi.org/10.1109/ICPR.1994.576383","url":null,"abstract":"This paper addresses the manner in which the structure and motion of the 3D world are manifest in the stable singularities of its imaged visual motion field. By focusing on stable structures the approach enjoys a degree of robustness and invariance to the particulars of recovering visual motion. The specific results to be described address situations where an optical sensor is undergoing pure rotational or pure translational motion through its environment. For the case of pure rotation it is shown that the singularities provide information about the axes, senses and relative magnitudes of rotation. For the case of pure translation it is shown that the singularities provide information about the shape of viewed surfaces as well as information about the translation itself. Finally, empirical results demonstrate the feasibility of applying the analysis to natural images.","PeriodicalId":312019,"journal":{"name":"Proceedings of 12th International Conference on Pattern Recognition","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1994-10-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132831528","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Recognizing facial expressions by spatio-temporal analysis","authors":"Yaser Yaccob, Larry S. Davis","doi":"10.1109/ICPR.1994.576429","DOIUrl":"https://doi.org/10.1109/ICPR.1994.576429","url":null,"abstract":"An approach for analysis and representation of facial dynamics for recognition of facial expressions from image sequences is proposed. The algorithms the authors develop utilize optical flow computation to identify the direction of rigid and non-rigid motions that are caused by human facial expressions. A mid-level symbolic representation that is motivated by linguistic and psychological considerations is developed. Recognizing six facial expressions, as well as eye blinking, are demonstrated on a collection of image sequences.","PeriodicalId":312019,"journal":{"name":"Proceedings of 12th International Conference on Pattern Recognition","volume":"234 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1994-10-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133172660","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Estimation of 3-D parametric models from shading image using genetic algorithms","authors":"H. Saito, N. Tsunashima","doi":"10.1109/ICPR.1994.576394","DOIUrl":"https://doi.org/10.1109/ICPR.1994.576394","url":null,"abstract":"In this paper, a method for estimating parameters of a 3-D shape from a 2-D shading image using a genetic algorithms (GAs) is proposed. The shape of the object is represented by a superquadrics model, and then the model parameters are coded for application to GAs. The coded string is evaluated according to the similarity of the shading image calculated from the 3-D model shape represented by the parameters to the given 2-D shading image. By applying the GAs to the optimization of the evaluation value, the string having the minimum difference can be found. The parameters are estimated from some shading images of various 3-D shapes by using the proposed method, and the results are presented.","PeriodicalId":312019,"journal":{"name":"Proceedings of 12th International Conference on Pattern Recognition","volume":"150 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1994-10-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124195693","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Using grey-level models to improve active shape model search","authors":"Tim Cootes, C. Taylor","doi":"10.1109/ICPR.1994.576227","DOIUrl":"https://doi.org/10.1109/ICPR.1994.576227","url":null,"abstract":"We describe methods for using flexible models to locate structures in images. We have previously described statistical models of shape and shape variability which can be used for this purpose (active shape models). In this paper we show how statistical models of grey-level appearance can be incorporated, leading to improved reliability and accuracy. We describe experiments designed to: 1) test how well an active shape model can locate an object in a new image; 2) to assess the effects on performance of varying the model parameters; and 3) to compare the results using grey-level models with those using a search for strongest edges. The results demonstrate that the addition of grey-level models leads to considerable improvement over earlier schemes.","PeriodicalId":312019,"journal":{"name":"Proceedings of 12th International Conference on Pattern Recognition","volume":"42 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1994-10-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117048975","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Spatio-chromatic model for colour image processing","authors":"S. Wolf, R. Ginosar, Y. Zeevi","doi":"10.1109/ICPR.1994.576372","DOIUrl":"https://doi.org/10.1109/ICPR.1994.576372","url":null,"abstract":"Traditional colour image enhancement techniques, have ignored the design of the human visual system and have treated colour images as three independent grayscale images. Contrary to this the authors first discuss an opponent colour model of the early visual system and demonstrate it's use in spatial enhancement of both image brightness and chroma. The opponent transform is calculated such that cone redundancies are reduced and information compressed, similar to the way it is performed in the visual system. This results in an achromatic and two opponent chromatic channels. Laplacian based edge enhancement is then performed independently on each channel. The spatial enhancement of chroma highlights small colour detail and enhances colour edges thus increasing image detail in areas where achromatic enhancement has no effect. Due to spatial limitations of the visual system, the best enhanced image is obtained when both achromatic and chromatic edge enhancement are used together.","PeriodicalId":312019,"journal":{"name":"Proceedings of 12th International Conference on Pattern Recognition","volume":"69 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1994-10-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116040271","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A new and efficient iterative approach to image matching","authors":"Zhengyou Zhang","doi":"10.1109/ICPR.1994.576359","DOIUrl":"https://doi.org/10.1109/ICPR.1994.576359","url":null,"abstract":"This paper addresses the problem of matching two images with unknown epipolar geometry. A new and efficient iterative algorithm is proposed, which is a part of the author's project on developing a robust image matching technique. The author defines a new measure of matching support, which allows less contribution and higher tolerance of deformation with respect to affine transformations from distant matches than from nearby ones. A new strategy for updating matches is developed, which only selects those matches having both high matching support and low matching ambiguity. The update strategy is different from the classical \"winner-take-all\", which evolves too soon and is easily stuck at a local minimum, and also from \"loser-take-nothing\", which is usually very slow. The proposed algorithm has been tested with two dozen image pairs of very different types of scenes, and very good results have been obtained. It works remarkably well in a scene with many repetitive patterns.","PeriodicalId":312019,"journal":{"name":"Proceedings of 12th International Conference on Pattern Recognition","volume":"29 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1994-10-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121999390","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}