{"title":"Object tracking via the dynamic velocity Hough transform","authors":"P. Lappas, J. Carter, R. Damper","doi":"10.1109/ICIP.2001.958505","DOIUrl":"https://doi.org/10.1109/ICIP.2001.958505","url":null,"abstract":"Motion tracking is an important task in computer vision. A new technique, the dynamic velocity Hough transform (DVHT), for tracking of parametric objects is described that extends the velocity Hough transform (VHT) to cater for arbitrary motion. Like the VHT, the new technique processes the whole image sequence, gathering global evidence of motion and structure. However, we do not assume constant linear velocity but rather allow arbitrary velocity. The method tries to find an optimal, smooth trajectory in the parameter space with maximum energy, where the latter incorporates both the structure of the moving object and the smoothness of motion. The constrained optimisation problem is solved using a temporal (time-delay) dynamic programming algorithm. Tracking in noise is much superior to the standard Hough transform.","PeriodicalId":291827,"journal":{"name":"Proceedings 2001 International Conference on Image Processing (Cat. No.01CH37205)","volume":"94 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-10-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125259678","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
B. Kapralos, M. Jenkin, E. Milios, John K. Tsotsos
{"title":"Eyes 'n ears face detection","authors":"B. Kapralos, M. Jenkin, E. Milios, John K. Tsotsos","doi":"10.1109/ICIP.2001.958954","DOIUrl":"https://doi.org/10.1109/ICIP.2001.958954","url":null,"abstract":"We present a robust and portable visual-based skin and face detection system developed for use in a multiple speaker teleconferencing system, employing both audio and video cues. An omni-directional video sensor is used to provide a view of the entire visual hemisphere, thereby allowing for multiple dynamic views of all the participants. Regions of skin are detected using simple statistical methods, along with histogram color models for both skin and non-skin color classes. Regions of skin belonging to the same person are grouped together, and using simple spatial properties, the position of each person's face is inferred. Preliminary results suggest the system is capable of detecting human faces present in an omni-directional image despite the poor resolution inherent with such an omni-directional sensor.","PeriodicalId":291827,"journal":{"name":"Proceedings 2001 International Conference on Image Processing (Cat. No.01CH37205)","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-10-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125265130","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Mosaics of video sequences with moving objects","authors":"Chiou-Ting Hsu, Yu-Chun Tsan","doi":"10.1109/ICIP.2001.958509","DOIUrl":"https://doi.org/10.1109/ICIP.2001.958509","url":null,"abstract":"We propose an efficient method for creating a mosaic of a video sequence in the presence of moving objects. This method includes two principal processes. The first one removes the moving objects from the background, and as a side effect, obtains the global motion. This global motion provides a good initial estimation to the next stage. Second, we employ a feature-based technique to derive the precise global motion with eight projective parameters. The performance of our work is demonstrated by experiments.","PeriodicalId":291827,"journal":{"name":"Proceedings 2001 International Conference on Image Processing (Cat. No.01CH37205)","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-10-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125364576","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Y. Kawata, N. Niki, H. Ohmatsu, M. Kusumoto, R. Kakinuma, K. Mori, H. Nishiyama, K. Eguchi, M. Kaneko, N. Moriyama
{"title":"Computerized analysis of 3-D pulmonary nodule images in surrounding and internal structure feature spaces","authors":"Y. Kawata, N. Niki, H. Ohmatsu, M. Kusumoto, R. Kakinuma, K. Mori, H. Nishiyama, K. Eguchi, M. Kaneko, N. Moriyama","doi":"10.1109/ICIP.2001.958637","DOIUrl":"https://doi.org/10.1109/ICIP.2001.958637","url":null,"abstract":"We are developing computerized feature extraction and classification methods to analyze malignant and benign pulmonary nodules in three-dimensional (3-D) thoracic CT images. Surrounding structure features were designed to characterize the relationships between nodules and their surrounding structures such as vessel, bronchi, and pleura. Internal structure features were derived from CT density and 3-D curvatures to characterize the inhomogeneous of CT density distribution inside the nodule. The stepwise linear discriminant classifier was used to select the best feature subset from multidimensional feature spaces. The discriminant scores output from the classifier were analyzed by the receiver operating characteristic (ROC) method and the classification accuracy was quantified by the area, Az, under the ROC curve. We analyzed a data set of 248 pulmonary nodules in this study. The internal structure features (Az=0.88) were more effective than the surrounding structure features (Az=0.69) in distinguishing malignant and benign nodules. The highest classification accuracy (Az=0.94) was obtained in the combined internal and surrounding structure feature space. The improvement was statistically significant in comparison to classification in either the internal structure or the surrounding structure feature space alone. The results of this study indicate the potential of using combined internal and surrounding structure features for computer-aided classification of pulmonary nodules.","PeriodicalId":291827,"journal":{"name":"Proceedings 2001 International Conference on Image Processing (Cat. No.01CH37205)","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-10-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125397277","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"VISMap: an interactive image/video retrieval system using visualization and concept maps","authors":"William Chen, Shih-Fu Chang","doi":"10.1109/ICIP.2001.958187","DOIUrl":"https://doi.org/10.1109/ICIP.2001.958187","url":null,"abstract":"Images and videos can be indexed by multiple features at different levels, such as color, texture, motion and text annotation. Organizing this information into a system so that users can query effectively is a challenging and important problem. We present VISMap, a visual information seeking system that extends the traditional query paradigms of query-by-example and query-by-sketch and replaces the models of relevance feedback with principles from information visualization and concept representation. Users no longer perform lengthy \"one-shot\" queries or rely on hidden relevance feedback mechanisms. Instead, we provide a rich set of tools that allow users to construct personal views of the video database and directly visualize and manipulate various views and comprehend effects of individual query criteria on the final search results. The set of tools include: (1) a feature space browser for feature-based exploration and navigation; (2) a distance map for metric comparison and setting and (3) a novel concept map for query representation and creation.","PeriodicalId":291827,"journal":{"name":"Proceedings 2001 International Conference on Image Processing (Cat. No.01CH37205)","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-10-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125415341","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Improved image thresholding for object extraction in IR images","authors":"B. Kamgar-Parsi","doi":"10.1109/ICIP.2001.959156","DOIUrl":"https://doi.org/10.1109/ICIP.2001.959156","url":null,"abstract":"Object extraction from an IR image background is of great interest both to the military and the commercial sector. A convenient and popular approach to object extraction is image thresholding. In this paper, we describe a new and easy to implement approach for extracting object(s) in single frame IR images, which has many similarities to image thresholding. Both on the basis of theoretical considerations and experimental results, however, our approach appears to be noticeably more dependable than image thresholding for IR images.","PeriodicalId":291827,"journal":{"name":"Proceedings 2001 International Conference on Image Processing (Cat. No.01CH37205)","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-10-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126619973","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
C. Pattichis, C. Christodoulou, M. Pattichis, M. Pantziaris, A. Nicolaides
{"title":"An integrated system for the assessment of ultrasonic imaging atherosclerotic carotid plaques","authors":"C. Pattichis, C. Christodoulou, M. Pattichis, M. Pantziaris, A. Nicolaides","doi":"10.1109/ICIP.2001.959019","DOIUrl":"https://doi.org/10.1109/ICIP.2001.959019","url":null,"abstract":"The objective of this work is to develop a system that will facilitate the automated characterization of ultrasonic imaging carotid plaques for the identification of individuals with asymptomatic carotid stenosis at risk of stroke. A total of 166 images were collected which were classified into: symptomatic because of ipsilateral hemispheric symptoms, or asymptomatic because they were not connected with ipsilateral hemispheric events. Ten different texture feature sets were extracted: first order statistics, spatial gray level dependence matrices, gray level difference statistics, neighbourhood gray tone difference matrix, statistical feature matrix, Laws texture energy measures, fractal dimension texture analysis, Fourier power spectrum and shape parameters. A modular neural network classifier was developed composed of self-organizing map (SOM) classifiers, achieving an overall diagnostic yield of 76.4%. The results of this work show that it is possible to identify a group of patients at risk of stroke based on texture features.","PeriodicalId":291827,"journal":{"name":"Proceedings 2001 International Conference on Image Processing (Cat. No.01CH37205)","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-10-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126964581","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Processing concept queries with object motions in video databases","authors":"Chia-Han Lin, Arbee L. P. Chen","doi":"10.1109/ICIP.2001.958575","DOIUrl":"https://doi.org/10.1109/ICIP.2001.958575","url":null,"abstract":"This paper proposes an indexing and query processing approach for content-based video retrieval The concept hierarchy is used to provide high-level concepts for users to specify queries. The index buckets and suffix trees are proposed to index the static attributes and motions of video objects A flexible query processing strategy is designed based on the index structure. Also, strategies for processing approximate queries and queries involving object motions are considered.","PeriodicalId":291827,"journal":{"name":"Proceedings 2001 International Conference on Image Processing (Cat. No.01CH37205)","volume":"30 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-10-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115178641","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Extraction of rotation invariant signature based on fractal geometry","authors":"Yu Tao, T. Ioerger, Y. Tang","doi":"10.1109/ICIP.2001.959239","DOIUrl":"https://doi.org/10.1109/ICIP.2001.959239","url":null,"abstract":"A new method of feature extraction with a rotation invariant property is presented. One of the main contributions of this study is that a rotation invariant signature of 2D contours is selected based on fractal theory. The rotation invariant signature is a measure of the fractal dimensions, which is rotation invariant based on a series of central projection transform (CPT) groups. As the CPT is applied to a 2D object, a unique contour is obtained. In the unfolding process, this contour is further spread into a central projection unfolded curve, which can be viewed as a periodic function due to the different orientations of the pattern. We consider the unfolded curves to be non-empty and bounded sets in IR/sup n/, and the central projection unfolded curves with respect to the box computing dimension are rotation invariant. Some experiments with positive results have been conducted. This approach is applicable to a wide range of areas such as image analysis, pattern recognition etc.","PeriodicalId":291827,"journal":{"name":"Proceedings 2001 International Conference on Image Processing (Cat. No.01CH37205)","volume":"6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-10-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115336332","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A fast method for identifying graphical objects in large engineering drawings","authors":"A. Chakraborty","doi":"10.1109/ICIP.2001.959165","DOIUrl":"https://doi.org/10.1109/ICIP.2001.959165","url":null,"abstract":"Document processing and understanding is important for a variety of applications such as office automation, creation of electronic manuals, online documentation and annotation etc. However the understanding and identifying of graphical objects in large engineering drawings which often can be of the order of 10000/spl times/10000 pixels using traditional methods can be a very challenging and time consuming task due to the sheer size. We describe a method that circumvents that by using a hierarchical top-down approach for recognizing graphical objects. The speedup is almost of the order of 1000 and this is achieved by following a strategy that exploits the characteristics of such diagrams such as low pixel density.","PeriodicalId":291827,"journal":{"name":"Proceedings 2001 International Conference on Image Processing (Cat. No.01CH37205)","volume":"55 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-10-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116383164","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}