{"title":"Image classification using compression distance","authors":"Yuxuan Lan, R. Harvey","doi":"10.2312/VVG.20051023","DOIUrl":"https://doi.org/10.2312/VVG.20051023","url":null,"abstract":"The normalised compression distance measures the mutual compressibility of two signals. We show that this distance can be used for classification on real images. Furthermore, the same compressor can also operate on derived features with no further modification. We consider derived features consisting of trees indicating the containment and relative area of connected sets within the image. It had been previously postulated that such trees might be useful features, but they are too complicated for conventional classifiers. The new classifier operating on these trees produces results that are very similar to those obtained on the raw images thus allowing, for the first time, classification using the full trees.","PeriodicalId":137668,"journal":{"name":"International Conference on Vision, Video and Graphics","volume":"16 3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134258592","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Motion Analysis in Video: Dolls, Dynamic Cues and Modern Art","authors":"J. Collomosse, Peter A. Hall","doi":"10.2312/VVG.20051014","DOIUrl":"https://doi.org/10.2312/VVG.20051014","url":null,"abstract":"This paper addresses the problem of synthesising animations from video clips; in particular emphasising the motion of tracked objects. We introduce idynamic cuesi as a class of motion emphasis cue, encompassing traditional animation techniques such as anticipation and exaggeration. We present methods for automatically synthesising such cues within video premised upon the recovery of articulated gures, and the subsequent manipulation of the recovered pose trajectories. Additionally, we apply our motion emphasis framework to emulate artwork in the Futurist style, popularised by Duchamp.","PeriodicalId":137668,"journal":{"name":"International Conference on Vision, Video and Graphics","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128126361","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Cartoon-Style Rendering of Motion from Video","authors":"J. Collomosse, P. Hall","doi":"10.2312/vvg.20031016","DOIUrl":"https://doi.org/10.2312/vvg.20031016","url":null,"abstract":"The contribution of this paper is a novel non-photorealistic rendering (NPR) system capable of rendering motion within a video sequence in artistic styles. A variety of cartoon-style motion cues may be inserted into a video sequence, including augmentation cues (such as streak lines, ghosting, or blurring) and deformation cues (such as squash and stretch or drag effects). Users may select from the gamut of available styles by setting parameters","PeriodicalId":137668,"journal":{"name":"International Conference on Vision, Video and Graphics","volume":"33 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122123283","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A Region Adjacency Tree Approach to the Detection and Design of Fiducials","authors":"E. Costanza, J. Robinson","doi":"10.2312/vvg.20031009","DOIUrl":"https://doi.org/10.2312/vvg.20031009","url":null,"abstract":"We report a topological approach to fiducial recognition for real-time applications. Independence from geometry makes the system tolerant to severe distortion, and allows encoding of extra information. The method is based on region adjacency trees. After describing the mathematical foundations, we present a set of simulations to evaluate the algorithm and optimise the fiducial design.","PeriodicalId":137668,"journal":{"name":"International Conference on Vision, Video and Graphics","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121813490","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Digitisation to Presentation - Building Virtual Museum Exhibitions","authors":"M. Patel, M. White, K. Walczak, P. Sayd","doi":"10.2312/vvg.20031026","DOIUrl":"https://doi.org/10.2312/vvg.20031026","url":null,"abstract":"We describe an innovative system designed for museums to create, manage and present multimedia based representations of museum artefacts in virtual exhibitions both inside and outside museums. Our system creates digital surrogates through a novel stereo photogrammetry system with little user interaction. The resulting 3D objects are refined using state-of-the-art 3D modelling software configured for ease of use by museum staff. A repository of such digital surrogates is managed in an XML enabled relational database and provides the basis for the creation and presentation of virtual museum exhibitions, allowing current museum websites to evolve from a 2D to a 3D multimedia-rich domain. In this paper, we discuss the modelling and refinement processes which are based on stereo photogrammetry and the creation and visualisation of virtual museum exhibitions using virtual and augmented reality techniques.","PeriodicalId":137668,"journal":{"name":"International Conference on Vision, Video and Graphics","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114960246","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A Novel Form of Pointing Device","authors":"H. Cantzler, C. Hoile","doi":"10.2312/vvg.20031008","DOIUrl":"https://doi.org/10.2312/vvg.20031008","url":null,"abstract":"This paper presents a novel approach for man machine interaction applying real time computer vision techniques. We use a handheld camera to control the mouse cursor on a computer display. The camera captures an image of the display in its eld of view and this can be used to judge the camera’s position and orientation relative to the display. The problem is modelled as a plane-to-plane projection (homography). Once the mapping of the display in the camera view to the real world display is known, the intersection between the central axis of the camera and the surface of the display can be computed. The mouse pointer is then moved to the corresponding display position. This calculation can be iterated continuously to update the mouse cursor position as the camera position and orientation changes. The camera can then be used to control the mouse cursor just like a laser pointer controls a laser dot. A prototype has been developed to demonstrate the approach.","PeriodicalId":137668,"journal":{"name":"International Conference on Vision, Video and Graphics","volume":"48 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125173058","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Iterative Multi-Planar Camera Calibration: Improving Stability using Model Selection","authors":"J. Vigueras, M. Berger, Gilles Simon","doi":"10.2312/vvg.20031021","DOIUrl":"https://doi.org/10.2312/vvg.20031021","url":null,"abstract":"Tracking, or camera pose determination, is the main technical challenge in numerous applications in computer vision and especially in Augmented Reality. However, pose computation processes commonly exhibit some fluctuations and lack of precision in the estimation of the parameters. This leads to unpleasant visual impressions when augmented scenes are considered. In this paper, we propose an efficient and reliable method for real time camera tracking which avoid unpleasant statistical fluctuations. This method is based on the knowledge of a piecewise planar structure in the scene and makes use of model selection to reduce fluctuations. Videos are attached to this paper which prove the effectiveness of our approach.","PeriodicalId":137668,"journal":{"name":"International Conference on Vision, Video and Graphics","volume":"17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116228585","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Color Reduction by Using a new Self-Growing and Self-Organized Neural Network","authors":"A. Atsalakis, N. Papamarkos","doi":"10.2312/VVG.20051007","DOIUrl":"https://doi.org/10.2312/VVG.20051007","url":null,"abstract":"A new method for the reduction of the number of colors in a digital image is proposed. The new method is based on the developed of a new neural network classifier that combines the advantages of the Growing Neural Gas (GNG) and the Kohonen Self-Organized Feature Map (SOFM) neural networks. We call the new neural network: Self-Growing and SelfOrganized Neural Gas (SGONG). Its main advantage is that it defines the number of the created neurons and their topology in an automatic way. As a consecutive, isolated color classes, which may correspond to significant image details, can be obtained. The SGONG is fed by the color components and additional spatial features. To speed up the entire algorithm and to reduce memory requirements, a fractal scanning sub-sampling technique is used. The method is applicable to any type of color images and it can accommodate any type of color space.","PeriodicalId":137668,"journal":{"name":"International Conference on Vision, Video and Graphics","volume":"02 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130203899","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"3D Image Analysis and Synthesis at MPI Informatik","authors":"C. Theobalt, M. Magnor, H. Seidel","doi":"10.2312/VVG.20051011","DOIUrl":"https://doi.org/10.2312/VVG.20051011","url":null,"abstract":"In the talk, we give a brief overview of the research done in the Computer Graphics Group and the GraphicsOptics-Vision Group of the Max-Planck-Institut für Informatik in the area of 3D Image Analysis and Synthesis. In this context, we address the whole pipeline ranging from the acquisition of computational scene models, over the algorithmic processing of these scene descriptions, to their photo-realistic rendition in the computer. This paper illustrates the questions that we are trying to answer by means of one of our research projects, video-based rendering. We have developed a model-based system to acquire, reconstruct and render free-viewpoint videos of human actors that nicely illustrates the concept of 3D Image Analysis and Synthesis.","PeriodicalId":137668,"journal":{"name":"International Conference on Vision, Video and Graphics","volume":"314 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115501044","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Investigating Occlusion and Discretization Problems in Image-Based Blurring Techniques","authors":"B. Barsky, Michael J. Tobias, D. Horn, D. P. Chu","doi":"10.2312/vvg.20031013","DOIUrl":"https://doi.org/10.2312/vvg.20031013","url":null,"abstract":"Traditional computer graphics methods render images that appear sharp at all depths. Adding blur can add realism to a scene, provide a sense of scale, and draw a viewer’s attention to a particular region of a scene. Our image based blur algorithm needs to distinguish whether a portion of an image is either from a single object or is part of more than one object. This motivates two approaches to identify objects after an image has been rendered. We illustrate how these techniques can be used in conjunction with our image space method to add blur to a scene.","PeriodicalId":137668,"journal":{"name":"International Conference on Vision, Video and Graphics","volume":"03 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130051906","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}