MULTIMEDIA '00Pub Date : 2000-11-04DOI: 10.1145/357744.357930
Gang Wei, I. Sethi
{"title":"Omni-face detection for video/image content description","authors":"Gang Wei, I. Sethi","doi":"10.1145/357744.357930","DOIUrl":"https://doi.org/10.1145/357744.357930","url":null,"abstract":"An omni-face detection scheme for image/video content description is proposed in this paper. It provides the ability to extract high-level features in terms of human activities rather than low-level features like color, texture and shape. The system relies on an omni-face detection system capable of locating human faces over a broad range of views in color images or videos with complex scenes. It uses the presence of skin-tone pixels coupled with shape, edge pattern and face-specific features to locate faces. The main distinguishing contribution of this work is being able to detect faces irrespective of their poses, including frontal-view and side-view, whereas contemporary systems deal with frontal-view faces only. The other novel aspects of the work lie in its iterative candidate filtering to segment objects from extraneous region, the use of Hausdorff distance-based normalized similarity measure to identify side-view facial profiles, and the exploration of hidden Markov model (HMM) to verify the presence of a side-view face. Image and video can be assigned with semantic descriptors based on human face information for later indexing and retrieval.","PeriodicalId":234597,"journal":{"name":"MULTIMEDIA '00","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123918720","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
MULTIMEDIA '00Pub Date : 2000-11-04DOI: 10.1145/357744.357762
C. Dorai, R. Bolle, N. Dimitrova, L. Agnihotri, Gang Wei
{"title":"On the evolution of videotext description scheme and its validation experiments for MPEG-7","authors":"C. Dorai, R. Bolle, N. Dimitrova, L. Agnihotri, Gang Wei","doi":"10.1145/357744.357762","DOIUrl":"https://doi.org/10.1145/357744.357762","url":null,"abstract":"Videotext refers to text superimposed on still images and video frames, and can be used in many MPEG-7 applications that deal with archival and delivery of images and video. It can be used to annotate and index large video and image collections, and enables text based search, automatic video logging, and video cataloging. This paper describes the joint work of IBM and Philips Research Laboratories on designing an MPEG-7 description scheme based on videotext. It describes the elements comprising the Videotext DS and discusses validation experiments performed to demonstrate the effectiveness of the DS in video browsing and classification tasks.","PeriodicalId":234597,"journal":{"name":"MULTIMEDIA '00","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134535630","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
MULTIMEDIA '00Pub Date : 2000-11-04DOI: 10.1145/357744.361911
Renato O. Stehling, M. Nascimento, A. Falcão
{"title":"On “shapes” of colors for content-based image retrieval","authors":"Renato O. Stehling, M. Nascimento, A. Falcão","doi":"10.1145/357744.361911","DOIUrl":"https://doi.org/10.1145/357744.361911","url":null,"abstract":"Color is a commonly used feature for realizing content-based image retrieval (CBIR). Towards this goal, this paper presents a new approach for CBIR which is based on well known and widely used color histograms. Contrasting to previous approaches, such as using a single color histogram for the whole image, or local color histograms for a fixed number of image cells, the one we propose (named Color Shape) uses a variable number of histograms, depending only on the actual number of colors present in the image. Our experiments using a large set of heterogeneous images and pre-defined query/answer sets show that the Color Shape approach offers good retrieval quality with relatively low space overhead, outperforming previous approaches.","PeriodicalId":234597,"journal":{"name":"MULTIMEDIA '00","volume":"24 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134474904","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
MULTIMEDIA '00Pub Date : 2000-11-04DOI: 10.1145/357744.357934
Yi Zhang, Tat-Seng Chua
{"title":"Detection of text captions in compressed domain video","authors":"Yi Zhang, Tat-Seng Chua","doi":"10.1145/357744.357934","DOIUrl":"https://doi.org/10.1145/357744.357934","url":null,"abstract":"In this paper, we propose a new method for detecting text captions in MPEG video stream. It is based on the observation that text captions in video frames typically possess high contrast against the background for them to be visible. The method operates on DCT coefficients in MPEG domain. The main contribution of this work is in developing a binarzied contrast feature domain in which the presence of text in video frames can be highlighted. A weighting function is defined to deduce the probability of a text frame. From the set of representative text frames located, further techniques are developed to outline, segment, and recognise text captions. The techniques have been tested on commercial and news videos from the MPEG7 data set and local TCS news video. The results demonstrate that our approach is effective in detecting and locating text captions in general video.","PeriodicalId":234597,"journal":{"name":"MULTIMEDIA '00","volume":"06 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117272837","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
MULTIMEDIA '00Pub Date : 2000-11-04DOI: 10.1145/357744.357758
D. Park, Yoon Seok Jeon, C. Won
{"title":"Efficient use of local edge histogram descriptor","authors":"D. Park, Yoon Seok Jeon, C. Won","doi":"10.1145/357744.357758","DOIUrl":"https://doi.org/10.1145/357744.357758","url":null,"abstract":"The purpose of this paper is to show how the edge histogram descriptor for MPEG-7 can be efficiently utilized for image matching. Since the edge histogram descriptor recommended for the MPEG-7 standard represents only local edge distribution in an image, the matching performance for image retrieval may not be satisfactory. In this paper, to increase the matching performance, we propose to use the global and semi-local edge histograms generated directly from the local histogram bins. Then, the global, semi-global, and local histograms of two images are compared to evaluate the similarity measure. Since we exploit the absolute locations of edge in the image as well as its global composition, the proposed matching method is considered to be a more image content-based retrieval. Experimental results support this claim. Experiments on test images for MPEG-7 core experiment show that the proposed method yields better retrieval performance especially for semantic similarity.","PeriodicalId":234597,"journal":{"name":"MULTIMEDIA '00","volume":"6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126138162","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
MULTIMEDIA '00Pub Date : 2000-11-04DOI: 10.1145/357744.357892
P. Wohlmacher
{"title":"Digital certificates: a survey of revocation methods","authors":"P. Wohlmacher","doi":"10.1145/357744.357892","DOIUrl":"https://doi.org/10.1145/357744.357892","url":null,"abstract":"Digital certificates form a basis that allows entities to trust each other. Due to different constraints, a certificate is only valid within a specific period of time. Coming from several threats, there are important reasons why its validity must be terminated sooner than assigned and thus, the certificate needs to be revoked. This paper provides a classification of revocation methods and gives an overview of the main methods like CRL, CRS, CRT, and OCSP. If and in which way a revocation method is suited must be analyzed in accordance to their purpose.","PeriodicalId":234597,"journal":{"name":"MULTIMEDIA '00","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129643797","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
MULTIMEDIA '00Pub Date : 2000-11-04DOI: 10.1145/357744.357750
J. Gallagher
{"title":"The merging of the arts with technology","authors":"J. Gallagher","doi":"10.1145/357744.357750","DOIUrl":"https://doi.org/10.1145/357744.357750","url":null,"abstract":"This paper supports the author's opinion that the merging of the arts with technology has been a natural process. She argues her point by citing a brief history of multi-media installations. Additionally, she discusses her recent work as an artist.","PeriodicalId":234597,"journal":{"name":"MULTIMEDIA '00","volume":"6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126309323","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
MULTIMEDIA '00Pub Date : 2000-11-04DOI: 10.1145/357744.357754
C. Strohecker, A. H. Slaughter, M. A. Horvath, Noah J. Appleton
{"title":"Zyklodeon: a software construction kit modeling cyclic timing patterns","authors":"C. Strohecker, A. H. Slaughter, M. A. Horvath, Noah J. Appleton","doi":"10.1145/357744.357754","DOIUrl":"https://doi.org/10.1145/357744.357754","url":null,"abstract":"Zyklodeon is a software construction kit in which Players create colorful Dancers and set them into motion. Dancers' appearances are inspired by Picasso's post-Cubist works, and movements are inspired by Martha Graham's emphasis on the torso as life center. The figures breathe visibly and move gracefully as they encounter other Dancers. Zyklodeon Players construct the figures and set parameters for their movements, thus experimenting with time cycles and notions of emergence in dynamic systems. Our design challenge is to make such complex relationships accessible through coordination of image and sound. We illustrate Dancers' movements as well as measures for breathing and dancing cycles. Graphical communication of key moments in the cycles is augmented by triggering of sounds that combine to form lively music.","PeriodicalId":234597,"journal":{"name":"MULTIMEDIA '00","volume":"85 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116592013","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
MULTIMEDIA '00Pub Date : 2000-11-04DOI: 10.1145/357744.357945
Y. Ariki
{"title":"Organization and retrieval of continuous media","authors":"Y. Ariki","doi":"10.1145/357744.357945","DOIUrl":"https://doi.org/10.1145/357744.357945","url":null,"abstract":"Because of the media digitization, a large amount of information such as speech, audio and video data is produced everyday. In order to retrieve data quickly and precisely from these databases, multimedia technologies for organizing and retrieving of speech, audio and video data are strongly required. In this paper, we overview the multimedia technologies such as organization and retrieval of speech, audio and video data, speaker indexing, audio summarization and cross media retrieval existing today. The main purpose of the organization is to produce tables of contents and indices from audio and video data automatically. In order to make these technologies feasible, first, processing units such as words on audio data and shots on video data are extracted. On a second step, they are meaningfully integrated into topics. Furthermore, the units extracted from different types of media are integrated for higher functions.","PeriodicalId":234597,"journal":{"name":"MULTIMEDIA '00","volume":"42 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125206262","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
MULTIMEDIA '00Pub Date : 2000-11-04DOI: 10.1145/357744.357880
J. Charlesworth, Philip N. Garner
{"title":"Spoken content metadata and MPEG-7","authors":"J. Charlesworth, Philip N. Garner","doi":"10.1145/357744.357880","DOIUrl":"https://doi.org/10.1145/357744.357880","url":null,"abstract":"The words spoken in an audio stream form an obvious descriptor essential to most audio-visual metadata standards. When derived using automatic speech recognition systems, the spoken content fits into neither low-level (representative) nor high-level (semantic) metadata categories. This results in difficulties in creating a representation that can support both interoperability between different extraction and application utilities while retaining robustness to the limitations of the extraction process. In this paper, we discuss the issues encountered in the design of the MPEG-7 spoken content descriptor and their applicability to other metadata standards.","PeriodicalId":234597,"journal":{"name":"MULTIMEDIA '00","volume":"48 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114055980","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}