Sung-Cheal Byun, Il Lo Lee, Taehwan Shin, Byung-Ha Ahn
{"title":"A public-key based watermarking for color image authentication","authors":"Sung-Cheal Byun, Il Lo Lee, Taehwan Shin, Byung-Ha Ahn","doi":"10.1109/ICME.2002.1035851","DOIUrl":"https://doi.org/10.1109/ICME.2002.1035851","url":null,"abstract":"We present a watermarking scheme for color image authentication based on cryptography. The scheme shows that if even a single bit has been changed, the watermarked image is regarded as unauthentic. Moreover, the scheme indicates the specific locations where changes have been made. To increase the security, the scheme uses two keys for the permutation of the input of the hash function and for the encryption (decryption) of watermark bits during the watermark insertion and extraction procedures, respectively. Results show that our watermark is correctly extracted if there is no change of watermarked image. However, any change of a watermarked image results in no watermark. The pixel changes of the watermarked image are reflected in the extracted watermark, which indicates the attacked area.","PeriodicalId":90694,"journal":{"name":"Proceedings. IEEE International Conference on Multimedia and Expo","volume":"112 1","pages":"593-596 vol.1"},"PeriodicalIF":0.0,"publicationDate":"2002-11-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"87799668","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Fast music retrieval using polyphonic binary feature vectors","authors":"H. Nagano, K. Kashino, H. Murase","doi":"10.1109/ICME.2002.1035728","DOIUrl":"https://doi.org/10.1109/ICME.2002.1035728","url":null,"abstract":"We propose a method for retrieving similar music from a polyphonic-music audio database using a polyphonic audio signal as a query. In this task, we must consider similarities among polyphonic signals of the music, and achieve quick retrieval. Therefore, we first introduce the polyphonic binary feature vector to represent the presence of multiple notes. This feature is suitable for the search based on the similarities among polyphonic audio signals. Then, we propose a new search method, which is quicker than the exhaustive use of DP matching. The search is accelerated using a \"similarity matrix\" to limit the search space. Experiments using a test database containing 216 music pieces show that the search accuracy of the proposed feature is 89%, which is approximately 26% higher than that of the conventional spectrum feature. It is also shown that the new search method retrieves similar music without significant accuracy degradation as well as the exhaustive search does and the computational complexity of the new search method is about 1/4 that of exhaustive search.","PeriodicalId":90694,"journal":{"name":"Proceedings. IEEE International Conference on Multimedia and Expo","volume":"385 1","pages":"101-104 vol.1"},"PeriodicalIF":0.0,"publicationDate":"2002-11-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"86819258","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Popularity-adaptive index scheme for fast music retrieval","authors":"Dongmoon Park, Eenjun Hwang","doi":"10.1109/ICME.2002.1035733","DOIUrl":"https://doi.org/10.1109/ICME.2002.1035733","url":null,"abstract":"Proliferation of audio databases on the WWW (World Wide Web) necessitates an audio retrieval system to find certain audio content within the audio corpus. Many papers have presented concepts, methodologies and systems to offer users ways to retrieve melody from collections of music contents. In this paper, we present a new index scheme to retrieve music based on the accumulated data of previous user queries to music. We first describe the current status of existing music information retrieval systems, and then present the design and implementation of our prototype system. We report the results obtained from an empirical evaluation of our approach.","PeriodicalId":90694,"journal":{"name":"Proceedings. IEEE International Conference on Multimedia and Expo","volume":"1 1","pages":"121-124 vol.1"},"PeriodicalIF":0.0,"publicationDate":"2002-11-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"86335468","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Category-based search using metadatabase in image retrieval","authors":"Yimin Wu, A. Zhang","doi":"10.1109/ICME.2002.1035752","DOIUrl":"https://doi.org/10.1109/ICME.2002.1035752","url":null,"abstract":"We present a self-adjustable metadatabase aimed at improving the performance of the relevance feedback module extensively used in content-based image retrieval systems. Our metadatabase provides a mechanism for accumulating the optimized relevance feedback records (which are called metadata records) obtained from previous queries. Each metadata record in the metadatabase includes optimal query, feature weights, and identifiers of relevant and/or irrelevant images, and can be effectively used to guide future queries. With the metadatabase, the relevance feedback module admits a noticeable improvement on its performance for category-based search, especially when the relevant images form multiple classes in the feature space. Experiments on a Corel image set (with 31,438 images) show that our method has at least a 15% improvement on average precision and recall over relevance-feedback-only approaches.","PeriodicalId":90694,"journal":{"name":"Proceedings. IEEE International Conference on Multimedia and Expo","volume":"20 1","pages":"197-200 vol.1"},"PeriodicalIF":0.0,"publicationDate":"2002-11-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"88690521","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Multi-level memory prefetching for media and stream processing","authors":"Jason E. Fritts","doi":"10.1109/ICME.2002.1035522","DOIUrl":"https://doi.org/10.1109/ICME.2002.1035522","url":null,"abstract":"This paper presents a multi-level memory prefetch hierarchy for media and stream processing applications. Two major bottlenecks in the performance of multimedia and network applications are long memory latencies and limited off-chip processor bandwidth. Aggressive prefetching can be used to mitigate the memory latency problem, but overly aggressive prefetching may overload the limited external processor bandwidth. To accommodate both problems, we propose multilevel memory prefetching. The multi-level organization enables conservative prefetching on-chip and more aggressive prefetching off-chip. The combination provides aggressive prefetching while minimally impacting off-chip bandwidth, enabling more efficient memory performance for media and stream processing. This paper presents preliminary results for multi-level memory prefetching, which show that combining prefetching at the L1 and DRAM memory levels provides the most effective prefetching with minimal extra bandwidth.","PeriodicalId":90694,"journal":{"name":"Proceedings. IEEE International Conference on Multimedia and Expo","volume":"1 1","pages":"101-104 vol.2"},"PeriodicalIF":0.0,"publicationDate":"2002-11-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"88880766","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Error resilience of video transmission by rate-distortion optimization and adaptive packetization","authors":"Yuxin Liu, P. Salama, E. Delp","doi":"10.1109/ICME.2002.1035697","DOIUrl":"https://doi.org/10.1109/ICME.2002.1035697","url":null,"abstract":"We propose new schemes to introduce error resilience into the compressed video bitstreams for transmission over packet networks. First, we develop an adaptive packetization scheme that prohibits any dependency across packets, for error resilience purposes, while exploiting the dependency within each packet to improve the source coding performance. Secondly, we address a two-layer rate-distortion optimization scheme to serve our packetization method. We also use forward error correction (FEC) coding across packets to provide further error protection. Finally, we present a simplified version of our schemes to make it fully compliant with the current ITU video coding standard - H.263+.","PeriodicalId":90694,"journal":{"name":"Proceedings. IEEE International Conference on Multimedia and Expo","volume":"31 4 1","pages":"613-616 vol.2"},"PeriodicalIF":0.0,"publicationDate":"2002-11-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"88977770","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"New kernels for analyzing multimodal data in multimedia using kernel machines","authors":"H. Aradhye, C. Dorai","doi":"10.1109/ICME.2002.1035368","DOIUrl":"https://doi.org/10.1109/ICME.2002.1035368","url":null,"abstract":"Research in automated analysis of digital media content has led to a large collection of low-level feature extractors, such as face detectors, videotext extractors, speech and speaker identifiers, people/vehicle trackers, and event locators. These media metadata are often symbolic rather than continuous-valued, and pose significant difficulty to subsequent tasks such as classification and dimensionality reduction which traditionally deal with continuous-valued data. This paper proposes a novel mechanism that extends tasks traditionally limited to continuous-valued feature spaces, such as (a) dimensionality reduction, (b) de-noising, and (c) clustering, to domains with symbolic features. To this end, we introduce new kernels based on well-known distance metrics, and prove Mercer validity of these kernels for analyzing symbolic feature spaces. We demonstrate their usefulness within the context of kernel-space methods such as Kernel PCA and SVM, in classifying machine learning datasets from the UCI repository and in temporal clustering and tracking of videotext in multimedia. We show that the generalized kernels help capture information from symbolic feature spaces, visualize symbolic data, and aid tasks such as classification and clustering, and therefore are useful in multimodal analysis of multimedia.","PeriodicalId":90694,"journal":{"name":"Proceedings. IEEE International Conference on Multimedia and Expo","volume":"209 1","pages":"37-40 vol.2"},"PeriodicalIF":0.0,"publicationDate":"2002-11-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"89018099","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
J. Delaigle, C. Vleeschouwer, B. Macq, R. Lagendijk
{"title":"Human visual system features enabling watermarking","authors":"J. Delaigle, C. Vleeschouwer, B. Macq, R. Lagendijk","doi":"10.1109/ICME.2002.1035653","DOIUrl":"https://doi.org/10.1109/ICME.2002.1035653","url":null,"abstract":"Digital watermarking consists of hiding subliminal information into digital media content, also called host data. It can be the basis of many applications, including security and media asset management. In this paper, we focus on the imperceptibility requirement for image watermarking. We present the main features of the human visual system (HVS) to be translated into watermarking technology. This paper highlights the need for dedicated inputs from the human vision community. The human visual system (HVS) is very complex and able to deal with a huge amount of information. Roughly speaking, it is composed of a receiver with a pre-processing stage, the eye and the retina, a transmission channel, the optic nerve, and a processing engine, the visual cortex. Mainly because of our lack of knowledge about brain behavior, i.e. about the way a stimulus is processed through its huge neural network, the large effort to understand and model the HVS behavior has partly remained fruitless. The aim of this paper is not to provide a thorough description of the HVS. For complete HVS models and more specific details, the reader is referred to existing literature. Here, we only try to understand, in a synthetic way and from an engineering perspective, the HVS features on which the designer of a watermarking algorithm can rely, i.e. its sensitivity and masking capabilities.","PeriodicalId":90694,"journal":{"name":"Proceedings. IEEE International Conference on Multimedia and Expo","volume":"72 1","pages":"489-492 vol.2"},"PeriodicalIF":0.0,"publicationDate":"2002-11-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"80026484","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
A. Gil-Solla, J. Pazos-Arias, C. López-García, J. López-Ardao, R. F. R. Rubio, M. Cabrer, R. Redondo
{"title":"Surfing the Web on TV: the MHP approach","authors":"A. Gil-Solla, J. Pazos-Arias, C. López-García, J. López-Ardao, R. F. R. Rubio, M. Cabrer, R. Redondo","doi":"10.1109/ICME.2002.1035579","DOIUrl":"https://doi.org/10.1109/ICME.2002.1035579","url":null,"abstract":"In last decade, we have seen the first steps to the end of passive television. Thanks to continuous advances in hardware and software, digital TV technology is now mature enough to enhance traditional TV sets (limited to content reproduction) with computing capability to run multimedia software integrating richer formats. A significant example of this progress is Internet access through television, which is becoming a reality in the last generation of digital set-top boxes. However, both media (computers and TVs) are different enough to require noteworthy modifications in their respective computing models. The MHP (multimedia home platform) standard is the first one that tries to define regulations in that sense. We comment on some technical aspects of the MHP solution.","PeriodicalId":90694,"journal":{"name":"Proceedings. IEEE International Conference on Multimedia and Expo","volume":"60 1","pages":"285-288 vol.2"},"PeriodicalIF":0.0,"publicationDate":"2002-11-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"76680629","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Enhancement of colored engravings","authors":"V. Misic, R. Buckley, K. Parker","doi":"10.1109/ICME.2002.1035885","DOIUrl":"https://doi.org/10.1109/ICME.2002.1035885","url":null,"abstract":"Digitization of printed illustrations for archive purposes is not always a simple procedure. For example, colored engravings (a common book-illustration type of the XVIII and XIX century) consist of regions of both very fine lines and almost uniformly colored regions. In order to keep the compression rate high, compression algorithms tend to destroy fine lines (the high-frequency content) that are the essence of this type of image. However, the images compressed using the modified mixed raster content (mMRC) compression scheme retain the lines and show exceptional fidelity to their prototypes at low bit-rates when compared to alternate compression methods such as JPEG and JPEG2000. This paper presents the simple algorithm for data-filling of the MRC background layer, that improves the compression results, and - at the same time - enhances the reconstructed image. This is demonstrated on the case study of the colored engravings from the William Blake Archive.","PeriodicalId":90694,"journal":{"name":"Proceedings. IEEE International Conference on Multimedia and Expo","volume":"17 1","pages":"729-732 vol.1"},"PeriodicalIF":0.0,"publicationDate":"2002-11-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"81465555","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}