2003 International Conference on Multimedia and Expo. ICME '03. Proceedings (Cat. No.03TH8698)最新文献

筛选
英文 中文
Histogram-based image retrieval using Gauss mixture vector quantization 基于直方图的高斯混合矢量量化图像检索
Sangoh Jeong, C. Won, R. Gray
{"title":"Histogram-based image retrieval using Gauss mixture vector quantization","authors":"Sangoh Jeong, C. Won, R. Gray","doi":"10.1109/ICME.2003.1221637","DOIUrl":"https://doi.org/10.1109/ICME.2003.1221637","url":null,"abstract":"Histogram-based image retrieval requires some form of quantization since the raw color images result in large dimensionality in the histogram representation. Simple uniform quantization disregards the spatial information among pixels in making histograms. Since traditional vector quantization (VQ) with squared-error distortion employs only the first moment, it neglects the relationship among vectors. We propose Gauss mixture vector quantization (GMVQ) as the quantization method for a histogram-based image retrieval to capture the spatial information in the image via the Gaussian covariance structure. Two common histogram distance measures are used to evaluate the similarity of histograms resulting from GMVQ. Our result shows that GMVQ with a quadratic discriminant analysis (QDA) distortion outperforms the two typical quantization methods in the histogram- based image retrieval.","PeriodicalId":118560,"journal":{"name":"2003 International Conference on Multimedia and Expo. ICME '03. Proceedings (Cat. No.03TH8698)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-07-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129893538","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
On the rate constraint of transmitting multiple priority classes with QoS 基于QoS的多优先级传输速率约束研究
W. Kumwilaisak, Qian Zhang, Wenwu Zhu, C.-C. Jay Kuo, Ya-Qin Zhang
{"title":"On the rate constraint of transmitting multiple priority classes with QoS","authors":"W. Kumwilaisak, Qian Zhang, Wenwu Zhu, C.-C. Jay Kuo, Ya-Qin Zhang","doi":"10.1109/ICME.2003.1220952","DOIUrl":"https://doi.org/10.1109/ICME.2003.1220952","url":null,"abstract":"The rate constraint of transmitting multiple priority classes over a time-varying service-rate channel is studied in this work. This constraint specifies the maximum data rate that can be transmitted reliably with QoS (quality of service) guarantee. In our framework, the time-varying service channel is modeled by an N-state discrete Markov process, where each Markov state is associated with a channel service rate, and the absolute priority scheduling is used to transport packets of different classes. The transmission rate constraint is derived based on effective bandwidth and capacity. To be more specific, given channel parameters and the maximum buffer size for each priority class, statistical QoS guarantees in terms of packet loss probabilities can be determined and translated to the transmission rate constraint. The derived result is verified by simulation in a time-varying wireless environment.","PeriodicalId":118560,"journal":{"name":"2003 International Conference on Multimedia and Expo. ICME '03. Proceedings (Cat. No.03TH8698)","volume":"65 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-07-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130347302","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Combining classifiers for face recognition 结合分类器进行人脸识别
Xiaoguang Lu, Yunhong Wang, Anil K. Jain
{"title":"Combining classifiers for face recognition","authors":"Xiaoguang Lu, Yunhong Wang, Anil K. Jain","doi":"10.1109/ICME.2003.1221236","DOIUrl":"https://doi.org/10.1109/ICME.2003.1221236","url":null,"abstract":"Current two-dimensional face recognition approaches can obtain a good performance only under constrained environments. However, in the real applications, face appearance changes significantly due to different illumination, pose, and expression. Face recognizers based on different representations of the input face images have different sensitivity to these variations. Therefore, a combination of different face classifiers which can integrate the complementary information should lead to improved classification accuracy. We use the sum rule and RBF-based integration strategies to combine three commonly used face classifiers based on PCA, ICA and LDA representations. Experiments conducted on a face database containing 206 subjects (2,060 face images) show that the proposed classifier combination approaches outperform individual classifiers.","PeriodicalId":118560,"journal":{"name":"2003 International Conference on Multimedia and Expo. ICME '03. Proceedings (Cat. No.03TH8698)","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-07-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123861916","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 173
Concept learning and transplantation for dynamic image databases 动态图像数据库的概念学习与移植
Anlei Dong, B. Bhanu
{"title":"Concept learning and transplantation for dynamic image databases","authors":"Anlei Dong, B. Bhanu","doi":"10.1109/ICME.2003.1221030","DOIUrl":"https://doi.org/10.1109/ICME.2003.1221030","url":null,"abstract":"The task of a content-based image retrieval (CBIR) system is to cater to users who expect to get relevant images with high precision and efficiency in response to query images. This paper presents a concept learning approach that integrates a mixture model of the data, relevance feedback and long-term continuous learning. The concepts are incrementally refined with increased retrieval experiences. The concept knowledge can be immediately transplanted to deal with the dynamic database situations such as insertion of new images, removal of existing images and query images, which are outside the database. Experimental results on Corel database show the efficacy of our approach.","PeriodicalId":118560,"journal":{"name":"2003 International Conference on Multimedia and Expo. ICME '03. Proceedings (Cat. No.03TH8698)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-07-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123961256","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Expressive lighting for interactive entertainment 用于互动娱乐的表现力照明
M. S. El-Nasr, I. Horswill
{"title":"Expressive lighting for interactive entertainment","authors":"M. S. El-Nasr, I. Horswill","doi":"10.1109/ICME.2003.1220945","DOIUrl":"https://doi.org/10.1109/ICME.2003.1220945","url":null,"abstract":"Lighting design is an essential element of visual storytelling. In cinematic theory, lighting design plays an important role in illumination, directing viewer's gaze to important locations, and portraying visual tension. However, lighting design is a complex process, and is especially problematic for interactive entertainment systems, since important design parameters such as spatial configuration, dramatic focus, and dramatic intensity, cannot be determined in advance. Indeed, they often change dramatically during interaction. Additionally, manually adjusting colors, positions, and angles of each light in a scene is a time consuming and tedious process. In this paper, we describe a system developed based on cinematic and theatrical lighting design theory to automatically adjust positions, colors, and angles of each light in real-time to accommodate the continually evolving dramatic situation, while maintaining the desired style, and ensuring visual continuity.","PeriodicalId":118560,"journal":{"name":"2003 International Conference on Multimedia and Expo. ICME '03. Proceedings (Cat. No.03TH8698)","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-07-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121187244","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
No-reference JPEG-2000 image quality metric 无参考JPEG-2000图像质量度量
E. Ong, Weisi Lin, Zhongkang Lu, S. Yao, Xiaokang Yang, Lijun Jiang
{"title":"No-reference JPEG-2000 image quality metric","authors":"E. Ong, Weisi Lin, Zhongkang Lu, S. Yao, Xiaokang Yang, Lijun Jiang","doi":"10.1109/ICME.2003.1220975","DOIUrl":"https://doi.org/10.1109/ICME.2003.1220975","url":null,"abstract":"In this paper, a method for measuring the perceptual image quality of JPEG-2000 coded images has been proposed. The image quality is characterized by the average edge-spread in the image, or more specifically the average extent of the slope's spread of an edge in the opposing gradients' directions. The proposed method is, in effect, a way of measuring the amount of blurring in the image. The effectiveness of such method is validated using subjective tests and the experimental results show that the proposed method can provide results that correlate relatively well with human subjective ratings.","PeriodicalId":118560,"journal":{"name":"2003 International Conference on Multimedia and Expo. ICME '03. Proceedings (Cat. No.03TH8698)","volume":"93 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-07-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124527126","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 35
A secure and robust approach to scalable video authentication 一种安全可靠的可扩展视频认证方法
Qibin Sun, D. He, Zhishou Zhang, Q. Tian
{"title":"A secure and robust approach to scalable video authentication","authors":"Qibin Sun, D. He, Zhishou Zhang, Q. Tian","doi":"10.1109/ICME.2003.1221590","DOIUrl":"https://doi.org/10.1109/ICME.2003.1221590","url":null,"abstract":"In this paper, we present a secure and robust content authentication scheme for scalable video streaming. In our authentication scheme we consider three common video transcoding methods as acceptable content manipulations, when the streaming bit-rate needs to be reduced, namely frame resizing, frame dropping and multi-cycle coding. By employing error correction coding (ECC) in different ways, the proposed scheme is insensitive to those incidental distortions introduced during the transcoding (i.e., robust) while is still sensitive to other intentional distortions such as frame alterations and insertion (i.e., secure). One key feature in our scheme is that it achieves an end-to-end authentication independent of transcoding infrastructure and obtains a good compromise between system robustness and security.","PeriodicalId":118560,"journal":{"name":"2003 International Conference on Multimedia and Expo. ICME '03. Proceedings (Cat. No.03TH8698)","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-07-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126365146","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 20
A handheld mirror simulation 手持式镜像模拟
A. François, E.-Y. Kang
{"title":"A handheld mirror simulation","authors":"A. François, E.-Y. Kang","doi":"10.1109/ICME.2003.1221724","DOIUrl":"https://doi.org/10.1109/ICME.2003.1221724","url":null,"abstract":"We present the design and construction of a handheld mirror simulation device. The perception of the world reflected through a mirror depends on the viewer's position with respect to the mirror and the 3-D geometry of the world. In order to simulate a real mirror on a computer screen, images of the observed world, consistent with the viewer's position, must be synthesized and displayed in real- time. Our system is build around a LCD screen manipulated by the user, a single camera fixed on the screen, and a tracking device. The continuous input video stream and tracker data is used to synthesize, in real-time, a continuous video stream displayed on the LCD screen. The synthesized video stream is a close approximation of what the user would see on the screen surface if it were a real mirror. Our system provides a generic interface for applications involving rich, first-person interaction, such as the virtual daguerreotype.","PeriodicalId":118560,"journal":{"name":"2003 International Conference on Multimedia and Expo. ICME '03. Proceedings (Cat. No.03TH8698)","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-07-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128068091","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 25
Object-based coding for long-term archive of surveillance video 基于对象的监控视频长期存档编码
A. Vetro, T. Haga, K. Sumi, Huifang Sun
{"title":"Object-based coding for long-term archive of surveillance video","authors":"A. Vetro, T. Haga, K. Sumi, Huifang Sun","doi":"10.1109/ICME.2003.1221642","DOIUrl":"https://doi.org/10.1109/ICME.2003.1221642","url":null,"abstract":"This paper describes video coding and segmentation techniques that can be used to achieve significant increase in storage capacity. Specifically, we examine the possibility to use object- based coding for efficient long-term archiving of surveillance video. We consider surveillance systems with many camera sources in which we are required to store several months of video data for each source, thus storage capacity is a major concern. The paper considers several automatic segmentation algorithms. With each algorithm, we analyze the shape coding overhead and implication on overall storage requirements, as well as the effect each algorithm has on the reconstructed quality of frames. Additionally, this paper reviews techniques to dynamically control the temporal rate of objects in the scene and perform bit allocation. Experimental results show that up to 90% savings in storage can be achieved with the proposed method compared to frame-based video coding techniques. The cost for this savings is that the accuracy of the background is compromised; however, we feel that this is satisfactory for the application under consideration.","PeriodicalId":118560,"journal":{"name":"2003 International Conference on Multimedia and Expo. ICME '03. Proceedings (Cat. No.03TH8698)","volume":"30 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-07-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125658643","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 66
Scene reconstruction using distributed microphone arrays 使用分布式麦克风阵列的场景重建
P. Aarabi, B. Mungamuru
{"title":"Scene reconstruction using distributed microphone arrays","authors":"P. Aarabi, B. Mungamuru","doi":"10.1109/ICME.2003.1221246","DOIUrl":"https://doi.org/10.1109/ICME.2003.1221246","url":null,"abstract":"A method for the joint localization and orientation estimation of a directional sound source using distributed microphones is presented. By modeling the signal attenuation due to the microphone directivity, the source directivity, and the source-microphone distance, a multi-dimensional search over all possible sound source scene reconstruction algorithm is presented in the context of an experiment with 24 microphones and a dynamic speech source. At a signal-to-noise ratio of 20 dB and with a reverberation time of approximately 0.1 s, accurate location estimates (20 cm error) and orientation estimates (less than 10/spl deg/ average error) are obtained.","PeriodicalId":118560,"journal":{"name":"2003 International Conference on Multimedia and Expo. ICME '03. Proceedings (Cat. No.03TH8698)","volume":"45 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-07-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128175513","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信