Proceedings of the 20th ACM international conference on Multimedia最新文献

筛选
英文 中文
Mining in-class social networks for large-scale pedagogical analysis 为大规模教学分析挖掘课堂社会网络
Proceedings of the 20th ACM international conference on Multimedia Pub Date : 2012-10-29 DOI: 10.1145/2393347.2393436
Xiao-Yong Wei, Zhen-Qun Yang
{"title":"Mining in-class social networks for large-scale pedagogical analysis","authors":"Xiao-Yong Wei, Zhen-Qun Yang","doi":"10.1145/2393347.2393436","DOIUrl":"https://doi.org/10.1145/2393347.2393436","url":null,"abstract":"Modeling the in-class student social networks is a highly desired goal in educational literature. However, due to the difficulty to collect social data, most of the conventional studies can only be conducted in a qualitative way on a small-scale of dataset obtained through questionnaires or interviews. We propose to solve the problems of data collection, social network construction and analysis with multimedia technology, in the way that we can automatically recognize the positions and identities of the students in classroom and construct the in-class social networks accordingly. With the social networks and the statistics on a large-scale dataset, we have demonstrated that the pedagogical analysis for investigating the co-learning patterns among the students can be conducted in a quantitative way, which provides the statistical clues about why prior studies reach conflicting conclusions on the relation between the students' positions in social networks and their academic performances. The experimental results have validated the effectiveness of the proposed approaches in both technical and pedagogical senses.","PeriodicalId":212654,"journal":{"name":"Proceedings of the 20th ACM international conference on Multimedia","volume":"70 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-10-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114170134","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
Name that room: room identification using acoustic features in a recording 命名那个房间:使用录音中的声学特征来识别房间
Proceedings of the 20th ACM international conference on Multimedia Pub Date : 2012-10-29 DOI: 10.1145/2393347.2396326
Nils Peters, Howard Lei, G. Friedland
{"title":"Name that room: room identification using acoustic features in a recording","authors":"Nils Peters, Howard Lei, G. Friedland","doi":"10.1145/2393347.2396326","DOIUrl":"https://doi.org/10.1145/2393347.2396326","url":null,"abstract":"This paper presents a system for identifying the room in an audio or video recording through the analysis of acoustical properties. The room identification system was tested using a corpus of 13440 reverberant audio samples. With no common content between the training and testing data, an accuracy of 61% for musical signals and 85% for speech signals was achieved. This approach could be applied in a variety of scenarios where knowledge about the acoustical environment is desired, such as location estimation, music recommendation, or emergency response systems.","PeriodicalId":212654,"journal":{"name":"Proceedings of the 20th ACM international conference on Multimedia","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-10-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114788239","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 30
SIGMM Technical Achievement Award SIGMM技术成就奖
Proceedings of the 20th ACM international conference on Multimedia Pub Date : 2012-10-29 DOI: 10.1145/2393347.2441736
R. Lienhart
{"title":"SIGMM Technical Achievement Award","authors":"R. Lienhart","doi":"10.1145/2393347.2441736","DOIUrl":"https://doi.org/10.1145/2393347.2441736","url":null,"abstract":"","PeriodicalId":212654,"journal":{"name":"Proceedings of the 20th ACM international conference on Multimedia","volume":"32 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-10-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114503085","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
PDSS: patch-descriptor-similarity space for effective face verification PDSS:用于有效人脸验证的patch-descriptor-similarity space
Proceedings of the 20th ACM international conference on Multimedia Pub Date : 2012-10-29 DOI: 10.1145/2393347.2396357
Xiaohua Zhai, Yuxin Peng, Jianguo Xiao
{"title":"PDSS: patch-descriptor-similarity space for effective face verification","authors":"Xiaohua Zhai, Yuxin Peng, Jianguo Xiao","doi":"10.1145/2393347.2396357","DOIUrl":"https://doi.org/10.1145/2393347.2396357","url":null,"abstract":"In this paper, we propose the Patch-Descriptor-Similarity Space (PDSS) for unconstrained face verification, which is challenging due to image variations in pose, lighting, facial expression, and occlusion. Our proposed PDSS considers jointly patch, descriptor and similarity measure, which are ignored by the existing work. PDSS is extremely effective for face verification because each axis of PDSS will boost each other and could maximize the effect of every axis. Each point in PDSS reflects a distinct partial-matching between two facial images, which could be robust to variations in the facial images. Moreover, by selecting the discriminating point subset from PDSS, we could describe accurately the characteristic similarities and differences between two facial images, and further decide whether they represent the same person. In PDSS, each axis can describe effectively the distinct features of the faces: each patch (the first axis) reflects a distinct trait of a face; the descriptor (the second axis) is used to describe such face trait; and the similarity between two features can be measured by a certain kind of similarity measure (the third axis). The experiment adopts the extensively-used Labeled Face in the Wild (LFW) unconstrained face recognition dataset (13K faces), and our proposed PDSS approach achieves the best result, compared with the state-of-the-art methods.","PeriodicalId":212654,"journal":{"name":"Proceedings of the 20th ACM international conference on Multimedia","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-10-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115143964","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Leveraging high-level and low-level features for multimedia event detection 利用高级和低级特性进行多媒体事件检测
Proceedings of the 20th ACM international conference on Multimedia Pub Date : 2012-10-29 DOI: 10.1145/2393347.2393412
Lu Jiang, Alexander Hauptmann, Guang Xiang
{"title":"Leveraging high-level and low-level features for multimedia event detection","authors":"Lu Jiang, Alexander Hauptmann, Guang Xiang","doi":"10.1145/2393347.2393412","DOIUrl":"https://doi.org/10.1145/2393347.2393412","url":null,"abstract":"This paper addresses the challenge of Multimedia Event Detection by proposing a novel method for high-level and low-level features fusion based on collective classification. Generally, the method consists of three steps: training a classifier from low-level features; encoding high-level features into graphs; and diffusing the scores on the established graph to obtain the final prediction. The final prediction is derived from multiple graphs each of which corresponds to a high-level feature. The paper investigates two graph construction methods using logarithmic and exponential loss functions, respectively and two collective classification algorithms, i.e. Gibbs sampling and Markov random walk. The theoretical analysis demonstrates that the proposed method converges and is computationally scalable and the empirical analysis on TRECVID 2011 Multimedia Event Detection dataset validates its outstanding performance compared to state-of-the-art methods, with an added benefit of interpretability.","PeriodicalId":212654,"journal":{"name":"Proceedings of the 20th ACM international conference on Multimedia","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-10-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115262742","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 79
Searching for diversified landmarks by photo 通过照片搜索各种地标
Proceedings of the 20th ACM international conference on Multimedia Pub Date : 2012-10-29 DOI: 10.1145/2393347.2396472
Junfeng Ye, Jia Chen, Zejia Chen, Yihe Zhu, Shenghua Bao, Zhong Su, Yong Yu
{"title":"Searching for diversified landmarks by photo","authors":"Junfeng Ye, Jia Chen, Zejia Chen, Yihe Zhu, Shenghua Bao, Zhong Su, Yong Yu","doi":"10.1145/2393347.2396472","DOIUrl":"https://doi.org/10.1145/2393347.2396472","url":null,"abstract":"This demo focuses on the problem of searching for diversified landmarks with photos as input. More particularly, we propose a system called DLMSearch that allows a user to upload a photo as a query and searches for a diverse set of relevant landmarks in real time. It also presents a photo summary for each retrieved landmark, considering both visual representativeness and diversity. Our online demo is available at http://lm.apexlab.org/landmark/demo.","PeriodicalId":212654,"journal":{"name":"Proceedings of the 20th ACM international conference on Multimedia","volume":"68 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-10-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123094685","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Interactive data-driven search and discovery of temporal behavior patterns from media streams 交互式数据驱动搜索和发现媒体流中的临时行为模式
Proceedings of the 20th ACM international conference on Multimedia Pub Date : 2012-10-29 DOI: 10.1145/2393347.2396512
Chreston A. Miller
{"title":"Interactive data-driven search and discovery of temporal behavior patterns from media streams","authors":"Chreston A. Miller","doi":"10.1145/2393347.2396512","DOIUrl":"https://doi.org/10.1145/2393347.2396512","url":null,"abstract":"The presented thesis work addresses how social scientists may derive patterns of human behavior captured in media streams. Currently, media streams are being segmented into sequences of events describing the actions captured in the streams, such as the interactions among humans. This segmentation creates a challenging data space to search characterized by non-numerical, temporal, descriptive data, e.g., Person A walks up to Person B at time T. We present an approach that allows one to interactively search and discover temporal behavior patterns within such a data space.","PeriodicalId":212654,"journal":{"name":"Proceedings of the 20th ACM international conference on Multimedia","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-10-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124765254","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Spacetime freeview generation using image-based rendering, relighting, and augmented telepresence 使用基于图像的渲染、重照明和增强远程呈现的时空自由视图生成
Proceedings of the 20th ACM international conference on Multimedia Pub Date : 2012-10-29 DOI: 10.1145/2393347.2396513
Fumio Okura
{"title":"Spacetime freeview generation using image-based rendering, relighting, and augmented telepresence","authors":"Fumio Okura","doi":"10.1145/2393347.2396513","DOIUrl":"https://doi.org/10.1145/2393347.2396513","url":null,"abstract":"This paper proposes an freeview generation technique providing the users to change their viewpoints beyond time and space. The study consists of three technical elements: image-based rendering, relighting, and augmented telepresence. Before now, we have developed two systems relating this study: an augmented telepresence system and a full spherical HDR aerial imaging system.","PeriodicalId":212654,"journal":{"name":"Proceedings of the 20th ACM international conference on Multimedia","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-10-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121777503","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Situation recognition: an evolving problem for heterogeneous dynamic big multimedia data 情境识别:异构动态大多媒体数据的演进问题
Proceedings of the 20th ACM international conference on Multimedia Pub Date : 2012-10-29 DOI: 10.1145/2393347.2396421
V. Singh, Mingyan Gao, R. Jain
{"title":"Situation recognition: an evolving problem for heterogeneous dynamic big multimedia data","authors":"V. Singh, Mingyan Gao, R. Jain","doi":"10.1145/2393347.2396421","DOIUrl":"https://doi.org/10.1145/2393347.2396421","url":null,"abstract":"With the growth in social media, internet of things, and planetary-scale sensing there is an unprecedented need to assimilate spatio-temporally distributed multimedia streams into actionable information. Consequently the concepts like objects, scenes, and events, need to be extended to recognize situations (e.g. epidemics, traffic jams, seasons, flash mobs). This paper motivates and computationally grounds the problem of situation recognition. It describes a systematic approach for combining multimodal real-time big data into actionable situations. Specifically it presents a generic approach for modeling and recognizing situations. A set of generic building blocks and guidelines help the domain experts model their situations of interest. The created models can be tested, refined, and deployed into practice using a developed system (EventShop). Results of applying this approach to create multiple situation-aware applications by combining heterogeneous streams (e.g. Twitter, Google Insights, Satellite imagery, Census) are presented.","PeriodicalId":212654,"journal":{"name":"Proceedings of the 20th ACM international conference on Multimedia","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-10-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121909242","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 38
Hi, magic closet, tell me what to wear! 嗨,神奇衣橱,告诉我该穿什么!
Proceedings of the 20th ACM international conference on Multimedia Pub Date : 2012-10-29 DOI: 10.1145/2393347.2393433
Si Liu, Jiashi Feng, Zheng Song, Tianzhu Zhang, Hanqing Lu, Changsheng Xu, Shuicheng Yan
{"title":"Hi, magic closet, tell me what to wear!","authors":"Si Liu, Jiashi Feng, Zheng Song, Tianzhu Zhang, Hanqing Lu, Changsheng Xu, Shuicheng Yan","doi":"10.1145/2393347.2393433","DOIUrl":"https://doi.org/10.1145/2393347.2393433","url":null,"abstract":"In this paper, we aim at a practical system, magic closet, for automatic occasion-oriented clothing recommendation. Given a user-input occasion, e.g., wedding, shopping or dating, magic closet intelligently suggests the most suitable clothing from the user's own clothing photo album, or automatically pairs the user-specified reference clothing (upper-body or lower-body) with the most suitable one from online shops. Two key criteria are explicitly considered for the magic closet system. One criterion is to wear properly, e.g., compared to suit pants, it is more decent to wear a cocktail dress for a banquet occasion. The other criterion is to wear aesthetically, e.g., a red T-shirt matches better white pants than green pants. To narrow the semantic gap between the low-level features of clothing and the high-level occasion categories, we adopt middle-level clothing attributes (e.g., clothing category, color, pattern) as a bridge. More specifically, the clothing attributes are treated as latent variables in our proposed latent Support Vector Machine (SVM) based recommendation model. The wearing properly criterion is described in the model through a feature-occasion potential and an attribute-occasion potential, while the wearing aesthetically criterion is expressed by an attribute-attribute potential. To learn a generalize-well model and comprehensively evaluate it, we collect a large clothing What-to-Wear (WoW) dataset, and thoroughly annotate the whole dataset with 7 multi-value clothing attributes and 10 occasion categories via Amazon Mechanic Turk. Extensive experiments on the WoW dataset demonstrate the effectiveness of the magic closet system for both occasion-oriented clothing recommendation and pairing.","PeriodicalId":212654,"journal":{"name":"Proceedings of the 20th ACM international conference on Multimedia","volume":"14 1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-10-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121474392","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 271
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信