HuEvent '14Pub Date : 2014-11-07DOI: 10.1145/2660505.2660514
O. Etzion, Fabiana Fournier
{"title":"On the Personalization of Event-Based Systems","authors":"O. Etzion, Fabiana Fournier","doi":"10.1145/2660505.2660514","DOIUrl":"https://doi.org/10.1145/2660505.2660514","url":null,"abstract":"In this paper we describe our position about personalization as a paradigm shift that is going to affect life in many areas. For Internet of Things applications, personalization is a critical success factor, especially in detecting situations in real-time. We discuss the need for personalization, and compare the Internet of Things to the traditional Internet to draw conclusions on the gaps. We also discuss \"The Event Model\" as a direction towards such personalization. Finally we present multi-disciplinary research challenges for enabling personalization for Internet of Things applications.","PeriodicalId":434817,"journal":{"name":"HuEvent '14","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-11-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129810214","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
HuEvent '14Pub Date : 2014-11-07DOI: 10.1145/2660505.2660513
Sejong Yoon, V. Pavlovic
{"title":"Sentiment Flow for Video Interestingness Prediction","authors":"Sejong Yoon, V. Pavlovic","doi":"10.1145/2660505.2660513","DOIUrl":"https://doi.org/10.1145/2660505.2660513","url":null,"abstract":"Computational analysis and prediction of digital media interestingness is a challenging task, largely driven by subjective nature of interestingness. Several attempts were made to construct a reliable measure and obtain a better understanding of interestingness based on various psychological study results. However, most current works focus on interestingness prediction for images. While the video affective analysis has been studied for quite some time, there are few works that explictly try to predict interestingness of videos. In this work, we extend a recent pilot study on the video interestingness prediction by using a mid-level representation of sentiment (emotion) sequence. We evaluate our proposed framework on three datasets including the datasets proposed by the pilot study and show that the result effectively verifies a promising utility of the approach.","PeriodicalId":434817,"journal":{"name":"HuEvent '14","volume":"41 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-11-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133006603","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
HuEvent '14Pub Date : 2014-11-07DOI: 10.1145/2660505.2660506
Ishani Chakraborty, Hui Cheng, O. Javed
{"title":"Entity centric Feature Pooling for Complex Event Detection","authors":"Ishani Chakraborty, Hui Cheng, O. Javed","doi":"10.1145/2660505.2660506","DOIUrl":"https://doi.org/10.1145/2660505.2660506","url":null,"abstract":"In this paper, we propose an entity centric region of interest detection and visual-semantic pooling scheme for complex event detection in YouTube-like videos. Our method is based on the hypothesis that many YouTube-like videos involve people interacting with each other and objects in their vicinity. Based on this hypothesis, we first discover an Area of Interest (AoI) map in image keyframes and then use the AoI map for localized pooling of features. The AoI map is derived from image based saliency cues weighted by the actionable space of the person involved in the event. We extract the actionable space of the person based on human position and gaze based attention allocated per region. Based on the AoI map, we divide the image into disparate regions, pool features separately from each region and finally combine them into a single image signature. To this end, we show that our proposed semantically pooled image signature contains discriminative information that detects visual events favorably as compared to state of the art approaches.","PeriodicalId":434817,"journal":{"name":"HuEvent '14","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-11-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133243167","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
HuEvent '14Pub Date : 2014-11-07DOI: 10.1145/2660505.2660509
Mario Guggenberger, M. Riegler, M. Lux, P. Halvorsen
{"title":"Event Understanding in Endoscopic Surgery Videos","authors":"Mario Guggenberger, M. Riegler, M. Lux, P. Halvorsen","doi":"10.1145/2660505.2660509","DOIUrl":"https://doi.org/10.1145/2660505.2660509","url":null,"abstract":"Event detection and understanding is an important area in computer science and especially multimedia. The term event is very broad, and we want to propose a novel event based view on endoscopic surgeries. Thus, with the novel view on surgery in this paper, we want to provide a better understanding and possible way of segmentation of the whole event surgery but also the included sub-events. To achieve this sophisticated goal, we present an annotation tool in combination with a thinking aloud test with an experienced surgeon.","PeriodicalId":434817,"journal":{"name":"HuEvent '14","volume":"126 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-11-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115505934","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
HuEvent '14Pub Date : 2014-11-07DOI: 10.1145/2660505.2660507
C. Papagiannopoulou, V. Mezaris
{"title":"Concept-based Image Clustering and Summarization of Event-related Image Collections","authors":"C. Papagiannopoulou, V. Mezaris","doi":"10.1145/2660505.2660507","DOIUrl":"https://doi.org/10.1145/2660505.2660507","url":null,"abstract":"In this work we deal with the problem of summarizing image collections that correspond to a single event each. For this, we adopt a clustering-based approach, and we perform a comparative study of different clustering algorithms and image representations. As part of this study, we propose and examine the possibility of using trained concept detectors so as to represent each image with a vector of concept detector responses, which is then used as input to the clustering algorithms. A technique which indicates which concepts are the most informative ones for clustering is also introduced, allowing us to prune the employed concept detectors. Following the clustering, a summary of the collection (thus, also of the event) can be formed by selecting one or more images per cluster, according to different possible criteria. The combination of clustering and concept-based image representation is experimentally shown to result in the formation of clusters and summaries that match well the human expectations.","PeriodicalId":434817,"journal":{"name":"HuEvent '14","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-11-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128099360","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
HuEvent '14Pub Date : 2014-11-07DOI: 10.1145/2660505.2660512
S. Wei, N. C. Tang, Yen-Yu Lin, Ming-Fang Weng, H. Liao
{"title":"Skeleton-augmented Human Action Understanding by Learning with Progressively Refined Data","authors":"S. Wei, N. C. Tang, Yen-Yu Lin, Ming-Fang Weng, H. Liao","doi":"10.1145/2660505.2660512","DOIUrl":"https://doi.org/10.1145/2660505.2660512","url":null,"abstract":"With the aim at accurate action video retrieval, we firstly present an approach that can infer the implicit skeleton structure for a query action, an RGB video, and then propose to expand this query with the inferred skeleton for improving the performance of retrieval. It is inspired by the observation that skeleton structures can compactly and effectively represent human actions, and are helpful in bridging the semantic gap in action retrieval. The focal point is hence on action skeleton estimation in RGB videos. Specifically, an iterative training procedure is developed to select relevant training data for inferring the skeleton of an input action, since corrupt training data not only degrades performance but also complicates the learning process. Through the iterations, relevant training data are gradually revealed, while more accurate skeletons are inferred with the refined training set. The proposed approach is evaluated on ChaLearn 2013. Significant performance gains in action retrieval are achieved with the aid of the inferred skeletons.","PeriodicalId":434817,"journal":{"name":"HuEvent '14","volume":"158 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-11-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133063013","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
HuEvent '14Pub Date : 2014-11-07DOI: 10.1145/2660505.2660511
Zengkai Wang, Junqing Yu
{"title":"Using Minute-by-Minute Match Report for Semantic Event Annotation in Soccer Video","authors":"Zengkai Wang, Junqing Yu","doi":"10.1145/2660505.2660511","DOIUrl":"https://doi.org/10.1145/2660505.2660511","url":null,"abstract":"In this work, we propose a soccer video annotation approach based on semantic matching with coarse time constraint, where video event and external text information - match report are synchronized by their semantic correspondence along the temporal sequences. Different from the state of the art soccer video analysis methods which assume that the time of event occurrence is given precisely in second, this work solves the problem that how to annotate the soccer video using the match report with coarse gained time information. Compared with previous approaches, the contributions of our approach include the following. 1) The approach synchronizes the video content and text description by their high-level semantics with coarse time constraint instead of the exact timestamp. In fact, most of the text descriptions from the famous sport websites provide the coarse time information in minutes rather than seconds. Therefore, we argue that our approach is more generalized. 2) We propose an attack-defense transition analysis (ADTA) based soccer video event boundary detection method. The previous methods give coarse boundaries which could be refined, or simply give the clips with fixed duration which may cause larger bias. The results of our method are more in line with the development process of soccer events. 3) Different with the existing audio features analysis based whistle detection method, we propose a novel Hough transformation based whistle detection algorithm from the perspective of image processing, which facilitates the game start time detection combing with the ellipse detection algorithm, and further helps the synchronization of video and text events. The experimental results conducted on large amount of soccer videos validated the effectiveness of our proposed approach.","PeriodicalId":434817,"journal":{"name":"HuEvent '14","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-11-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114624827","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
HuEvent '14Pub Date : 2014-11-07DOI: 10.1145/2660505.2660508
T. Kajiyama, S. Satoh
{"title":"User Emotion Sensing in Search Process based on Chromatic Sensation","authors":"T. Kajiyama, S. Satoh","doi":"10.1145/2660505.2660508","DOIUrl":"https://doi.org/10.1145/2660505.2660508","url":null,"abstract":"The sensing of user emotion in the Web space is typically performed using user logs, e.g., pages visiting by users and/or the texts input by users. These techniques are essentially based on text to identify the target in which the users showed interest. However, it is difficult to estimate the emotion only by text. On the other hand, it is known that the color sensation has direct connection to human's affective sensation and thus is suitable for emotion sensing. Based on this, we propose a model for sensing user emotion on the basis of the psychological principle, namely, color sensation. The model extracts the present feeling of the user by using the colors browsed or selected by the user. The model has five elements: an algorithm for extracting feature colors that potentially represent user emotion, an emotion database for describing the relationships between emotions and colors, an algorithm for extracting the user emotion from the extracted feature color, images symbolizing information for making it easier for users to find relevant information that match their present feeling, and an interface for browsing information from sense-related viewpoints. As a first step in implementing this model, it was tested using 50 people searching for an application using an intuitive interface. The results revealed that the colors they selected were potentially related to their feelings especially for users with ambiguous information needs.","PeriodicalId":434817,"journal":{"name":"HuEvent '14","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-11-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116849711","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}