{"title":"Smart Multimodal Interaction through Big Data","authors":"Cafer Tosun","doi":"10.1145/2663204.2669985","DOIUrl":"https://doi.org/10.1145/2663204.2669985","url":null,"abstract":"Smart phones and mobile technologies have changed software usage dramatically. Ease of use and simplicity has made software accessible to a huge number of users. In addition, technological advancements in multimodal interaction are opening new frontiers in software. Users are interacting with software systems through multiple channels such as gestures and speech. Touch screens, cameras, sensors, and wearable devices are enablers of this interaction. The user expectation is that the interaction with business software also becomes as simple as the interaction with consumer software. In particular, through the usage of mobile devices, consumer and business software is coming closer together. Next generation software systems and applications will have to enable smart, seamless and contextual multimodal interaction capabilities. New tools, technologies and solutions will be required to increase the ease of use and to build the user experience of the future.","PeriodicalId":389037,"journal":{"name":"Proceedings of the 16th International Conference on Multimodal Interaction","volume":"32 7","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131776334","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Measuring Child Visual Attention using Markerless Head Tracking from Color and Depth Sensing Cameras","authors":"Jonathan Bidwell, Irfan Essa, A. Rozga, G. Abowd","doi":"10.1145/2663204.2663235","DOIUrl":"https://doi.org/10.1145/2663204.2663235","url":null,"abstract":"A child's failure to respond to his or her name being called is an early warning sign for autism and response to name is currently assessed as a part of standard autism screening and diagnostic tools. In this paper, we explore markerless child head tracking as an unobtrusive approach for automatically predicting child response to name. Head turns are used as a proxy for visual attention. We analyzed 50 recorded response to name sessions with the goal of predicting if children, ages 15 to 30 months, responded to name calls by turning to look at an examiner within a defined time interval. The child's head turn angles and hand annotated child name call intervals were extracted from each session. Human assisted tracking was employed using an overhead Kinect camera, and automated tracking was later employed using an additional forward facing camera as a proof-of-concept. We explore two distinct analytical approaches for predicting child responses, one relying on rule-based approached and another on random forest classification. In addition, we derive child response latency as a new measurement that could provide researchers and clinicians with finer grain quantitative information currently unavailable in the field due to human limitations. Finally we reflect on steps for adapting our system to work in less constrained natural settings.","PeriodicalId":389037,"journal":{"name":"Proceedings of the 16th International Conference on Multimodal Interaction","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128415214","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"An Outline of Opportunities for Multimodal Research","authors":"D. Heylen, A. Vinciarelli","doi":"10.1145/2663204.2668320","DOIUrl":"https://doi.org/10.1145/2663204.2668320","url":null,"abstract":"This paper summarizes the contributions to the Workshop \"Roadmapping the Future of Multimodal Interaction Research including Business Opportunities and Challenges\". We present major challenges and ideas for making progress in the field of social signal processing and related fields as presented by the contributors of the workshop.","PeriodicalId":389037,"journal":{"name":"Proceedings of the 16th International Conference on Multimodal Interaction","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128745536","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Data-Driven Model of Nonverbal Behavior for Socially Assistive Human-Robot Interactions","authors":"H. Admoni, B. Scassellati","doi":"10.1145/2663204.2663263","DOIUrl":"https://doi.org/10.1145/2663204.2663263","url":null,"abstract":"Socially assistive robotics (SAR) aims to develop robots that help people through interactions that are inherently social, such as tutoring and coaching. For these interactions to be effective, socially assistive robots must be able to recognize and use nonverbal social cues like eye gaze and gesture. In this paper, we present a preliminary model for nonverbal robot behavior in a tutoring application. Using empirical data from teachers and students in human-human tutoring interactions, the model can be both predictive (recognizing the context of new nonverbal behaviors) and generative (creating new robot nonverbal behaviors based on a desired context) using the same underlying data representation.","PeriodicalId":389037,"journal":{"name":"Proceedings of the 16th International Conference on Multimodal Interaction","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114820796","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Chun-Yen Hsu, Ying-Chao Tung, Han-Yu Wang, Silvia Chyou, Jer-Wei Lin, Mike Y. Chen
{"title":"Glass Shooter: Exploring First-Person Shooter Game Control with Google Glass","authors":"Chun-Yen Hsu, Ying-Chao Tung, Han-Yu Wang, Silvia Chyou, Jer-Wei Lin, Mike Y. Chen","doi":"10.1145/2663204.2669623","DOIUrl":"https://doi.org/10.1145/2663204.2669623","url":null,"abstract":"Smart Glasses offer the opportunity to use head mounted sensors, such as gyroscope and accelerometers, to enable new types of game interaction. To better understand game play experience on Smart Glasses, we recruited 24 participants to play four current games on Google Glass that uses different interaction methods, including gyroscope, voice, touchpad, and in-air gesture. Study results showed that participants were concerned with comfort and social acceptance. Also, their favorite input method was gyroscope, and their favorite game type was First-Person Shooter (FPS) game. Hence, we implemented a FPS game on Google Glass using gyroscope for changing the viewport, and divide FPS controls into four categories: (a)Viewport Control, (b)Aim Control, (c)Fire Control, (d) Move Control. We implemented multiple control method in each category to evaluate and explore glass game control design.","PeriodicalId":389037,"journal":{"name":"Proceedings of the 16th International Conference on Multimodal Interaction","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127005659","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Nicolas Riesterer, C. Becker-Asano, J. Hué, C. Dornhege, B. Nebel
{"title":"The hybrid Agent MARCO","authors":"Nicolas Riesterer, C. Becker-Asano, J. Hué, C. Dornhege, B. Nebel","doi":"10.1145/2663204.2669629","DOIUrl":"https://doi.org/10.1145/2663204.2669629","url":null,"abstract":"We present MARCO, a hybrid, chess playing agent equipped with a custom-built robotic arm and a virtual agent's face displaying emotions. MARCO was built to investigate the hypothesis that hybrid agents capable of displaying emotions make playing chess more personal and enjoyable. In addition, we aim to explore means of achieving emotional contagion between man and machine.","PeriodicalId":389037,"journal":{"name":"Proceedings of the 16th International Conference on Multimodal Interaction","volume":"67 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121866768","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Increasing Customers' Attention using Implicit and Explicit Interaction in Urban Advertisement","authors":"Matthias Wölfel, Luigi Bucchino","doi":"10.1145/2663204.2663268","DOIUrl":"https://doi.org/10.1145/2663204.2663268","url":null,"abstract":"Online advertising campaigns are gaining customers' attention in comparison to advertisement campaigns in the urban space. How to bring back users' attention to advertisement in the urban space by using implicit and explicit interactions of the user is investigated in this publication. We have used age, gender and position estimates to automatically customize the advertisement campaign and 3D gestures to allow the customer to interact with the shown content. To evaluate the overall acceptability and particular aspects of such kind of targeted and interactive advertisement we have developed a prototypical implementation and placed it into a crowded shopping mall. In total 98 random visitors of the mall have experienced the system and answered a questionnaire afterwards.","PeriodicalId":389037,"journal":{"name":"Proceedings of the 16th International Conference on Multimodal Interaction","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126481869","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Multi-Resident Human Behaviour Identification in Ambient Assisted Living Environments","authors":"H. Alemdar","doi":"10.1145/2663204.2666286","DOIUrl":"https://doi.org/10.1145/2663204.2666286","url":null,"abstract":"Multimodal interactions in ambient assisted living environments require human behaviour to be recognized and monitored automatically. The complex nature of human behaviour makes it extremely difficult to infer and adapt to, especially in multi-resident environments. This proposed research aims to contribute to the multimodal interaction community by (i) providing publicly available, naturalistic, rich and annotated datasets for human behaviour modeling, (ii) introducing evaluation methods of several inference methods from a behaviour monitoring perspective, (iii) developing novel methods for recognizing individual behaviour in multi-resident smart environments without assuming any person identification, (iv) proposing methods for mitigating the scalability issues by using transfer, active, and semi-supervised learning techniques. The proposed studies will address both practical and methodological aspects of human behaviour recognition in smart interactive environments.","PeriodicalId":389037,"journal":{"name":"Proceedings of the 16th International Conference on Multimodal Interaction","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127632188","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Exploring multimodality for translator-computer interaction","authors":"Julián Zapata","doi":"10.1145/2663204.2666280","DOIUrl":"https://doi.org/10.1145/2663204.2666280","url":null,"abstract":"Multimodal interaction has the potential to become one of the most efficient, cost-effective and ergonomic working strategies for translation professionals, in the near future. This paper describes our doctoral research project on multimodality in translator-computer interaction (TCI). The specific objective of this project is to observe and analyze the translator experience (TX) with off-the-shelf voice-and-touch-enabled multimodal interfaces, as compared to the interaction with traditional keyboard-and-mouse graphical user interfaces. Overall, this project aims to collect user-centered data and provide recommendations for better-grounded translation tool design. Thus, the study will make an original contribution to the advancement of knowledge in the fields of TCI and human-computer interaction in general.","PeriodicalId":389037,"journal":{"name":"Proceedings of the 16th International Conference on Multimodal Interaction","volume":"37 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134417984","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Written Activity, Representations and Fluency as Predictors of Domain Expertise in Mathematics","authors":"S. Oviatt, Adrienne Cohen","doi":"10.1145/2663204.2663245","DOIUrl":"https://doi.org/10.1145/2663204.2663245","url":null,"abstract":"The emerging field of multimodal learning analytics evaluates natural communication modalities (digital pen, speech, images) to identify domain expertise, learning, and learning-oriented precursors. Using the Math Data Corpus, this research investigated students' digital pen input as small groups collaborated on solving math problems. Compared with non-experts, findings indicated that domain experts have an opposite pattern of accelerating total written activity as problem difficulty increases, a lower written and spoken disfluency rate, and they express different content--including a higher ratio of nonlinguistic symbolic representations and structured diagrams to elemental marks. Implications are discussed for developing reliable multimodal learning analytics systems that incorporate digital pen input to automatically track the consolidation of domain expertise. This includes prediction based on a combination of activity patterns, fluency, and content analysis. New MMLA systems are expected to have special utility on cell phones, which already have multimodal interfaces and are the dominant educational platform worldwide.","PeriodicalId":389037,"journal":{"name":"Proceedings of the 16th International Conference on Multimodal Interaction","volume":"98 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133182717","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}