{"title":"Voice Puppetry: Speech Synthesis Adventures in Human Centred AI","authors":"M. Aylett, Yolanda Vazquez-Alvarez","doi":"10.1145/3379336.3381478","DOIUrl":"https://doi.org/10.1145/3379336.3381478","url":null,"abstract":"State-of-the-art speech synthesis owes much to modern AI machine learning, with recurrent neural networks becoming the new standard. However, how you say something is just as important as what you say. If we draw inspiration from human dramatic performance, ideas such as artistic direction can help us design interactive speech synthesis systems which can be finely controlled by a human voice. This \"voice puppetry\" has many possible applications from film dubbing to the pre-creation of prompts for a conversational agent. Previous work in voice puppetry has raised the question of how such a system should work and how we might interact with it. Here, we share the results of a focus group discussing voice puppetry and responding to a voice puppetry demo. Results highlight a main challenge in user-centred AI: where is the trade-off between control and automation? and how may users control this trade-off?","PeriodicalId":335081,"journal":{"name":"Proceedings of the 25th International Conference on Intelligent User Interfaces Companion","volume":"31 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-03-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115805283","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Multimodal Analytics for early Diagnosis of Alzheimer's Disease","authors":"Chathurika Jayangani Palliya Guruge","doi":"10.1145/3379336.3381508","DOIUrl":"https://doi.org/10.1145/3379336.3381508","url":null,"abstract":"Alzheimer's Disease(AD) is a neurocognitive disease that causes impairments of cognition as well as Activities of Daily Living (ADLs). This research investigates the possibility of diagnosing the early stages of AD using the impairments of ADLs. It explores the areas which are difficult to assess and undiagnosed in clinical settings using behavioral analytics. The main focus of the research is to identify unrecognized and abnormal behavioral patterns associated with ADLs with the use of multimodal technologies such as hand movements, facial responses, eye gaze, and spontaneous speech patterns which are visible at prodromal stages of AD and evaluate their association with the corresponding mental status.","PeriodicalId":335081,"journal":{"name":"Proceedings of the 25th International Conference on Intelligent User Interfaces Companion","volume":"291 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-03-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122794691","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Exploiting a Large-scale Knowledge Graph for Question Generation in Food Preference Interview Systems","authors":"Jie Zeng, Y. Nakano","doi":"10.1145/3379336.3381504","DOIUrl":"https://doi.org/10.1145/3379336.3381504","url":null,"abstract":"This paper presents a dialogue system that acquires user's food preference through a conversation. First, we proposed a method for selecting relevant topics and generating questions based on Freebase, a large-scale knowledge graph. To select relevant topics, using the Wikipedia corpus, we created a topic-embedding model that represents the correlation among topics. For missing entities in Freebase, knowledge completion was applied using knowledge graph embedding. We incorporated these functions into a dialogue system and conducted a user study. The results reveal that the proposed dialogue system more efficiently elicited words related to food and common nouns, and these words were highly correlated in a word embedding space.","PeriodicalId":335081,"journal":{"name":"Proceedings of the 25th International Conference on Intelligent User Interfaces Companion","volume":"43 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-03-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128029785","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Designing Adaptive Behavior in a Social Robot","authors":"Eleonora Zedda","doi":"10.1145/3379336.3381510","DOIUrl":"https://doi.org/10.1145/3379336.3381510","url":null,"abstract":"Robots are becoming more and more present in our daily activities. In order to improve user interaction with them, it is important to design behaviors in robots that show social attitude and ability to adapt to the users. For this purpose, robots should adapt their behavior recognizing the user's emotion, also considering the actual user with cognitive and physical disabilities. However, most contemporary approaches rarely attempt to consider recognized emotional features in an active manner to modulate robot decision-making and dialogue for the benefit of the user. In this project, I aim to design and implement a module in a humanoid robot to create an adaptive behavior in a Social Robot for older adults who may have cognitive impairments.","PeriodicalId":335081,"journal":{"name":"Proceedings of the 25th International Conference on Intelligent User Interfaces Companion","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-03-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127509414","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Towards Naturally Responsive Spoken Dialog Systems by Modelling Pragmatic-Prosody Correlations of Discourse Markers","authors":"Anindita Nath","doi":"10.1145/3379336.3381490","DOIUrl":"https://doi.org/10.1145/3379336.3381490","url":null,"abstract":"Human speakers in a dialog adapt their responses and the way they convey them to their interlocutors by appropriately tuning their prosody, taking into account the context in which the dialog takes place. Today's spoken dialog systems are incapable of exhibiting such natural responsive behavior. Hence, there is a need for models that enable the selection of better prosody in system responses to make them appropriate to the pragmatic intentions and the dialog context. This submission includes the detailed description of my preliminary study on the prosody of discourse markers, the methods used and my initial findings that corroborate the existence of correlations between prosody and pragmatic intentions of discourse markers in human-human dialogs. These correlations, if modeled accurately, can help dialog systems respond with context-appropriate prosody.","PeriodicalId":335081,"journal":{"name":"Proceedings of the 25th International Conference on Intelligent User Interfaces Companion","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-03-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129632830","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"CrowdForest","authors":"Eun-jeong Kang, Cin Young Hur, Y. S. Choi","doi":"10.1145/3379336.3381486","DOIUrl":"https://doi.org/10.1145/3379336.3381486","url":null,"abstract":"The large number of social platforms developed enable users to express their opinions and access information more freely. However, their algorithmic strategies can have a high possibility of exacerbating a filter bubble or echo chambers which may evoke distinctive emotions response with others. Herein, we present a new online visualization tool for opinion sharing, called CrowdForest, which allows users to visualize their opinions, interacting with others based on semantic figurative metaphors driven by sentiment analysis.","PeriodicalId":335081,"journal":{"name":"Proceedings of the 25th International Conference on Intelligent User Interfaces Companion","volume":"79 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-03-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116807563","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"AI-Based 360-degree Video Generation from Monocular Video for Immersive Experience","authors":"K. Asano, Ryota Suzuki, Yue Qiu, Y. Satoh","doi":"10.1145/3379336.3381477","DOIUrl":"https://doi.org/10.1145/3379336.3381477","url":null,"abstract":"We propose an artificial intelligence (AI)-based framework for generating 360-degree videos from videos recorded by monocular cameras. We also show immersive virtual reality content generation using AI through an analysis of user experience that compares manually designed and AI-generated 360-degree videos based on the proposed framework. The production of 360-degree videos conventionally requires special equipment, such as omni-directional cameras. Our framework is applicable to a massive amount of existing camera and video, hence it increases the availability of 360-degree videos. We implemented our framework in two steps. First, we generate a three-dimensional point cloud from the input video. Then, we apply AI-based methods to interpolate the sparse point cloud based on geometric and semantic information. Our framework will be applicable to several usages such as assisting surveying past traffic accident videos and education on showing the historical townscape of 360 degrees, etc.","PeriodicalId":335081,"journal":{"name":"Proceedings of the 25th International Conference on Intelligent User Interfaces Companion","volume":"33 6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-03-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123806268","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Kashyap Todi, J. Vanderdonckt, Xiaojuan Ma, Jeffrey A. Nichols, N. Banovic
{"title":"AI4AUI","authors":"Kashyap Todi, J. Vanderdonckt, Xiaojuan Ma, Jeffrey A. Nichols, N. Banovic","doi":"10.1145/3379336.3379359","DOIUrl":"https://doi.org/10.1145/3379336.3379359","url":null,"abstract":"This workshop aims at exploring how adaptive user interfaces, i.e., user interface that can modify, change, or adapt themselves based on the user, or their context of use, can benefit from Artificial Intelligence (AI) in general, and Machine Learning (ML) techniques in particular, towards objectively improving some software quality properties, such as usability, aesthetics, reliability, or security. For this purpose, participants will present a case study, and classify their proposed technique in terms of several criteria, such as (but not limited to): input, technique, output, adaptation steps covered, adaptation time, level of automation, software quality properties addressed, measurement method, potential benefits, and drawbacks. These will be then clustered for group discussions according to the aforementioned criteria, such as by technique family or property addressed. From these discussions, an AI4AUI framework will emerge that will be used for positioning, comparing presented techniques, and for generating future avenues.","PeriodicalId":335081,"journal":{"name":"Proceedings of the 25th International Conference on Intelligent User Interfaces Companion","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-03-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127735876","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Antrea Chrysanthou, Pinar Barlas, K. Kyriakou, S. Kleanthous, Jahna Otterbacher
{"title":"Bursting the Bubble: Tool for Awareness and Research about Overpersonalization in Information Access Systems","authors":"Antrea Chrysanthou, Pinar Barlas, K. Kyriakou, S. Kleanthous, Jahna Otterbacher","doi":"10.1145/3379336.3381863","DOIUrl":"https://doi.org/10.1145/3379336.3381863","url":null,"abstract":"Modern information access systems extensively use personalization, automatically filtering and/or ranking content based on the user profile, to guide users to the most relevant material. However, this can also lead to unwanted effects such as the \"filter bubble.\" We present an interactive demonstration system, designed as an educational and research tool, which imitates a search engine, personalizing the search results returned for a query based on the user's characteristics. The system can be tailored to suit any type of audience and context, as well as enabling the collection of responses and interaction data.","PeriodicalId":335081,"journal":{"name":"Proceedings of the 25th International Conference on Intelligent User Interfaces Companion","volume":"44 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-03-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128082611","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Takuya Yonezawa, Yuanyuan Wang, Yukiko Kawai, K. Sumiya
{"title":"Dynamic Video Tag Cloud: A Cooking Support System for Recipe Short Videos","authors":"Takuya Yonezawa, Yuanyuan Wang, Yukiko Kawai, K. Sumiya","doi":"10.1145/3379336.3381476","DOIUrl":"https://doi.org/10.1145/3379336.3381476","url":null,"abstract":"Recently, recipe short videos such as Kurashiru and DELISH KITCHEN have been rapidly gaining attention. These recipe videos can help people learn the essentials of cooking in a short time. However, it is difficult to understand cooking operations by watching a video only one time. Also, since these short recipe videos do not take into account the user's cooking level, anyone can see only the same videos. Therefore, in this paper, we propose a novel cooking support system for recipe short videos, called Dynamic Video Tag Cloud. For this, we first extract cooking operations from a text recipe included in an existing recipe short video. Next, we extract various supplementary recipe information (videos) based on users' cooking levels by weighting the appearance frequency of cooking operations for each cooking genre. Then, the system visualizes supplementary recipe information (videos) to the users in a tag that could interface.","PeriodicalId":335081,"journal":{"name":"Proceedings of the 25th International Conference on Intelligent User Interfaces Companion","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-03-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128157036","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}