Robert G. Capra, Manuel A. Pérez-Quiñones, Naren Ramakrishnan
{"title":"WebContext: remote access to shared context","authors":"Robert G. Capra, Manuel A. Pérez-Quiñones, Naren Ramakrishnan","doi":"10.1145/971478.971517","DOIUrl":"https://doi.org/10.1145/971478.971517","url":null,"abstract":"In this paper, we describe a system and architecture for building and remotely accessing shared context between a user and a computer. The system is designed to allow a user to browse web pages on a personal computer and then remotely make queries about information seen on the web pages using a telephone-based voice user interface.","PeriodicalId":416822,"journal":{"name":"Workshop on Perceptive User Interfaces","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-11-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134080122","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Recognizing movements from the ground reaction force","authors":"R. Headon, R. Curwen","doi":"10.1145/971478.971523","DOIUrl":"https://doi.org/10.1145/971478.971523","url":null,"abstract":"This paper presents a novel approach to movement recognition, using the vertical component of a person's Ground Reaction Force (GRF). Typical primitive movements such as taking a step, jumping, drop-landing, sitting down, rising to stand and crouching are decomposed and recognized in terms of the GRF signal observed by a weight sensitive floor. Previous works focused on vision processing for movement recognition. This work provides a new sensor modality for a larger research effort, that of sentient computing, which is concerned with giving computers awareness of their environment and inhabitants.","PeriodicalId":416822,"journal":{"name":"Workshop on Perceptive User Interfaces","volume":"48 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-11-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134415677","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
P. Kakumanu, R. Gutierrez-Osuna, A. Esposito, R. Bryll, A. Goshtasby, O. Garcia
{"title":"Speech driven facial animation","authors":"P. Kakumanu, R. Gutierrez-Osuna, A. Esposito, R. Bryll, A. Goshtasby, O. Garcia","doi":"10.1145/971478.971488","DOIUrl":"https://doi.org/10.1145/971478.971488","url":null,"abstract":"The results reported in this article are an integral part of a larger project aimed at achieving perceptually realistic animations, including the individualized nuances, of three-dimensional human faces driven by speech. The audiovisual system that has been developed for learning the spatio-temporal relationship between speech acoustics and facial animation is described, including video and speech processing, pattern analysis, and MPEG-4 compliant facial animation for a given speaker. In particular, we propose a perceptual transformation of the speech spectral envelope, which is shown to capture the dynamics of articulatory movements. An efficient nearest-neighbor algorithm is used to predict novel articulatory trajectories from the speech dynamics. The results are very promising and suggest a new way to approach the modeling of synthetic lip motion of a given speaker driven by his/her speech. This would also provide clues toward a more general cross-speaker realistic animation.","PeriodicalId":416822,"journal":{"name":"Workshop on Perceptive User Interfaces","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-11-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130012068","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
E. Campana, Jason Baldridge, J. Dowding, Beth Ann Hockey, R. Remington, L. Stone
{"title":"Using eye movements to determine referents in a spoken dialogue system","authors":"E. Campana, Jason Baldridge, J. Dowding, Beth Ann Hockey, R. Remington, L. Stone","doi":"10.1145/971478.971489","DOIUrl":"https://doi.org/10.1145/971478.971489","url":null,"abstract":"Most computational spoken dialogue systems take a \"literary\" approach to reference resolution. With this type of approach, entities that are mentioned by a human interactor are unified with elements in the world state based on the same principles that guide the process during text interpretation. In human-to-human interaction, however, referring is a much more collaborative process. Participants often under-specify their referents, relying on their discourse partners for feedback if more information is needed to uniquely identify a particular referent. By monitoring eye-movements during this interaction, it is possible to improve the performance of a spoken dialogue system on referring expressions that are underspecified according to the literary model. This paper describes a system currently under development that employs such a strategy.","PeriodicalId":416822,"journal":{"name":"Workshop on Perceptive User Interfaces","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-11-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126797285","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"User and social interfaces by observing human faces for intelligent wheelchairs","authors":"Y. Kuno, Y. Murakami, N. Shimada","doi":"10.1145/971478.971514","DOIUrl":"https://doi.org/10.1145/971478.971514","url":null,"abstract":"With the increase in the number of senior citizens, there is a growing demand for human-friendly wheelchairs as mobility aids. Thus several intelligent wheelchairs have been proposed recently. However, they consider friendliness only to their users. Since wheelchairs move among people, they should also be friendly to people around them. In other words, they should have a social-friendly interface as well as a user-friendly interface. We propose an intelligent wheelchair that is friendly to both user and people around it by observing the faces of both user and others. The user can control it by turning his/her face in the direction where he/she would like to turn. It observes pedestrian's face and changes its collision avoidance method depending on whether or not he/she notices it. Here we assume that the pedestrian notices the wheelchair if his/her face often faces toward the wheelchair.","PeriodicalId":416822,"journal":{"name":"Workshop on Perceptive User Interfaces","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-11-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127393180","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Perceptive user interfaces workshop","authors":"D. Schmorrow, J. Patrey","doi":"10.1145/971478.971485","DOIUrl":"https://doi.org/10.1145/971478.971485","url":null,"abstract":"One of the shortcomings of traditional human-computer interaction is its failure to be tailored specifically for human cognition. Human cognition has particular virtues and limitations; designing interfaces to maximize the virtues and bolster the limitations could produce substantial gains in information management capability. These cognition-centric design principles strive to move beyond mere human-computer interaction and towards human-computer symbiosis – a catalytic marriage between the heuristic-driven, contextsensitive powers of human cognition and the detail-oriented, data crunching might of computer computation. This symbiosis becomes feasible due to progress made during the “Decade of the Brain” in expanding the understanding of brain mechanisms and introduction of novel non-invasive assessment tools (such as fMRI), the ongoing “Cognitive Revolution” in behavioral science producing advances in the science of problems solving, reasoning, and decision making, and the growth of digital technologies in pure computing power, miniaturization and ruggedization, data mining sophistication, and evolving advancements in robust input/output devices. These advances produce four significant content domains: multimodal interaction, shared context, interested management, and a new generation of human factors issues. Traditional computer systems rely almost solely on visual information (with a meager auditory component) – future systems will be inherently multimodal, relying on all sensory and motor processing channels for receiving and conveying information. Traditional computer systems also are restricted because humans and computers operate within different contexts – computers are wholly unaware of cues that humans give the highest priority or how to capitalize on those cues to help humans better process information. Similarly, computers lack the ability to truly ‘serve’ the user and determine what information in an environment should be omitted, what should be highlighted, and what should be portrayed with accuracy (and what determines sufficient accuracy). Finally, the advent of these new tools in the human and computer domain requires a new generation of human factors design issues be addressed. Our panel has LCDR Dylan Schmorrow acting as Chair. LCDR Schmorrow is the program officer for the Defense Advance Research Projects Agency Information Technology Office’s “Augmented Cognition” program. LT Jim Patrey will act as cochair; LT Patrey is the Director of International Programs for DARPA’s “Augmented Cognition” program. The proposed panel will each address one of the four content areas identified as significant within “Augmented Cognition.” Dr. Phil Cohen of the Center for Human Computer Interaction will address issues of multimodality. Dr. Denny Proffitt will discuss the value of context as an augmentation tool. Dr. Mike Zyda, of the","PeriodicalId":416822,"journal":{"name":"Workshop on Perceptive User Interfaces","volume":"48 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-11-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127635996","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
K. Matsui, Yumi Wakita, T. Konuma, K. Mizutani, M. Endo, M. Murata
{"title":"An experimental multilingual speech translation system","authors":"K. Matsui, Yumi Wakita, T. Konuma, K. Mizutani, M. Endo, M. Murata","doi":"10.1145/971478.971491","DOIUrl":"https://doi.org/10.1145/971478.971491","url":null,"abstract":"In this paper, we describe an experimental speech translation system utilizing small, PC-based hardware with multi-modal user interface. Two major problems for people using an automatic speech translation device are speech recognition errors and language translation errors. In this paper we focus on developing techniques to overcome these problems. The techniques include a new language translation approach based on example sentences, simplified expression rules, and a multi-modal user interface which shows possible speech recognition candidates retrieved from the example sentences. Combination of the proposed techniques can provide accurate language translation performance even if the speech recognition result contains some errors. We propose to use keyword classes by looking at the dependency between keywords to detect the misrecognized keywords and to search the example expressions. Then, the suitable example expression is chosen using a touch panel or by pushing buttons. The language translation picks up the expression in the other language, which should always be grammatically correct. Simplified translated expressions are realized by speech-act based simplifying rules so that the system can avoid various redundant expressions. A simple comparison study showed that the proposed method outputs almost 2 to 10 times faster than a conventional translation device.","PeriodicalId":416822,"journal":{"name":"Workshop on Perceptive User Interfaces","volume":"65 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-11-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122811469","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"The Bayes Point Machine for computer-user frustration detection via pressuremouse","authors":"Yuan Qi, Carson Reynolds, Rosalind W. Picard","doi":"10.1145/971478.971495","DOIUrl":"https://doi.org/10.1145/971478.971495","url":null,"abstract":"We mount eight pressure sensors on a computer mouse and collect mouse pressure signals from subjects who fill out web forms containing usability bugs. This approach is based on a hypothesis that subjects tend to apply excess pressure to the mouse after encountering frustrating events. We then train a Bayes Point Machine in an attempt to classify two regions of each user's behavior: mouse pressure where the form- filling process is proceeding smoothly, and mouse pressure following a usability bug. Different from current popular classifiers such as the Support Vector Machine, the Bayes Point Machine is a new classification technique rooted in the Bayesian theory. Trained with a new efficient Bayesian approximation algorithm, Expectation Propagation, the Bayes Point Machine achieves a person-dependent classification accuracy rate of 88%, which outperforms the Support Vector Machine in our experiments. The resulting system can be used for many applications in human-computer interaction including adaptive interface design.","PeriodicalId":416822,"journal":{"name":"Workshop on Perceptive User Interfaces","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-11-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116319856","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A multimodal presentation planner for a home entertainment environment","authors":"C. Elting, G. Michelitsch","doi":"10.1145/971478.971492","DOIUrl":"https://doi.org/10.1145/971478.971492","url":null,"abstract":"In this paper we outline the design and the implementation of the multimodal presentation planner PMO, which is part of the EMBASSI intelligent user interface for home entertainment devices. We provide details about the concepts we use to produce cohesive and coherent output as well as illustrate the software architecture of the PMO. We compare our approach with the state of the art in presentation planning and conclude with an illustration of our future work.","PeriodicalId":416822,"journal":{"name":"Workshop on Perceptive User Interfaces","volume":"74 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-11-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132453825","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Naturally conveyed explanations of device behavior","authors":"Michael Oltmans, Randall Davis","doi":"10.1145/971478.971498","DOIUrl":"https://doi.org/10.1145/971478.971498","url":null,"abstract":"Designers routinely explain their designs to one another using sketches and verbal descriptions of behavior, both of which can be understood long before the device has been fully specified. But current design tools fail almost completely to support this sort of interaction, instead not only forcing designers to specify details of the design, but typically requiring that they do so by navigating a forest of menus and dialog boxes, rather than directly describing the behaviors with sketches and verbal explanations. We have created a prototype system, called assistance, capable of interpreting multimodal explanations for simple 2-D kinematic devices. The program generates a model of the events and the causal relationships between events that have been described via hand drawn sketches, sketched annotations, and verbal descriptions. Our goal is to make the designer's interaction with the computer more like interacting with another designer. This requires the ability not only to understand physical devices but also to understand the means by which the explanations of these devices are conveyed.","PeriodicalId":416822,"journal":{"name":"Workshop on Perceptive User Interfaces","volume":"24 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-11-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121526965","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}