Workshop on Perceptive User Interfaces最新文献

筛选
英文 中文
Speech driven facial animation 语音驱动面部动画
Workshop on Perceptive User Interfaces Pub Date : 2001-11-15 DOI: 10.1145/971478.971488
P. Kakumanu, R. Gutierrez-Osuna, A. Esposito, R. Bryll, A. Goshtasby, O. Garcia
{"title":"Speech driven facial animation","authors":"P. Kakumanu, R. Gutierrez-Osuna, A. Esposito, R. Bryll, A. Goshtasby, O. Garcia","doi":"10.1145/971478.971488","DOIUrl":"https://doi.org/10.1145/971478.971488","url":null,"abstract":"The results reported in this article are an integral part of a larger project aimed at achieving perceptually realistic animations, including the individualized nuances, of three-dimensional human faces driven by speech. The audiovisual system that has been developed for learning the spatio-temporal relationship between speech acoustics and facial animation is described, including video and speech processing, pattern analysis, and MPEG-4 compliant facial animation for a given speaker. In particular, we propose a perceptual transformation of the speech spectral envelope, which is shown to capture the dynamics of articulatory movements. An efficient nearest-neighbor algorithm is used to predict novel articulatory trajectories from the speech dynamics. The results are very promising and suggest a new way to approach the modeling of synthetic lip motion of a given speaker driven by his/her speech. This would also provide clues toward a more general cross-speaker realistic animation.","PeriodicalId":416822,"journal":{"name":"Workshop on Perceptive User Interfaces","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-11-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130012068","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 38
WebContext: remote access to shared context WebContext:远程访问共享上下文
Workshop on Perceptive User Interfaces Pub Date : 2001-11-15 DOI: 10.1145/971478.971517
Robert G. Capra, Manuel A. Pérez-Quiñones, Naren Ramakrishnan
{"title":"WebContext: remote access to shared context","authors":"Robert G. Capra, Manuel A. Pérez-Quiñones, Naren Ramakrishnan","doi":"10.1145/971478.971517","DOIUrl":"https://doi.org/10.1145/971478.971517","url":null,"abstract":"In this paper, we describe a system and architecture for building and remotely accessing shared context between a user and a computer. The system is designed to allow a user to browse web pages on a personal computer and then remotely make queries about information seen on the web pages using a telephone-based voice user interface.","PeriodicalId":416822,"journal":{"name":"Workshop on Perceptive User Interfaces","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-11-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134080122","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 14
Recognizing movements from the ground reaction force 识别来自地面的反作用力
Workshop on Perceptive User Interfaces Pub Date : 2001-11-15 DOI: 10.1145/971478.971523
R. Headon, R. Curwen
{"title":"Recognizing movements from the ground reaction force","authors":"R. Headon, R. Curwen","doi":"10.1145/971478.971523","DOIUrl":"https://doi.org/10.1145/971478.971523","url":null,"abstract":"This paper presents a novel approach to movement recognition, using the vertical component of a person's Ground Reaction Force (GRF). Typical primitive movements such as taking a step, jumping, drop-landing, sitting down, rising to stand and crouching are decomposed and recognized in terms of the GRF signal observed by a weight sensitive floor. Previous works focused on vision processing for movement recognition. This work provides a new sensor modality for a larger research effort, that of sentient computing, which is concerned with giving computers awareness of their environment and inhabitants.","PeriodicalId":416822,"journal":{"name":"Workshop on Perceptive User Interfaces","volume":"48 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-11-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134415677","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 46
Using eye movements to determine referents in a spoken dialogue system 在口语对话系统中使用眼球运动来确定指涉物
Workshop on Perceptive User Interfaces Pub Date : 2001-11-15 DOI: 10.1145/971478.971489
E. Campana, Jason Baldridge, J. Dowding, Beth Ann Hockey, R. Remington, L. Stone
{"title":"Using eye movements to determine referents in a spoken dialogue system","authors":"E. Campana, Jason Baldridge, J. Dowding, Beth Ann Hockey, R. Remington, L. Stone","doi":"10.1145/971478.971489","DOIUrl":"https://doi.org/10.1145/971478.971489","url":null,"abstract":"Most computational spoken dialogue systems take a \"literary\" approach to reference resolution. With this type of approach, entities that are mentioned by a human interactor are unified with elements in the world state based on the same principles that guide the process during text interpretation. In human-to-human interaction, however, referring is a much more collaborative process. Participants often under-specify their referents, relying on their discourse partners for feedback if more information is needed to uniquely identify a particular referent. By monitoring eye-movements during this interaction, it is possible to improve the performance of a spoken dialogue system on referring expressions that are underspecified according to the literary model. This paper describes a system currently under development that employs such a strategy.","PeriodicalId":416822,"journal":{"name":"Workshop on Perceptive User Interfaces","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-11-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126797285","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 38
User and social interfaces by observing human faces for intelligent wheelchairs 通过观察人脸为智能轮椅提供用户和社交界面
Workshop on Perceptive User Interfaces Pub Date : 2001-11-15 DOI: 10.1145/971478.971514
Y. Kuno, Y. Murakami, N. Shimada
{"title":"User and social interfaces by observing human faces for intelligent wheelchairs","authors":"Y. Kuno, Y. Murakami, N. Shimada","doi":"10.1145/971478.971514","DOIUrl":"https://doi.org/10.1145/971478.971514","url":null,"abstract":"With the increase in the number of senior citizens, there is a growing demand for human-friendly wheelchairs as mobility aids. Thus several intelligent wheelchairs have been proposed recently. However, they consider friendliness only to their users. Since wheelchairs move among people, they should also be friendly to people around them. In other words, they should have a social-friendly interface as well as a user-friendly interface. We propose an intelligent wheelchair that is friendly to both user and people around it by observing the faces of both user and others. The user can control it by turning his/her face in the direction where he/she would like to turn. It observes pedestrian's face and changes its collision avoidance method depending on whether or not he/she notices it. Here we assume that the pedestrian notices the wheelchair if his/her face often faces toward the wheelchair.","PeriodicalId":416822,"journal":{"name":"Workshop on Perceptive User Interfaces","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-11-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127393180","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 20
Perceptive user interfaces workshop 感知用户界面研讨会
Workshop on Perceptive User Interfaces Pub Date : 2001-11-15 DOI: 10.1145/971478.971485
D. Schmorrow, J. Patrey
{"title":"Perceptive user interfaces workshop","authors":"D. Schmorrow, J. Patrey","doi":"10.1145/971478.971485","DOIUrl":"https://doi.org/10.1145/971478.971485","url":null,"abstract":"One of the shortcomings of traditional human-computer interaction is its failure to be tailored specifically for human cognition. Human cognition has particular virtues and limitations; designing interfaces to maximize the virtues and bolster the limitations could produce substantial gains in information management capability. These cognition-centric design principles strive to move beyond mere human-computer interaction and towards human-computer symbiosis – a catalytic marriage between the heuristic-driven, contextsensitive powers of human cognition and the detail-oriented, data crunching might of computer computation. This symbiosis becomes feasible due to progress made during the “Decade of the Brain” in expanding the understanding of brain mechanisms and introduction of novel non-invasive assessment tools (such as fMRI), the ongoing “Cognitive Revolution” in behavioral science producing advances in the science of problems solving, reasoning, and decision making, and the growth of digital technologies in pure computing power, miniaturization and ruggedization, data mining sophistication, and evolving advancements in robust input/output devices. These advances produce four significant content domains: multimodal interaction, shared context, interested management, and a new generation of human factors issues. Traditional computer systems rely almost solely on visual information (with a meager auditory component) – future systems will be inherently multimodal, relying on all sensory and motor processing channels for receiving and conveying information. Traditional computer systems also are restricted because humans and computers operate within different contexts – computers are wholly unaware of cues that humans give the highest priority or how to capitalize on those cues to help humans better process information. Similarly, computers lack the ability to truly ‘serve’ the user and determine what information in an environment should be omitted, what should be highlighted, and what should be portrayed with accuracy (and what determines sufficient accuracy). Finally, the advent of these new tools in the human and computer domain requires a new generation of human factors design issues be addressed. Our panel has LCDR Dylan Schmorrow acting as Chair. LCDR Schmorrow is the program officer for the Defense Advance Research Projects Agency Information Technology Office’s “Augmented Cognition” program. LT Jim Patrey will act as cochair; LT Patrey is the Director of International Programs for DARPA’s “Augmented Cognition” program. The proposed panel will each address one of the four content areas identified as significant within “Augmented Cognition.” Dr. Phil Cohen of the Center for Human Computer Interaction will address issues of multimodality. Dr. Denny Proffitt will discuss the value of context as an augmentation tool. Dr. Mike Zyda, of the","PeriodicalId":416822,"journal":{"name":"Workshop on Perceptive User Interfaces","volume":"48 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-11-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127635996","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
An experimental multilingual speech translation system 一个实验性的多语种语音翻译系统
Workshop on Perceptive User Interfaces Pub Date : 2001-11-15 DOI: 10.1145/971478.971491
K. Matsui, Yumi Wakita, T. Konuma, K. Mizutani, M. Endo, M. Murata
{"title":"An experimental multilingual speech translation system","authors":"K. Matsui, Yumi Wakita, T. Konuma, K. Mizutani, M. Endo, M. Murata","doi":"10.1145/971478.971491","DOIUrl":"https://doi.org/10.1145/971478.971491","url":null,"abstract":"In this paper, we describe an experimental speech translation system utilizing small, PC-based hardware with multi-modal user interface. Two major problems for people using an automatic speech translation device are speech recognition errors and language translation errors. In this paper we focus on developing techniques to overcome these problems. The techniques include a new language translation approach based on example sentences, simplified expression rules, and a multi-modal user interface which shows possible speech recognition candidates retrieved from the example sentences. Combination of the proposed techniques can provide accurate language translation performance even if the speech recognition result contains some errors. We propose to use keyword classes by looking at the dependency between keywords to detect the misrecognized keywords and to search the example expressions. Then, the suitable example expression is chosen using a touch panel or by pushing buttons. The language translation picks up the expression in the other language, which should always be grammatically correct. Simplified translated expressions are realized by speech-act based simplifying rules so that the system can avoid various redundant expressions. A simple comparison study showed that the proposed method outputs almost 2 to 10 times faster than a conventional translation device.","PeriodicalId":416822,"journal":{"name":"Workshop on Perceptive User Interfaces","volume":"65 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-11-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122811469","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
The Bayes Point Machine for computer-user frustration detection via pressuremouse 通过压力鼠标检测计算机用户挫败感的贝叶斯点机
Workshop on Perceptive User Interfaces Pub Date : 2001-11-15 DOI: 10.1145/971478.971495
Yuan Qi, Carson Reynolds, Rosalind W. Picard
{"title":"The Bayes Point Machine for computer-user frustration detection via pressuremouse","authors":"Yuan Qi, Carson Reynolds, Rosalind W. Picard","doi":"10.1145/971478.971495","DOIUrl":"https://doi.org/10.1145/971478.971495","url":null,"abstract":"We mount eight pressure sensors on a computer mouse and collect mouse pressure signals from subjects who fill out web forms containing usability bugs. This approach is based on a hypothesis that subjects tend to apply excess pressure to the mouse after encountering frustrating events. We then train a Bayes Point Machine in an attempt to classify two regions of each user's behavior: mouse pressure where the form- filling process is proceeding smoothly, and mouse pressure following a usability bug. Different from current popular classifiers such as the Support Vector Machine, the Bayes Point Machine is a new classification technique rooted in the Bayesian theory. Trained with a new efficient Bayesian approximation algorithm, Expectation Propagation, the Bayes Point Machine achieves a person-dependent classification accuracy rate of 88%, which outperforms the Support Vector Machine in our experiments. The resulting system can be used for many applications in human-computer interaction including adaptive interface design.","PeriodicalId":416822,"journal":{"name":"Workshop on Perceptive User Interfaces","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-11-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116319856","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 57
A multimodal presentation planner for a home entertainment environment 用于家庭娱乐环境的多模式演示计划器
Workshop on Perceptive User Interfaces Pub Date : 2001-11-15 DOI: 10.1145/971478.971492
C. Elting, G. Michelitsch
{"title":"A multimodal presentation planner for a home entertainment environment","authors":"C. Elting, G. Michelitsch","doi":"10.1145/971478.971492","DOIUrl":"https://doi.org/10.1145/971478.971492","url":null,"abstract":"In this paper we outline the design and the implementation of the multimodal presentation planner PMO, which is part of the EMBASSI intelligent user interface for home entertainment devices. We provide details about the concepts we use to produce cohesive and coherent output as well as illustrate the software architecture of the PMO. We compare our approach with the state of the art in presentation planning and conclude with an illustration of our future work.","PeriodicalId":416822,"journal":{"name":"Workshop on Perceptive User Interfaces","volume":"74 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-11-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132453825","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 14
Human-robot interface based on the mutual assistance between speech and vision 基于语音和视觉相互辅助的人机界面
Workshop on Perceptive User Interfaces Pub Date : 2001-11-15 DOI: 10.1145/971478.971483
Mitsutoshi Yoshizaki, Y. Kuno, A. Nakamura
{"title":"Human-robot interface based on the mutual assistance between speech and vision","authors":"Mitsutoshi Yoshizaki, Y. Kuno, A. Nakamura","doi":"10.1145/971478.971483","DOIUrl":"https://doi.org/10.1145/971478.971483","url":null,"abstract":"This paper presents a user interface for a service robot that can bring the objects asked by the user. Speech-based interface is appropriate for this application. However, it alone is not sufficient. The system needs a vision-based interface to recognize gestures as well. Moreover, it needs vision capabilities to obtain the real world information about the objects mentioned in the user's speech. For example, the robot needs to find the target object ordered by speech to carry out the task. This can be considered that vision assists speech. However, vision sometimes fails to detect the objects. Moreover, there are objects for which vision cannot be expected to work well. In these cases, the robot tells the current status to the user so that he/she can give advice by speech to the robot. This can be considered that speech assists vision through the user. This paper presents how the mutual assistance between speech and vision works and demonstrates promising results through experiments.","PeriodicalId":416822,"journal":{"name":"Workshop on Perceptive User Interfaces","volume":"43 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-11-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124113781","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信