Companion Publication of the 2022 International Conference on Multimodal Interaction最新文献

筛选
英文 中文
Enabling Non-Technical Domain Experts to Create Robot-Assisted Therapeutic Scenarios via Visual Programming 使非技术领域的专家能够通过可视化编程创建机器人辅助的治疗方案
Christian Schütze, André Groß, B. Wrede, Birte Richter
{"title":"Enabling Non-Technical Domain Experts to Create Robot-Assisted Therapeutic Scenarios via Visual Programming","authors":"Christian Schütze, André Groß, B. Wrede, Birte Richter","doi":"10.1145/3536220.3558072","DOIUrl":"https://doi.org/10.1145/3536220.3558072","url":null,"abstract":"In this paper, we present a visual programming software for enabling non-technical domain experts to create robot-assisted therapy scenarios for multiple robotic platforms. Our new approach is evaluated by comparing it with Choregraphe, the standard visual programming framework for the often used robotics platforms Pepper and NAO. We could show that our approach receives higher usability values and allows users to perform better in some practical tasks, including understanding, changing and creating small robot-assisted therapy scenarios.","PeriodicalId":186796,"journal":{"name":"Companion Publication of the 2022 International Conference on Multimodal Interaction","volume":"17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-11-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130198624","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Exploring Facial Metric Normalization For Within- and Between-Subject Comparisons in a Multimodal Health Monitoring Agent 在多模态健康监测代理中探索受试者内部和受试者之间比较的面部度量归一化
Oliver Roesler, Hardik Kothare, William Burke, Michael Neumann, J. Liscombe, Andrew Cornish, Doug Habberstad, D. Pautler, David Suendermann-Oeft, Vikram Ramanarayanan
{"title":"Exploring Facial Metric Normalization For Within- and Between-Subject Comparisons in a Multimodal Health Monitoring Agent","authors":"Oliver Roesler, Hardik Kothare, William Burke, Michael Neumann, J. Liscombe, Andrew Cornish, Doug Habberstad, D. Pautler, David Suendermann-Oeft, Vikram Ramanarayanan","doi":"10.1145/3536220.3558071","DOIUrl":"https://doi.org/10.1145/3536220.3558071","url":null,"abstract":"The use of facial metrics obtained through remote web-based platforms has shown promising results for at-home assessment of facial function in multiple neurological and mental disorders. However, an important factor influencing the utility of the obtained metrics is the variability within and across participant sessions due to position and movement of the head relative to the camera. In this paper, we investigate two different facial landmark predictors in combination with four different normalization methods with respect to their effect on the utility of facial metrics obtained through a multimodal assessment platform. We analyzed 38 people with Parkinson’s disease (pPD) and 22 healthy controls who were asked to complete four interactive sessions, a week apart from each other. We find that metrics extracted through MediaPipe clearly outperform metrics extracted through OpenCV and Dlib in terms of test-retest reliability and patient-control discriminability. Furthermore, our results suggest that using the inter-caruncular distance to normalize all raw visual measurements prior to metric computation is optimal for between-subject analyses, while raw measurements (without normalization) can also be used for within-subject comparisons.","PeriodicalId":186796,"journal":{"name":"Companion Publication of the 2022 International Conference on Multimodal Interaction","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-11-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126456515","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
ReCell: replicating recurrent cell for auto-regressive pose generation ReCell:复制复发细胞以产生自退姿态
V. Korzun, Anna Beloborodova, Arkady Ilin
{"title":"ReCell: replicating recurrent cell for auto-regressive pose generation","authors":"V. Korzun, Anna Beloborodova, Arkady Ilin","doi":"10.1145/3536220.3558801","DOIUrl":"https://doi.org/10.1145/3536220.3558801","url":null,"abstract":"This paper describes FineMotion’s gesture generating system entry for the GENEA Challenge 2022. Our system is based on auto-regressive approach imitating recurrent cell. Combined with a special windowed auto-encoder and training approach this system generates plausible gestures appropriate to input speech.","PeriodicalId":186796,"journal":{"name":"Companion Publication of the 2022 International Conference on Multimodal Interaction","volume":"58 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-11-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133612754","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
How can Interaction Data be Contextualized with Mobile Sensing to Enhance Learning Engagement Assessment in Distance Learning? 如何将互动数据与移动传感结合起来,以加强远程学习的学习投入评估?
George-Petru Ciordas-Hertel, Daniel Biedermann, M. Winter, Julia Mordel, H. Drachsler
{"title":"How can Interaction Data be Contextualized with Mobile Sensing to Enhance Learning Engagement Assessment in Distance Learning?","authors":"George-Petru Ciordas-Hertel, Daniel Biedermann, M. Winter, Julia Mordel, H. Drachsler","doi":"10.1145/3536220.3558037","DOIUrl":"https://doi.org/10.1145/3536220.3558037","url":null,"abstract":"Multimodal learning analytics can enrich interaction data with contextual information through mobile sensing. Information about, for example, the physical environment, movement, physiological signals, or smart wearable usage. Through the use of smart wearables, contextual information can thus be captured and made available again to students in further processing steps so that they can reflect and annotate it. This paper describes a software infrastructure and a study design that successfully captured contextual information utilizing mobile sensing using students’ smart wearables in distance learning. In the conducted study, data was collected from the smartphones of 76 students as they self-directedly participated in an online learning unit using a learning management system (LMS) over a two-week period. During the students’ active phases in the LMS, interaction data as well as state and trait measurements were collected by the LMS. Simultaneously, hardware sensor data, app usage data, interaction with notifications, and ecological momentary assessments (EMA) were automatically but transparently collected from the students’ smartphones. Finally, this paper describes some preliminary insights from the study process and their implications for further data processing.","PeriodicalId":186796,"journal":{"name":"Companion Publication of the 2022 International Conference on Multimodal Interaction","volume":"32 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-11-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121851671","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Predicting evaluations of entrepreneurial pitches based on multimodal nonverbal behavioral cues and self-reported characteristics 基于多模态非语言行为线索和自我报告特征的创业推介预测评价
Kostas Stoitsas, Itır Önal Ertuğrul, Werner Liebregts, Merel M. Jung
{"title":"Predicting evaluations of entrepreneurial pitches based on multimodal nonverbal behavioral cues and self-reported characteristics","authors":"Kostas Stoitsas, Itır Önal Ertuğrul, Werner Liebregts, Merel M. Jung","doi":"10.1145/3536220.3558041","DOIUrl":"https://doi.org/10.1145/3536220.3558041","url":null,"abstract":"Acquiring funding for a startup venture often involves pitching a business idea to potential investors. Insight into the nonverbal behavioral cues that impact the investment decision making process can help entrepreneurs to improve their persuasion skills and can provide valuable insights to investors and researchers. Previous research on the prediction of investment decisions in entrepreneurial pitches has primarily focused on analyzing (usually unimodal) behavioral cues from pitchers only. To address this gap, in this study we compare the predictive performance of different feature sets consisting of nonverbal behavior cues from different modalities (i.e., facial expressions, head movement, and vocal expressions) from both pitchers and investors and their self-reported characteristics. Our findings show promising results for the prediction of investor’s evaluations of entrepreneurial pitches. Multimodal behavioral cues, especially head movement and vocal expressions, were found to be most predictive.","PeriodicalId":186796,"journal":{"name":"Companion Publication of the 2022 International Conference on Multimodal Interaction","volume":"GE-25 5","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-11-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121009761","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Head Movement Patterns during Face-to-Face Conversations Vary with Age 面对面交谈时的头部运动模式随年龄而变化
Denisa Qori Mcdonald, C. Zampella, E. Sariyanidi, Aashvi Manakiwala, Ellis Dejardin, J. Herrington, R. Schultz, B. Tunç
{"title":"Head Movement Patterns during Face-to-Face Conversations Vary with Age","authors":"Denisa Qori Mcdonald, C. Zampella, E. Sariyanidi, Aashvi Manakiwala, Ellis Dejardin, J. Herrington, R. Schultz, B. Tunç","doi":"10.1145/3536220.3563366","DOIUrl":"https://doi.org/10.1145/3536220.3563366","url":null,"abstract":"Advances in computational behavior analysis have the potential to increase our understanding of behavioral patterns and developmental trajectories in neurotypical individuals, as well as in individuals with mental health conditions marked by motor, social, and emotional difficulties. This study focuses on investigating how head movement patterns during face–to–face conversations vary with age from childhood through adulthood. We rely on computer vision techniques due to their suitability for analysis of social behaviors in naturalistic settings, since video data capture can be unobtrusively embedded within conversations between two social partners. The methods in this work include unsupervised learning for movement pattern clustering, and supervised classification and regression as a function of age. The results demonstrate that 3–minute video recordings of head movements during conversations show patterns that distinguish between participants that are younger vs. older than 12 years with accuracy. Additionally, we extract relevant patterns of head movement upon which the age distinction was determined by our models.","PeriodicalId":186796,"journal":{"name":"Companion Publication of the 2022 International Conference on Multimodal Interaction","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-11-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128202449","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Automatic facial expressions, gaze direction and head movements generation of a virtual agent 自动面部表情,凝视方向和头部运动生成的虚拟代理
Alice Delbosc, M. Ochs, S. Ayache
{"title":"Automatic facial expressions, gaze direction and head movements generation of a virtual agent","authors":"Alice Delbosc, M. Ochs, S. Ayache","doi":"10.1145/3536220.3558806","DOIUrl":"https://doi.org/10.1145/3536220.3558806","url":null,"abstract":"In this article, we present two models to jointly and automatically generate the head, facial and gaze movements of a virtual agent from acoustic speech features. Two architectures are explored: a Generative Adversarial Network and an Adversarial Encoder-Decoder. Head movements and gaze orientation are generated as 3D coordinates, while facial expressions are generated using action units based on the facial action coding system. A large corpus of almost 4 hours of videos, involving 89 different speakers is used to train our models. We extract the speech and visual features automatically from these videos using existing tools. The evaluation of these models is conducted objectively with measures such as density evaluation and a visualisation from PCA reduction, as well as subjectively through a users perceptive study. Our proposed methodology shows that on 15 seconds sequences, encoder-decoder architecture drastically improves the perception of generated behaviours in two criteria: the coordination with speech and the naturalness. Our code can be found in : https://github.com/aldelb/non-verbal-behaviours-generation.","PeriodicalId":186796,"journal":{"name":"Companion Publication of the 2022 International Conference on Multimodal Interaction","volume":"30 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-11-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128634927","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Exploring the Benefits of Spatialized Multimodal Psychophysiological Insights for User Experience Research 探索空间化多模态心理生理学洞察对用户体验研究的好处
F. Simard, Tomy Aumont, Sayeed A. D. Kizuk, Pascal E. Fortin
{"title":"Exploring the Benefits of Spatialized Multimodal Psychophysiological Insights for User Experience Research","authors":"F. Simard, Tomy Aumont, Sayeed A. D. Kizuk, Pascal E. Fortin","doi":"10.1145/3536220.3558039","DOIUrl":"https://doi.org/10.1145/3536220.3558039","url":null,"abstract":"Conducting psychophysiological investigations outside of lab settings has a lot of potential for academic applications as well as for industries concerned about the quality of their user and customer experience. Prior work employing in-the-wild methodologies often focuses on a limited set of biometrics, which constrains the depth of the insights generated by such investigations. In this work, we field tested a new system for multimodal data acquisition and present exploratory results and insights from two ambulatory data collection sessions, held during public events. Through a spatially grounded analysis, we investigate the feasibility, and practicality of multimodal ambulatory psychophysiological inquiries, and hypothesize on areas of added values for public event organizers.","PeriodicalId":186796,"journal":{"name":"Companion Publication of the 2022 International Conference on Multimodal Interaction","volume":"254 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-11-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131521802","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Training Computational Models of Group Processes without Groundtruth: the Self- vs External Assessment’s Dilemma 无基础真理的群体过程训练计算模型:自我评估与外部评估的困境
Lucien Maman, G. Volpe, G. Varni
{"title":"Training Computational Models of Group Processes without Groundtruth: the Self- vs External Assessment’s Dilemma","authors":"Lucien Maman, G. Volpe, G. Varni","doi":"10.1145/3536220.3563687","DOIUrl":"https://doi.org/10.1145/3536220.3563687","url":null,"abstract":"Supervised learning relies on the availability and reliability of the labels used to train computational models. In research areas such as Affective Computing and Social Signal Processing, such labels are usually extracted from multiple self- and/or external assessments. Labels are, then, either aggregated to produce a single groundtruth label, or all used during training, potentially resulting in degrading performance of the models. Defining a “true” label is, however, complex. Labels can be gathered at different times, with different tools, and may contain biases. Furthermore, multiple assessments are usually available for a same sample with potential contradictions. Thus, it is crucial to devise strategies that can take advantage of both self- and external assessments to train computational models without a reliable groundtruth. In this study, we designed and tested 3 of such strategies with the aim of mitigating the biases and making the models more robust to uncertain labels. Results show that the strategy based on weighting the loss during training according to a measure of disagreement improved the performances of the baseline, hence, underlining the potential of such an approach.","PeriodicalId":186796,"journal":{"name":"Companion Publication of the 2022 International Conference on Multimodal Interaction","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-11-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131049782","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
To Improve Is to Change: Towards Improving Mood Prediction by Learning Changes in Emotion 改善就是改变:通过学习情绪变化来改善情绪预测
S. Narayana, Ramanathan Subramanian, Ibrahim Radwan, Roland Goecke
{"title":"To Improve Is to Change: Towards Improving Mood Prediction by Learning Changes in Emotion","authors":"S. Narayana, Ramanathan Subramanian, Ibrahim Radwan, Roland Goecke","doi":"10.1145/3536220.3563685","DOIUrl":"https://doi.org/10.1145/3536220.3563685","url":null,"abstract":"Although the terms mood and emotion are closely related and often used interchangeably, they are distinguished based on their duration, intensity and attribution. To date, hardly any computational models have (a) examined mood recognition, and (b) modelled the interplay between mood and emotional state in their analysis. In this paper, as a first step towards mood prediction, we propose a framework that utilises both dominant emotion (or mood) labels, and emotional change labels on the AFEW-VA database. Experiments evaluating unimodal (trained only using mood labels) and multimodal (trained with both mood and emotion change labels) convolutional neural networks confirm that incorporating emotional change information in the network training process can significantly improve the mood prediction performance, thus highlighting the importance of modelling emotion and mood simultaneously for improved performance in affective state recognition.","PeriodicalId":186796,"journal":{"name":"Companion Publication of the 2022 International Conference on Multimodal Interaction","volume":"56 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130578295","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信