{"title":"A Question of Fidelity: Comparing Different User Testing Methods for Evaluating In-Car Prompts","authors":"Anna-Maria Meck, C. Draxler, Thurid Vogt","doi":"10.1145/3543829.3544519","DOIUrl":null,"url":null,"abstract":"User studies are a major component in any user-centered design process. Testing methods thereby vary tremendously regarding the organizational, financial, and timely effort needed to conduct them. Driving simulator studies generally are the method of choice when dialogs need to be validated for in-car settings. These studies are highly time- and cost-consuming though. Online crowdsourcing studies can be an alternative as they allow for quick results and large sample sizes while at the same time being time- and cost-efficient. Still, voice user interface designers argue for a lack of applicability to concrete use cases. This is especially true for speech dialog systems in an in-car context where users experience voice as a secondary task with the primary task being driving. To compare the validity of different user testing methods, study participants in a between-subjects study design evaluated proactive in-car prompts presented a) in an online crowdsourcing study in text form, b) in an online crowdsourcing study via audio, and c) in a driving simulator. Prompt evaluations did not differ significantly between conditions a) and c) but diverged for condition b). Findings are explained by drawing from the Elaboration Likelihood Model and used to answer the question of how to efficiently validate in-car prompts.","PeriodicalId":138046,"journal":{"name":"Proceedings of the 4th Conference on Conversational User Interfaces","volume":"14 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-07-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 4th Conference on Conversational User Interfaces","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3543829.3544519","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1
Abstract
User studies are a major component in any user-centered design process. Testing methods thereby vary tremendously regarding the organizational, financial, and timely effort needed to conduct them. Driving simulator studies generally are the method of choice when dialogs need to be validated for in-car settings. These studies are highly time- and cost-consuming though. Online crowdsourcing studies can be an alternative as they allow for quick results and large sample sizes while at the same time being time- and cost-efficient. Still, voice user interface designers argue for a lack of applicability to concrete use cases. This is especially true for speech dialog systems in an in-car context where users experience voice as a secondary task with the primary task being driving. To compare the validity of different user testing methods, study participants in a between-subjects study design evaluated proactive in-car prompts presented a) in an online crowdsourcing study in text form, b) in an online crowdsourcing study via audio, and c) in a driving simulator. Prompt evaluations did not differ significantly between conditions a) and c) but diverged for condition b). Findings are explained by drawing from the Elaboration Likelihood Model and used to answer the question of how to efficiently validate in-car prompts.