{"title":"User interviews revisited: identifying user positions and system interpretations","authors":"Eeva Raita","doi":"10.1145/2399016.2399119","DOIUrl":null,"url":null,"abstract":"Throughout the history of HCI, interviews have been utilized for collecting users' subjective evaluations of interactive technology. This paper raises the issue that these interviews are often deployed in a manner overlooking two aspects of evaluation: the relative positions from which the system is evaluated and the interviewees' interpretations of the system. In the study, 14 users of a new information system were asked to evaluate provocative claims about the system's usability. The analyses of their responses reveal two sources of variation: what is being evaluated and who is evaluating it. Interviewees evaluated the system's usability from five user positions: end user, supervisor, organization's representative, co-developer, and outsider. Also, four \"faces\" of the system were interpreted: UI, utility, communication medium, and unknown entity. These findings are employed for drawing of broader conclusions about the system and its use, and procedures for improving user interviews in HCI are presented.","PeriodicalId":352513,"journal":{"name":"Nordic Conference on Human-Computer Interaction","volume":"3 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2012-10-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"7","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Nordic Conference on Human-Computer Interaction","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/2399016.2399119","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 7
Abstract
Throughout the history of HCI, interviews have been utilized for collecting users' subjective evaluations of interactive technology. This paper raises the issue that these interviews are often deployed in a manner overlooking two aspects of evaluation: the relative positions from which the system is evaluated and the interviewees' interpretations of the system. In the study, 14 users of a new information system were asked to evaluate provocative claims about the system's usability. The analyses of their responses reveal two sources of variation: what is being evaluated and who is evaluating it. Interviewees evaluated the system's usability from five user positions: end user, supervisor, organization's representative, co-developer, and outsider. Also, four "faces" of the system were interpreted: UI, utility, communication medium, and unknown entity. These findings are employed for drawing of broader conclusions about the system and its use, and procedures for improving user interviews in HCI are presented.