Testing software's changing features with environment-driven abstraction identification.

IF 2.1 3区 计算机科学 Q3 COMPUTER SCIENCE, INFORMATION SYSTEMS
Requirements Engineering Pub Date : 2022-01-01 Epub Date: 2022-09-20 DOI:10.1007/s00766-022-00390-8
Zedong Peng, Prachi Rathod, Nan Niu, Tanmay Bhowmik, Hui Liu, Lin Shi, Zhi Jin
{"title":"Testing software's changing features with environment-driven abstraction identification.","authors":"Zedong Peng,&nbsp;Prachi Rathod,&nbsp;Nan Niu,&nbsp;Tanmay Bhowmik,&nbsp;Hui Liu,&nbsp;Lin Shi,&nbsp;Zhi Jin","doi":"10.1007/s00766-022-00390-8","DOIUrl":null,"url":null,"abstract":"<p><p>ions are significant domain terms that have assisted in requirements elicitation and modeling. To extend the assistance toward requirements validation, we present in this paper an automated approach to identifying the abstractions for supporting requirements-based testing. We select relevant Wikipedia pages to serve as a domain corpus that is independent from any specific software system. We further define five novel patterns based on part-of-speech tagging and dependency parsing, and frame our candidate abstractions in the form of <key, value> pairs for better testability, where the \"key\" helps locate \"what to test\", and the \"value\" helps guide \"how to test it\" by feeding in concrete data. We evaluate our approach with six software systems in two application domains: Electronic health records and Web conferencing. The results show that our abstractions are more accurate than those generated by a state-of-the-art technique. While the initial findings indicate our abstractions' capabilities of revealing bugs and matching the environmental assumptions created manually, we articulate a new way to perform requirements-based testing by focusing on a software system's changing features. Specifically, we hypothesize that the same feature would behave differently under a pair of opposing environmental conditions and assess our abstractions' applicability to this new form of feature testing.</p>","PeriodicalId":20912,"journal":{"name":"Requirements Engineering","volume":"27 4","pages":"405-427"},"PeriodicalIF":2.1000,"publicationDate":"2022-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9486799/pdf/","citationCount":"4","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Requirements Engineering","FirstCategoryId":"94","ListUrlMain":"https://doi.org/10.1007/s00766-022-00390-8","RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2022/9/20 0:00:00","PubModel":"Epub","JCR":"Q3","JCRName":"COMPUTER SCIENCE, INFORMATION SYSTEMS","Score":null,"Total":0}
引用次数: 4

Abstract

ions are significant domain terms that have assisted in requirements elicitation and modeling. To extend the assistance toward requirements validation, we present in this paper an automated approach to identifying the abstractions for supporting requirements-based testing. We select relevant Wikipedia pages to serve as a domain corpus that is independent from any specific software system. We further define five novel patterns based on part-of-speech tagging and dependency parsing, and frame our candidate abstractions in the form of pairs for better testability, where the "key" helps locate "what to test", and the "value" helps guide "how to test it" by feeding in concrete data. We evaluate our approach with six software systems in two application domains: Electronic health records and Web conferencing. The results show that our abstractions are more accurate than those generated by a state-of-the-art technique. While the initial findings indicate our abstractions' capabilities of revealing bugs and matching the environmental assumptions created manually, we articulate a new way to perform requirements-based testing by focusing on a software system's changing features. Specifically, we hypothesize that the same feature would behave differently under a pair of opposing environmental conditions and assess our abstractions' applicability to this new form of feature testing.

Abstract Image

Abstract Image

Abstract Image

用环境驱动的抽象识别来测试软件不断变化的特性。
离子是重要的领域术语,有助于需求引出和建模。为了扩展对需求验证的帮助,我们在本文中提出了一种自动化的方法来识别支持基于需求的测试的抽象。我们选择相关的维基百科页面作为独立于任何特定软件系统的领域语料库。我们进一步定义了五种基于词性标记和依赖解析的新模式,并以对的形式构建我们的候选抽象,以获得更好的可测试性,其中“键”帮助定位“要测试什么”,“值”通过提供具体数据帮助指导“如何测试”。我们用两个应用领域的六个软件系统来评估我们的方法:电子健康记录和网络会议。结果表明,我们的抽象比最先进的技术生成的抽象更准确。虽然最初的发现表明我们的抽象具有揭示缺陷和匹配手动创建的环境假设的能力,但我们通过关注软件系统不断变化的特性,阐明了一种执行基于需求的测试的新方法。具体来说,我们假设相同的功能在一对相反的环境条件下会表现不同,并评估我们的抽象对这种新形式的功能测试的适用性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Requirements Engineering
Requirements Engineering 工程技术-计算机:软件工程
CiteScore
7.10
自引率
10.70%
发文量
27
审稿时长
>12 weeks
期刊介绍: The journal provides a focus for the dissemination of new results about the elicitation, representation and validation of requirements of software intensive information systems or applications. Theoretical and applied submissions are welcome, but all papers must explicitly address: -the practical consequences of the ideas for the design of complex systems -how the ideas should be evaluated by the reflective practitioner The journal is motivated by a multi-disciplinary view that considers requirements not only in terms of software components specification but also in terms of activities for their elicitation, representation and agreement, carried out within an organisational and social context. To this end, contributions are sought from fields such as software engineering, information systems, occupational sociology, cognitive and organisational psychology, human-computer interaction, computer-supported cooperative work, linguistics and philosophy for work addressing specifically requirements engineering issues.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信