nlrpBENCH: A Benchmark for Natural Language Requirements Processing

W. Tichy, Mathias Landhäußer, Sven J. Körner
{"title":"nlrpBENCH: A Benchmark for Natural Language Requirements Processing","authors":"W. Tichy, Mathias Landhäußer, Sven J. Körner","doi":"10.5445/IR/1000041252","DOIUrl":null,"url":null,"abstract":"We present nlrpBENCH: a new platform and framework to improve soft- \nware engineering research as well as teaching with focus on requirements engineering \nduring the software engineering process. It is available on http://nlrp.ipd. \nkit.edu. \nRecent advances in natural language processing have made it possible to process \ntextual software requirements automatically, for example checking them for flaws or \ntranslating them into software artifacts. This development is particularly fortunate, \nas the majority of requirements is written in unrestricted natural language. However, \nmany of the tools in in this young area of research have been evaluated only on limited \nsets of examples, because there is no accepted benchmark that could be used to assess \nand compare these tools. To improve comparability and thereby accelerate progress, \nwe have begun to assemble nlrpBENCH, a collection of requirements specifications \nmeant both as a challenge for tools and a yardstick for comparison. \nWe have gathered over 50 requirement texts of varying length and difficulty and \norganized them in benchmark sets. At present, there are two task types: model extrac- \ntion (e.g., generating UML models) and text correction (e.g., eliminating ambiguities). \nEach text is accompanied by the expected result and metrics for scoring results. This \npaper describes the composition of the benchmark and the sources. Due to the brevity \nof this paper, we omit example tools comparisons which are also available.","PeriodicalId":176893,"journal":{"name":"Software Engineering & Management","volume":"14 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"16","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Software Engineering & Management","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.5445/IR/1000041252","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 16

Abstract

We present nlrpBENCH: a new platform and framework to improve soft- ware engineering research as well as teaching with focus on requirements engineering during the software engineering process. It is available on http://nlrp.ipd. kit.edu. Recent advances in natural language processing have made it possible to process textual software requirements automatically, for example checking them for flaws or translating them into software artifacts. This development is particularly fortunate, as the majority of requirements is written in unrestricted natural language. However, many of the tools in in this young area of research have been evaluated only on limited sets of examples, because there is no accepted benchmark that could be used to assess and compare these tools. To improve comparability and thereby accelerate progress, we have begun to assemble nlrpBENCH, a collection of requirements specifications meant both as a challenge for tools and a yardstick for comparison. We have gathered over 50 requirement texts of varying length and difficulty and organized them in benchmark sets. At present, there are two task types: model extrac- tion (e.g., generating UML models) and text correction (e.g., eliminating ambiguities). Each text is accompanied by the expected result and metrics for scoring results. This paper describes the composition of the benchmark and the sources. Due to the brevity of this paper, we omit example tools comparisons which are also available.
nlrpBENCH:自然语言需求处理的基准
我们提出了nlrpBENCH:一个新的平台和框架,以改善软件工程研究和教学,重点是在软件工程过程中的需求工程。可以在http://nlrp.ipd上找到。kit.edu。自然语言处理的最新进展使得自动处理文本软件需求成为可能,例如检查它们的缺陷或将它们翻译成软件工件。这种开发是特别幸运的,因为大多数需求是用不受限制的自然语言编写的。然而,在这个年轻的研究领域中,许多工具只在有限的例子集上进行了评估,因为没有公认的基准可以用来评估和比较这些工具。为了提高可比性并因此加速进展,我们已经开始组装nlrpBENCH,这是一个需求规范的集合,它既是对工具的挑战,也是对比较的标准。我们收集了50多个不同长度和难度的需求文本,并将它们组织在基准集中。目前,有两种任务类型:模型提取(例如,生成UML模型)和文本更正(例如,消除歧义)。每个文本都附有预期结果和评分结果的度量标准。本文介绍了基准的组成和来源。由于本文的简短性,我们省略了同样可用的示例工具比较。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信