从用户撰写的场景和词嵌入中提取领域模型

Yuchen Shen, T. Breaux
{"title":"从用户撰写的场景和词嵌入中提取领域模型","authors":"Yuchen Shen, T. Breaux","doi":"10.1109/REW56159.2022.00036","DOIUrl":null,"url":null,"abstract":"Domain models are used by requirements analysts to rationalize domain phenomena into discrete entities that drive requirements elicitation and analysis. Domain models include entities, actors or agents, their actions, and desired qualities assigned to states in the domain. Domain models are acquired through a wide range of sources, including interviews with subject matter experts, and by analyzing text-based scenarios, regulations and policies. Requirements automation to assist with elicitation or text analysis can be supported using masked language models (MLM), which have been used to learn contextual information from natural language sentences and transfer this learning to natural language processing (NLP) tasks. The MLM can be used to predict the most likely missing word in a sentence, and thus be used to explore domain concepts encoded in a word embedding. In this paper, we explore an approach of extracting domain knowledge from user-authored scenarios using typed dependency parsing techniques. We also explore the efficacy of a complementary approach of using a BERT-based MLM to identify entities and associated qualities to build a domain model from a single-word seed term.","PeriodicalId":360738,"journal":{"name":"2022 IEEE 30th International Requirements Engineering Conference Workshops (REW)","volume":"39 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"4","resultStr":"{\"title\":\"Domain Model Extraction from User-authored Scenarios and Word Embeddings\",\"authors\":\"Yuchen Shen, T. Breaux\",\"doi\":\"10.1109/REW56159.2022.00036\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Domain models are used by requirements analysts to rationalize domain phenomena into discrete entities that drive requirements elicitation and analysis. Domain models include entities, actors or agents, their actions, and desired qualities assigned to states in the domain. Domain models are acquired through a wide range of sources, including interviews with subject matter experts, and by analyzing text-based scenarios, regulations and policies. Requirements automation to assist with elicitation or text analysis can be supported using masked language models (MLM), which have been used to learn contextual information from natural language sentences and transfer this learning to natural language processing (NLP) tasks. The MLM can be used to predict the most likely missing word in a sentence, and thus be used to explore domain concepts encoded in a word embedding. In this paper, we explore an approach of extracting domain knowledge from user-authored scenarios using typed dependency parsing techniques. We also explore the efficacy of a complementary approach of using a BERT-based MLM to identify entities and associated qualities to build a domain model from a single-word seed term.\",\"PeriodicalId\":360738,\"journal\":{\"name\":\"2022 IEEE 30th International Requirements Engineering Conference Workshops (REW)\",\"volume\":\"39 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2022-08-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"4\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2022 IEEE 30th International Requirements Engineering Conference Workshops (REW)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/REW56159.2022.00036\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 IEEE 30th International Requirements Engineering Conference Workshops (REW)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/REW56159.2022.00036","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 4

摘要

需求分析人员使用领域模型来将领域现象合理化为驱动需求提取和分析的离散实体。领域模型包括实体、参与者或代理、它们的行为,以及分配给领域中状态的期望质量。领域模型是通过广泛的来源获得的,包括与主题专家的访谈,以及通过分析基于文本的场景、法规和政策。可以使用掩码语言模型(MLM)来支持需求自动化,以协助引出或文本分析,这些模型已被用于从自然语言句子中学习上下文信息,并将这种学习转移到自然语言处理(NLP)任务中。MLM可用于预测句子中最可能缺失的单词,从而用于探索单词嵌入中编码的领域概念。在本文中,我们探索了一种使用类型依赖解析技术从用户编写的场景中提取领域知识的方法。我们还探索了一种互补方法的有效性,即使用基于bert的MLM来识别实体和相关质量,从而从单个单词种子术语构建领域模型。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Domain Model Extraction from User-authored Scenarios and Word Embeddings
Domain models are used by requirements analysts to rationalize domain phenomena into discrete entities that drive requirements elicitation and analysis. Domain models include entities, actors or agents, their actions, and desired qualities assigned to states in the domain. Domain models are acquired through a wide range of sources, including interviews with subject matter experts, and by analyzing text-based scenarios, regulations and policies. Requirements automation to assist with elicitation or text analysis can be supported using masked language models (MLM), which have been used to learn contextual information from natural language sentences and transfer this learning to natural language processing (NLP) tasks. The MLM can be used to predict the most likely missing word in a sentence, and thus be used to explore domain concepts encoded in a word embedding. In this paper, we explore an approach of extracting domain knowledge from user-authored scenarios using typed dependency parsing techniques. We also explore the efficacy of a complementary approach of using a BERT-based MLM to identify entities and associated qualities to build a domain model from a single-word seed term.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信