{"title":"使用符号动画和机器学习的基于模型的测试","authors":"Pierre-Christophe Bué, Frédéric Dadeau, Pierre-Cyrille Héam","doi":"10.1109/ICSTW.2010.43","DOIUrl":null,"url":null,"abstract":"We present in this paper a technique based on symbolic animation of models that aims at producing model-based tests. In order to guide the animation of the model, we rely on the use of a deterministic finite automaton (DFA) of the model that is built using a well-known machine learning algorithm, that considers a complex model as a black-box component, whose behavior is inferred. Since the DFA obtained in this way may be an over-approximation and, thus, admit traces that were not admitted on the original model, this abstraction is refined using counter-examples made of unfeasible traces. The computation of counter-examples is performed using a systematic coverage of the DFA states and transitions, producing test sequences that are replayed on the model, providing either test cases for offline testing, or counter-examples that aim at refining the abstraction.","PeriodicalId":117410,"journal":{"name":"2010 Third International Conference on Software Testing, Verification, and Validation Workshops","volume":"4020 2 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2010-04-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":"{\"title\":\"Model-Based Testing Using Symbolic Animation and Machine Learning\",\"authors\":\"Pierre-Christophe Bué, Frédéric Dadeau, Pierre-Cyrille Héam\",\"doi\":\"10.1109/ICSTW.2010.43\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"We present in this paper a technique based on symbolic animation of models that aims at producing model-based tests. In order to guide the animation of the model, we rely on the use of a deterministic finite automaton (DFA) of the model that is built using a well-known machine learning algorithm, that considers a complex model as a black-box component, whose behavior is inferred. Since the DFA obtained in this way may be an over-approximation and, thus, admit traces that were not admitted on the original model, this abstraction is refined using counter-examples made of unfeasible traces. The computation of counter-examples is performed using a systematic coverage of the DFA states and transitions, producing test sequences that are replayed on the model, providing either test cases for offline testing, or counter-examples that aim at refining the abstraction.\",\"PeriodicalId\":117410,\"journal\":{\"name\":\"2010 Third International Conference on Software Testing, Verification, and Validation Workshops\",\"volume\":\"4020 2 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2010-04-06\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"2\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2010 Third International Conference on Software Testing, Verification, and Validation Workshops\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICSTW.2010.43\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2010 Third International Conference on Software Testing, Verification, and Validation Workshops","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICSTW.2010.43","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Model-Based Testing Using Symbolic Animation and Machine Learning
We present in this paper a technique based on symbolic animation of models that aims at producing model-based tests. In order to guide the animation of the model, we rely on the use of a deterministic finite automaton (DFA) of the model that is built using a well-known machine learning algorithm, that considers a complex model as a black-box component, whose behavior is inferred. Since the DFA obtained in this way may be an over-approximation and, thus, admit traces that were not admitted on the original model, this abstraction is refined using counter-examples made of unfeasible traces. The computation of counter-examples is performed using a systematic coverage of the DFA states and transitions, producing test sequences that are replayed on the model, providing either test cases for offline testing, or counter-examples that aim at refining the abstraction.