{"title":"自适应分布式系统的基准测试与评估支持","authors":"A. Vilenica, W. Lamersdorf","doi":"10.1109/CISIS.2012.115","DOIUrl":null,"url":null,"abstract":"Increasingly, distributed systems have to deal with highly dynamic and hardly predictable environments. This trend, in conjunction with rising demands for sophisticated non-functional system requirements, challenges both the development and operation (i.e. management) of traditional distributed systems. One promising approach to cope with these challenges are self-adaptive distributed systems that are characterized by the capability to configure and maintain themselves. However, the inherent dynamic of self-adaptive systems requires intensive evaluation and benchmarking efforts in order to ensure the intended system behaviour. In order to support that, this paper presents a framework that aims at supporting the nominal-actual comparison of self-adaptive distributed systems as well as the comparison of different self-adaptive solutions with respect to a specific software implementation task. The underlying approach consists of (i) a declarative definition language and (ii) a software component that is capable of conducting evaluations and benchmarks on different software implementations.","PeriodicalId":158978,"journal":{"name":"2012 Sixth International Conference on Complex, Intelligent, and Software Intensive Systems","volume":"29 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2012-07-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"5","resultStr":"{\"title\":\"Benchmarking and Evaluation Support for Self-Adaptive Distributed Systems\",\"authors\":\"A. Vilenica, W. Lamersdorf\",\"doi\":\"10.1109/CISIS.2012.115\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Increasingly, distributed systems have to deal with highly dynamic and hardly predictable environments. This trend, in conjunction with rising demands for sophisticated non-functional system requirements, challenges both the development and operation (i.e. management) of traditional distributed systems. One promising approach to cope with these challenges are self-adaptive distributed systems that are characterized by the capability to configure and maintain themselves. However, the inherent dynamic of self-adaptive systems requires intensive evaluation and benchmarking efforts in order to ensure the intended system behaviour. In order to support that, this paper presents a framework that aims at supporting the nominal-actual comparison of self-adaptive distributed systems as well as the comparison of different self-adaptive solutions with respect to a specific software implementation task. The underlying approach consists of (i) a declarative definition language and (ii) a software component that is capable of conducting evaluations and benchmarks on different software implementations.\",\"PeriodicalId\":158978,\"journal\":{\"name\":\"2012 Sixth International Conference on Complex, Intelligent, and Software Intensive Systems\",\"volume\":\"29 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2012-07-04\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"5\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2012 Sixth International Conference on Complex, Intelligent, and Software Intensive Systems\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/CISIS.2012.115\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2012 Sixth International Conference on Complex, Intelligent, and Software Intensive Systems","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CISIS.2012.115","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Benchmarking and Evaluation Support for Self-Adaptive Distributed Systems
Increasingly, distributed systems have to deal with highly dynamic and hardly predictable environments. This trend, in conjunction with rising demands for sophisticated non-functional system requirements, challenges both the development and operation (i.e. management) of traditional distributed systems. One promising approach to cope with these challenges are self-adaptive distributed systems that are characterized by the capability to configure and maintain themselves. However, the inherent dynamic of self-adaptive systems requires intensive evaluation and benchmarking efforts in order to ensure the intended system behaviour. In order to support that, this paper presents a framework that aims at supporting the nominal-actual comparison of self-adaptive distributed systems as well as the comparison of different self-adaptive solutions with respect to a specific software implementation task. The underlying approach consists of (i) a declarative definition language and (ii) a software component that is capable of conducting evaluations and benchmarks on different software implementations.