{"title":"Enhancing competitiveness via a public fault and failure data repository","authors":"D. Wallace","doi":"10.1109/HASE.1998.731610","DOIUrl":"https://doi.org/10.1109/HASE.1998.731610","url":null,"abstract":"Today most software companies are under pressure to get reliable software products built quickly. To assist both industry and researchers, the Information Technology Laboratory (ITL) at NIST in the US, has initiated the Reference Data: Software Error, Fault, Failure Data Collection & Analysis Repository Project. Usually referred to as the EFF Project, the project treats the term \"error\" as the human action that produces the incorrect result; the term \"fault\" as the manifestation of an error in an artifact, and the term \"failure\" as the result of a fault that has been activated, or triggered, during operation of the software. The EFF project is collecting and analyzing data from either the development and maintenance processes or the operation of a delivered computer system. The data, with all proprietary information removed, will become available through a publicly accessible World Wide Web facility at NIST. Researchers may use the data for their experimentation. The information technology industry and researchers may use the resulting reference data to assist in many types of decisions for building better end user products and for developing new methods and tools to support industry. NIST encourages companies to consider the benefits of a public database. NIST will accept new or existing private data to augment the repository. Such data may be from either the software development or maintenance process, or may consist of failure data derived from systems already in service.","PeriodicalId":340424,"journal":{"name":"Proceedings Third IEEE International High-Assurance Systems Engineering Symposium (Cat. No.98EX231)","volume":"42 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1998-11-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124742609","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Good enough versus high assurance software testing and analysis methods","authors":"W. Howden","doi":"10.1109/HASE.1998.731609","DOIUrl":"https://doi.org/10.1109/HASE.1998.731609","url":null,"abstract":"Good enough evaluation methods are reviewed, including low volume statistical testing. High assurance software is viewed as requiring high volume statistical testing, or analysis methods that prove that a program has certain desired properties. The weaknesses of some previous high assurance strategies are discussed, and a new approach called pattern directed testing and analysis is described. In this approach, design patterns are used to give high assurance for critical generic properties, and enough assurance for problem dependent functionality.","PeriodicalId":340424,"journal":{"name":"Proceedings Third IEEE International High-Assurance Systems Engineering Symposium (Cat. No.98EX231)","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1998-11-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122004785","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Error and failure analysis of a UNIX server","authors":"Ronjeet Lal, G. Choi","doi":"10.1109/HASE.1998.731618","DOIUrl":"https://doi.org/10.1109/HASE.1998.731618","url":null,"abstract":"This paper presents a measurement-based dependability study of a UNIX server. The event logs of a UNIX server are collected to form the dependability data basis. Message logs spanning approximately eleven months were collected for this study. The event log data are classified and categorized to calculate parameters such as MTBF and availability. Component analysis is also performed to identify modules that are prone to errors in the system. Next, the system error activity proceeding each system failure is analyzed to identify error patterns that may be precursors of the observed failure events. Lastly, the error/failure results from the measurement are reviewed in the perspective of the fault/error assumptions made in several popular fault injection studies.","PeriodicalId":340424,"journal":{"name":"Proceedings Third IEEE International High-Assurance Systems Engineering Symposium (Cat. No.98EX231)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1998-11-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129496309","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"System survivability through security bug tolerance","authors":"C. Pu, Crispin Cowan","doi":"10.1109/HASE.1998.731601","DOIUrl":"https://doi.org/10.1109/HASE.1998.731601","url":null,"abstract":"Summary form only given. A traditional approach to system security is the construction of entirely new software that satisfy well defined security properties. However, the market pressure towards features seems to make such approach infeasible in the near term. Increasingly, commercial off-the-shelf software such as Windows NT is being used in mission critical information infrastructures. Therefore, alternative means must be found to protect large commercial software from attack. We present the security bug tolerance approach as an alternative to building highly secure software from scratch. Security bug tolerance accepts the idea that critical system software will contain vulnerabilities in the form of exploitable bugs, and seeks effective means to prevent these bugs from being exploited efficiently by attackers. We present a categorization scheme for security bug tolerance techniques, and populate it with techniques of our own and from the literature. The categorization is powerful enough to analyze and compare the similarities and differences of relatively diverse techniques such as firewalls, program type checking, and \"security through obscurity\". The goal of security bug tolerance techniques is system survivability (e.g., graceful degradation of system functionality or performance in the presence of attacks) in contrast to complete system security. Instead of ferreting out all potential vulnerabilities using the broad array of accepted techniques, such as formal methods and fault injection, we study means to stop and slow down attacks when they occur.","PeriodicalId":340424,"journal":{"name":"Proceedings Third IEEE International High-Assurance Systems Engineering Symposium (Cat. No.98EX231)","volume":"81 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1998-11-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128270126","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Assurance of transactions in allied heterogeneous systems by autonomous decentralized system techniques","authors":"Hideo Hanamura, K. Mori","doi":"10.1109/HASE.1998.731630","DOIUrl":"https://doi.org/10.1109/HASE.1998.731630","url":null,"abstract":"The autonomous decentralized system (DAS) has been designed for high flexibility to realize the on-line properties of on-line maintenance, on-line expansion and fault tolerance. However, users needs are increasingly varied and quickly change today. To satisfy those needs a homogeneous system such as a conventional ADS has limits in its capability to provide adequate services through heterogeneous application systems which are consist of different properties and quality data. Thus the integration of heterogeneous systems are required. This paper presents the technique for the integration of the heterogeneous systems in ADS based on DF architecture and discusses atomic actions on the DF, then shows the effectiveness of the allying techniques in implementation.","PeriodicalId":340424,"journal":{"name":"Proceedings Third IEEE International High-Assurance Systems Engineering Symposium (Cat. No.98EX231)","volume":"65 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1998-11-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122235199","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Estimating the number of residual defects [in software]","authors":"Y. Malaiya, J. Denton","doi":"10.1109/HASE.1998.731600","DOIUrl":"https://doi.org/10.1109/HASE.1998.731600","url":null,"abstract":"The number of residual defects is one of the most important factors that allows one to decide if a piece of software is ready to be released. In theory, one can find all the defects and count them. However, it is impossible to find all the defects within a reasonable amount of time. Estimating the defect density can become difficult for high-reliability software, since the remaining defects can be extremely hard to test for. One possible way is to apply the exponential software reliablility growth model (SRGM), and thus estimate the total number of defects present at the beginning of testing. In this paper, we show the problems with this approach and present a new approach based on software test coverage. Test coverage directly measures the thoroughness of testing, avoiding the problem of variations of test effectiveness. We apply this model to actual test data in order to project the residual number of defects. This method results in estimates that are more stable than the existing methods. The method is also easier to understand, and the convergence to the estimate can be observed visually.","PeriodicalId":340424,"journal":{"name":"Proceedings Third IEEE International High-Assurance Systems Engineering Symposium (Cat. No.98EX231)","volume":"1 3","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1998-11-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131775268","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A two-version approach for real-time concurrency control and recovery","authors":"Tei-Wei Kuo, Yuan-Ting Kao, Lihchyun Shu","doi":"10.1109/HASE.1998.731628","DOIUrl":"https://doi.org/10.1109/HASE.1998.731628","url":null,"abstract":"This paper proposes an integrated mechanism for concurrency control and recovery in real-time databases. A two-version approach is used to adjust the serializability of schedules in favor of higher-priority transactions, and, at the same time, to provide a consistent database image at client sites for local processing of read-only transactions. When the database crashes, a consistent database image is used, in return, to efficiently and predictably recover the database from failure. A two-version network-server-based architecture is proposed to support a diskless run-time logging mechanism and an effective write-through procedure.","PeriodicalId":340424,"journal":{"name":"Proceedings Third IEEE International High-Assurance Systems Engineering Symposium (Cat. No.98EX231)","volume":"18 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1998-11-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123861177","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Log-logistic software reliability growth model","authors":"S. Gokhale, Kishor S. Trivedi","doi":"10.1109/HASE.1998.731593","DOIUrl":"https://doi.org/10.1109/HASE.1998.731593","url":null,"abstract":"The finite-failure non-homogeneous Poisson process (NHPP) models proposed in the literature exhibit either constant, monotonic increasing or monotonic decreasing failure occurrence rates per fault, and are inadequate to describe the failure processes underlying certain failure data sets. In this paper, we propose the log-logistic reliability growth model, which can capture the increasing/decreasing nature of the failure occurrence rate per fault. Equations are developed to estimate the parameters of the existing finite-failure NHPP models, as well as the log-logistic model, based on failure data collected in the form of inter-failure times. We also present an analysis of two data sets, where the underlying failure process could not be adequately described by the existing models, which motivated the development of the log-logistic model.","PeriodicalId":340424,"journal":{"name":"Proceedings Third IEEE International High-Assurance Systems Engineering Symposium (Cat. No.98EX231)","volume":"33 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1998-11-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114282092","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
M. Pizza, L. Strigini, A. Bondavalli, F. Giandomenico
{"title":"Optimal discrimination between transient and permanent faults","authors":"M. Pizza, L. Strigini, A. Bondavalli, F. Giandomenico","doi":"10.1109/HASE.1998.731615","DOIUrl":"https://doi.org/10.1109/HASE.1998.731615","url":null,"abstract":"An important practical problem in fault diagnosis is discriminating between permanent faults and transient faults. In many computer systems, the majority of errors are due to transient faults. Many heuristic methods have been used for discriminating between transient and permanent faults; however, we have found no previous work stating this decision problem in clear probabilistic terms. We present an optimal procedure for discriminating between transient and permanent faults, based on applying Bayesian inference to the observed events (correct and erroneous results). We describe how the assessed probability that a module is permanently faulty must vary with observed symptoms. We describe and demonstrate our proposed method on a simple application problem, building the appropriate equations and showing numerical examples. The method can be implemented as a run-time diagnosis algorithm at little computational cost; it can also be used to evaluate any heuristic diagnostic procedure by comparison.","PeriodicalId":340424,"journal":{"name":"Proceedings Third IEEE International High-Assurance Systems Engineering Symposium (Cat. No.98EX231)","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1998-11-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129955162","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Predictable safety in the control of high consequence systems","authors":"J. M. Covan, J. A. Cooper","doi":"10.1109/HASE.1998.731612","DOIUrl":"https://doi.org/10.1109/HASE.1998.731612","url":null,"abstract":"Many industries transmit large amounts of energy under the control of safety critical systems. Inadvertent release of energy by such systems can result in negative high consequences. This paper describes a principle-based strategy for preventing inadvertent release due to normal operational stresses or abnormal (e.g., accident) stresses. The safety principles, developed by Sandia National Laboratories for imbedding detonation safety in nuclear weapons, include isolation, inoperability and incompatibility. These principles are defined in the paper. They are illustrated and contrasted to conventional practice via the application to a gas furnace control system.","PeriodicalId":340424,"journal":{"name":"Proceedings Third IEEE International High-Assurance Systems Engineering Symposium (Cat. No.98EX231)","volume":"9 6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1998-11-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116444655","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}