{"title":"Atomic commitment in database systems over active networks","authors":"Zhili Zhang, W. Perrizo, V. Shi","doi":"10.1109/ICDE.1999.754906","DOIUrl":"https://doi.org/10.1109/ICDE.1999.754906","url":null,"abstract":"Traditional communication networks have functioned only to deliver information packets from one end system to another. As the cost of computing power continues to decrease, it is worthwhile to consider the benefits of adding more of that computing power to the network itself, either to enhance services or to trade off against other costs such as time, bandwidth and storage. This new area of research is called Active Networking. Active networking represents a new approach in which network switches or nodes perform customized computing as the messages flowing through them. This paper proposes a novel approach to atomic commitment of transactions in distributed database systems. In our approach, active nodes participate in the process of atomic commitment of distributed transactions by maintaining passive \"blackboards\" which record voting status of participant subtransactions. One active node is delegated the coordinator function so that when rendezvous is there, commit decisions can be made. Since the decision is made from an internal network node (selected to be close to the participant nodes), the uncertainty time delay can be significantly reduced. This is especially critical in wide area networks. Other active nodes can be used to suppress duplicated votes sent in the upstream direction. Additionally, once a negative vote is received at any active node, children participants can be notified of the abort decision immediately and therefore much earlier than would otherwise be possible. This approach is shown to yield performance improvements over traditional methods, in terms of delay and network message traffic.","PeriodicalId":236128,"journal":{"name":"Proceedings 15th International Conference on Data Engineering (Cat. No.99CB36337)","volume":"236 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1999-03-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116985967","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
A. Bouguettaya, B. Benatallah, M. Ouzzani, Lily Hendra
{"title":"Using Java and CORBA for implementing Internet databases","authors":"A. Bouguettaya, B. Benatallah, M. Ouzzani, Lily Hendra","doi":"10.1109/ICDE.1999.754927","DOIUrl":"https://doi.org/10.1109/ICDE.1999.754927","url":null,"abstract":"We describe an architecture called WebFINDIT that allows dynamic couplings of Web accessible databases based on their content and interest. We propose an implementation using WWW, Java, JDBC, and CORBA's ORBs that communicate via the CORBA's IIOP protocol. The combination of these technologies offers a compelling middleware infrastructure to implement fluid-area enterprise applications. In addition to a discussion of WebFINDIT's core concepts and implementation architecture, we also discuss an experience of exiting WebFINDIT in a healthcare application.","PeriodicalId":236128,"journal":{"name":"Proceedings 15th International Conference on Data Engineering (Cat. No.99CB36337)","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1999-03-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128153938","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Systematic multiresolution and its application to the World Wide Web","authors":"S. Acharya, H. F. Korth, V. Poosala","doi":"10.1109/ICDE.1999.754896","DOIUrl":"https://doi.org/10.1109/ICDE.1999.754896","url":null,"abstract":"Many emerging environments are increasingly facing the problem where the requirements of applications easily outstrip the system resources. This is particularly acute in the World Wide Web (WWW) and many data-intensive applications like OLAP and multimedia databases. We address this problem in the Web context via systematic multiresolution, i.e., a framework for providing responses at different qualities (resolutions) and costs. We validate our conceptual contributions by implementing NetBlitz a multiresolution-based proxy server on the WWW. NetBlitz addresses two key problems facing the Web: high latencies and heterogeneity of client resources and requirements. It solves these problems by dynamically generating the \"required version\" of a Web object based on client preferences and capabilities. We also propose novel multiresolution-aware caching techniques that further improve performance. Finally we experimentally demonstrate the utility of multiresolution and the caching enhancements proposed.","PeriodicalId":236128,"journal":{"name":"Proceedings 15th International Conference on Data Engineering (Cat. No.99CB36337)","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1999-03-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134532182","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Scalable classification over SQL databases","authors":"S. Chaudhuri, U. Fayyad, J. Bernhardt","doi":"10.1109/ICDE.1999.754963","DOIUrl":"https://doi.org/10.1109/ICDE.1999.754963","url":null,"abstract":"We identify data-intensive operations that are common to classifiers and develop a middleware that decomposes and schedules these operations efficiently using a backend SQL database. Our approach has the added advantage of not requiring any specialized physical data organization. We demonstrate the scalability characteristics of our enhanced client with experiments on Microsoft SQL Server 7.0 by varying data size, number of attributes and characteristics of decision trees.","PeriodicalId":236128,"journal":{"name":"Proceedings 15th International Conference on Data Engineering (Cat. No.99CB36337)","volume":"114 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1999-03-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124121457","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Efficient time series matching by wavelets","authors":"K. Chan, A. Fu","doi":"10.1109/ICDE.1999.754915","DOIUrl":"https://doi.org/10.1109/ICDE.1999.754915","url":null,"abstract":"Time series stored as feature vectors can be indexed by multidimensional index trees like R-Trees for fast retrieval. Due to the dimensionality curse problem, transformations are applied to time series to reduce the number of dimensions of the feature vectors. Different transformations like Discrete Fourier Transform (DFT) Discrete Wavelet Transform (DWT), Karhunen-Loeve (KL) transform or Singular Value Decomposition (SVD) can be applied. While the use of DFT and K-L transform or SVD have been studied on the literature, to our knowledge, there is no in-depth study on the application of DWT. In this paper we propose to use Haar Wavelet Transform for time series indexing. The major contributions are: (1) we show that Euclidean distance is preserved in the Haar transformed domain and no false dismissal will occur, (2) we show that Haar transform can outperform DFT through experiments, (3) a new similarity model is suggested to accommodate vertical shift of time series, and (4) a two-phase method is proposed for efficient n-nearest neighbor query in time series databases.","PeriodicalId":236128,"journal":{"name":"Proceedings 15th International Conference on Data Engineering (Cat. No.99CB36337)","volume":"70 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1999-03-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129322319","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"The ECHO method: concurrency control method for a large-scale distributed database","authors":"Y. Shirota, A. Iizawa, Hiroko Mano, T. Yano","doi":"10.1109/ICDE.1999.754922","DOIUrl":"https://doi.org/10.1109/ICDE.1999.754922","url":null,"abstract":"When constructing a large-scale database system, data replication is useful to give users efficient access to that data. However for databases with a very large number of replica sites, using traditional networking concurrency control protocols makes communication too costly for practical use. In this paper we propose a new protocol called ECHO for concurrency control of replica data distributed on a large scale. Using broadcasting, such as satellite broadcasting, terrestrial broadcasting, and cable television, the ECHO method can keep communication costs constant, regardless of the number of sites. The ECHO method also has a backup mechanism for failed broadcast reception using the communication network.","PeriodicalId":236128,"journal":{"name":"Proceedings 15th International Conference on Data Engineering (Cat. No.99CB36337)","volume":"83 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1999-03-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114617980","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Chengfei Liu, M. Orlowska, Xiaofang Zhou, Xuemin Lin
{"title":"Confirmation: a solution for non-compensatability in workflow applications","authors":"Chengfei Liu, M. Orlowska, Xiaofang Zhou, Xuemin Lin","doi":"10.1109/ICDE.1999.754910","DOIUrl":"https://doi.org/10.1109/ICDE.1999.754910","url":null,"abstract":"The notion of a compensation is widely used in advanced transaction models as means of recovery from a failure. Similar concepts are adopted for providing \"transaction-like\" behaviour for long business processes supported by workflows technology. Generally, designing a compensating task in the context of a workflow process is a non-trivial job. In fact, not every task is compensatable. This work contributes to the study of the non-compensatability problem. A compensating task C of a task T semantically undoes the effect of T after T has been committed. For example, the compensating task of a deposit is a withdrawal. For a task to be compensatable, it must satisfy two conditions. Forcibility: The compensating task of the task must be forcible. In other words, after the task commits, the execution of its compensating task is guaranteed to succeed by the application semantics. Relaxation of isolation: The isolation requirement of the shared data resources which the task may access must be relaxed. This relaxation is required as the purpose of introducing compensation is to avoid long-duration waiting, otherwise, compensation may become useless. In this work, we carefully investigate the properties of shared resources and tasks which may be performed on these resources. As all its invoked operations must be compensatable as well if a task is compensatable, we only discuss the compensatability of operations defined on shared resources.","PeriodicalId":236128,"journal":{"name":"Proceedings 15th International Conference on Data Engineering (Cat. No.99CB36337)","volume":"17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1999-03-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114858784","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Speeding up heterogeneous data access by converting and pushing down string comparisons","authors":"Weiye Zhang, P. Larson","doi":"10.1109/ICDE.1999.754941","DOIUrl":"https://doi.org/10.1109/ICDE.1999.754941","url":null,"abstract":"Pushing down predicates to an external data source is critical to the performance for querying heterogeneous data sources. However, predicate-containing string comparisons cannot be pushed down unchanged if the external data source uses a different collating sequence. We describe a table-driven technique for rewriting such predicates to account for the differences in collation. In addition to precise conversion, we also consider imprecise conversion.","PeriodicalId":236128,"journal":{"name":"Proceedings 15th International Conference on Data Engineering (Cat. No.99CB36337)","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1999-03-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114887698","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"ROCK: a robust clustering algorithm for categorical attributes","authors":"S. Guha, R. Rastogi, Kyuseok Shim","doi":"10.1109/ICDE.1999.754967","DOIUrl":"https://doi.org/10.1109/ICDE.1999.754967","url":null,"abstract":"We study clustering algorithms for data with Boolean and categorical attributes. We show that traditional clustering algorithms that use distances between points for clustering are not appropriate for Boolean and categorical attributes. Instead, we propose a novel concept of links to measure the similarity/proximity between a pair of data points. We develop a robust hierarchical clustering algorithm, ROCK, that employs links and not distances when merging clusters. Our methods naturally extend to non-metric similarity measures that are relevant in situations where a domain expert/similarity table is the only source of knowledge. In addition to presenting detailed complexity results for ROCK, we also conduct an experimental study with real-life as well as synthetic data sets. Our study shows that ROCK not only generates better quality clusters than traditional algorithms, but also exhibits good scalability properties.","PeriodicalId":236128,"journal":{"name":"Proceedings 15th International Conference on Data Engineering (Cat. No.99CB36337)","volume":"24 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1999-03-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125355941","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Scalable web server design for distributed data management","authors":"S. Baker, Bongki Moon","doi":"10.1109/ICDE.1999.754904","DOIUrl":"https://doi.org/10.1109/ICDE.1999.754904","url":null,"abstract":"With the explosive popularity of the internet and the world wide web (WWW), there is a rapidly growing need to provide unprecedented access to globally distributed data sources through the internet. Web accessibility will be an essential component of the services that future digital libraries should provide for clients. This need has created a strong demand for database access capability through the internet, and high performance scalable web servers. As most popular web sites are experiencing overload from an increasing number of users accessing the sites at the same time, it is desired that scalable web servers should adapt to the changing access characteristics and should be capable of handling a large number of concurrent requests simultaneously, with reasonable response times and minimal request drop rates.","PeriodicalId":236128,"journal":{"name":"Proceedings 15th International Conference on Data Engineering (Cat. No.99CB36337)","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1999-03-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126410623","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}