Fatemeh Esfahani, Mahsa Daneshmand, Venkatesh Srinivasan, Alex Thomo, Kui Wu
{"title":"基于中心极限定理和h指数的可伸缩概率桁架分解。","authors":"Fatemeh Esfahani, Mahsa Daneshmand, Venkatesh Srinivasan, Alex Thomo, Kui Wu","doi":"10.1007/s10619-022-07415-9","DOIUrl":null,"url":null,"abstract":"<p><p>Truss decomposition is a popular notion of hierarchical dense substructures in graphs. In a nutshell, <i>k</i>-truss is the largest subgraph in which every edge is contained in at least <i>k</i> triangles. Truss decomposition aims to compute <i>k</i>-trusses for each possible value of <i>k</i>. There are many works that study truss decomposition in deterministic graphs. However, in probabilistic graphs, truss decomposition is significantly more challenging and has received much less attention; state-of-the-art approaches do not scale well to large probabilistic graphs. Finding the tail probabilities of the number of triangles that contain each edge is a critical challenge of those approaches. This is achieved using dynamic programming which has quadratic run-time and thus not scalable to real large networks which, quite commonly, can have edges contained in many triangles (in the millions). To address this challenge, we employ a special version of the Central Limit Theorem (CLT) to obtain the tail probabilities efficiently. Based on our CLT approach we propose a peeling algorithm for truss decomposition that scales to large probabilistic graphs and offers significant improvement over state-of-the-art. We also design a second method which progressively tightens the estimate of the truss value of each edge and is based on <i>h</i>-index computation. In contrast to our CLT-based approach, our <i>h</i>-index algorithm (1) is progressive by allowing the user to see near-results along the way, (2) does not sacrifice the exactness of final result, and (3) achieves all these while processing only one edge and its immediate neighbors at a time, thus resulting in smaller memory footprint. We perform extensive experiments to show the scalability of both of our proposed algorithms.</p>","PeriodicalId":50568,"journal":{"name":"Distributed and Parallel Databases","volume":" ","pages":"299-333"},"PeriodicalIF":0.9000,"publicationDate":"2022-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9310023/pdf/","citationCount":"1","resultStr":"{\"title\":\"Scalable probabilistic truss decomposition using central limit theorem and H-index.\",\"authors\":\"Fatemeh Esfahani, Mahsa Daneshmand, Venkatesh Srinivasan, Alex Thomo, Kui Wu\",\"doi\":\"10.1007/s10619-022-07415-9\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><p>Truss decomposition is a popular notion of hierarchical dense substructures in graphs. In a nutshell, <i>k</i>-truss is the largest subgraph in which every edge is contained in at least <i>k</i> triangles. Truss decomposition aims to compute <i>k</i>-trusses for each possible value of <i>k</i>. There are many works that study truss decomposition in deterministic graphs. However, in probabilistic graphs, truss decomposition is significantly more challenging and has received much less attention; state-of-the-art approaches do not scale well to large probabilistic graphs. Finding the tail probabilities of the number of triangles that contain each edge is a critical challenge of those approaches. This is achieved using dynamic programming which has quadratic run-time and thus not scalable to real large networks which, quite commonly, can have edges contained in many triangles (in the millions). To address this challenge, we employ a special version of the Central Limit Theorem (CLT) to obtain the tail probabilities efficiently. Based on our CLT approach we propose a peeling algorithm for truss decomposition that scales to large probabilistic graphs and offers significant improvement over state-of-the-art. We also design a second method which progressively tightens the estimate of the truss value of each edge and is based on <i>h</i>-index computation. In contrast to our CLT-based approach, our <i>h</i>-index algorithm (1) is progressive by allowing the user to see near-results along the way, (2) does not sacrifice the exactness of final result, and (3) achieves all these while processing only one edge and its immediate neighbors at a time, thus resulting in smaller memory footprint. We perform extensive experiments to show the scalability of both of our proposed algorithms.</p>\",\"PeriodicalId\":50568,\"journal\":{\"name\":\"Distributed and Parallel Databases\",\"volume\":\" \",\"pages\":\"299-333\"},\"PeriodicalIF\":0.9000,\"publicationDate\":\"2022-01-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9310023/pdf/\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Distributed and Parallel Databases\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://doi.org/10.1007/s10619-022-07415-9\",\"RegionNum\":4,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"2022/7/25 0:00:00\",\"PubModel\":\"Epub\",\"JCR\":\"Q3\",\"JCRName\":\"COMPUTER SCIENCE, INFORMATION SYSTEMS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Distributed and Parallel Databases","FirstCategoryId":"94","ListUrlMain":"https://doi.org/10.1007/s10619-022-07415-9","RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2022/7/25 0:00:00","PubModel":"Epub","JCR":"Q3","JCRName":"COMPUTER SCIENCE, INFORMATION SYSTEMS","Score":null,"Total":0}
Scalable probabilistic truss decomposition using central limit theorem and H-index.
Truss decomposition is a popular notion of hierarchical dense substructures in graphs. In a nutshell, k-truss is the largest subgraph in which every edge is contained in at least k triangles. Truss decomposition aims to compute k-trusses for each possible value of k. There are many works that study truss decomposition in deterministic graphs. However, in probabilistic graphs, truss decomposition is significantly more challenging and has received much less attention; state-of-the-art approaches do not scale well to large probabilistic graphs. Finding the tail probabilities of the number of triangles that contain each edge is a critical challenge of those approaches. This is achieved using dynamic programming which has quadratic run-time and thus not scalable to real large networks which, quite commonly, can have edges contained in many triangles (in the millions). To address this challenge, we employ a special version of the Central Limit Theorem (CLT) to obtain the tail probabilities efficiently. Based on our CLT approach we propose a peeling algorithm for truss decomposition that scales to large probabilistic graphs and offers significant improvement over state-of-the-art. We also design a second method which progressively tightens the estimate of the truss value of each edge and is based on h-index computation. In contrast to our CLT-based approach, our h-index algorithm (1) is progressive by allowing the user to see near-results along the way, (2) does not sacrifice the exactness of final result, and (3) achieves all these while processing only one edge and its immediate neighbors at a time, thus resulting in smaller memory footprint. We perform extensive experiments to show the scalability of both of our proposed algorithms.
期刊介绍:
Distributed and Parallel Databases publishes papers in all the traditional as well as most emerging areas of database research, including:
Availability and reliability;
Benchmarking and performance evaluation, and tuning;
Big Data Storage and Processing;
Cloud Computing and Database-as-a-Service;
Crowdsourcing;
Data curation, annotation and provenance;
Data integration, metadata Management, and interoperability;
Data models, semantics, query languages;
Data mining and knowledge discovery;
Data privacy, security, trust;
Data provenance, workflows, Scientific Data Management;
Data visualization and interactive data exploration;
Data warehousing, OLAP, Analytics;
Graph data management, RDF, social networks;
Information Extraction and Data Cleaning;
Middleware and Workflow Management;
Modern Hardware and In-Memory Database Systems;
Query Processing and Optimization;
Semantic Web and open data;
Social Networks;
Storage, indexing, and physical database design;
Streams, sensor networks, and complex event processing;
Strings, Texts, and Keyword Search;
Spatial, temporal, and spatio-temporal databases;
Transaction processing;
Uncertain, probabilistic, and approximate databases.