{"title":"分布式算法和lca中的采样和输出估计","authors":"Leonid Barenboim , Tzalik Maimon","doi":"10.1016/j.tcs.2025.115311","DOIUrl":null,"url":null,"abstract":"<div><div>We consider the distributed message-passing model and the Local Computational Algorithms (LCA) model. In both models a network is represented by an <em>n</em>-vertex graph <span><math><mi>G</mi><mo>=</mo><mo>(</mo><mi>V</mi><mo>,</mo><mi>E</mi><mo>)</mo></math></span>. We focus on labeling problems, such as vertex-coloring, edge-coloring, maximal independent set (MIS) and maximal matching. In the distributed model the vertices of <em>v</em> perform computations in parallel in order to compute their own solution for solving the problem for <em>G</em>. In contrast, in the LCA model probes are performed on certain vertices in order to compute their labels in a solution to a given problem. In this work we study the possibility of estimating a solution produced by an algorithm, much before the algorithm terminates. This estimation not only allows for size approximation of a solution, but also for early detection of failure in randomized algorithms. We do this such that a correcting procedure can be executed. To this end, we propose a sampling technique, in which the labels in the sampling are distributed proportionally to the distribution in the algorithm's output. However, the sampling running time is significantly smaller than that of the algorithm in hand.</div><div>We achieve the following results, in terms of the maximum degree Δ and the arboricity <em>a</em> of the input graph. The running time of our procedures is <span><math><mi>O</mi><mo>(</mo><mi>log</mi><mo></mo><mi>a</mi><mo>+</mo><mi>log</mi><mo></mo><mi>log</mi><mo></mo><mi>n</mi><mo>)</mo></math></span>, for sampling vertex-coloring, edge-coloring, maximal matching and MIS. This significantly improves upon previous sampling techniques, which incur additional dependency on the maximum degree Δ that can be much higher than the arboricity, as well as more significant dependency on <em>n</em>. Not only that, we also show that our technique extends naturally for the power graph <span><math><msup><mrow><mi>G</mi></mrow><mrow><mi>r</mi></mrow></msup></math></span> for any constant integer <span><math><mi>r</mi><mo>></mo><mn>1</mn></math></span> for the problems of MIS and coloring.</div><div>Our techniques for sampling in the distributed model provide a powerful and general tool for estimation in the LCA model. In this setting the goal is estimating the size of a solution to a given problem, by making as few vertex probes as possible. For the above-mentioned problems, we achieve estimations with probe complexity <span><math><msup><mrow><mi>d</mi></mrow><mrow><mi>O</mi><mo>(</mo><mi>log</mi><mo></mo><mi>a</mi><mo>+</mo><mi>log</mi><mo></mo><mi>log</mi><mo></mo><mi>n</mi><mo>)</mo></mrow></msup></math></span>, where <span><math><mi>d</mi><mo>=</mo><mi>m</mi><mi>i</mi><mi>n</mi><mo>(</mo><mi>Δ</mi><mo>,</mo><mi>a</mi><mo>⋅</mo><mi>p</mi><mi>o</mi><mi>l</mi><mi>y</mi><mo>(</mo><mi>log</mi><mo></mo><mo>(</mo><mi>n</mi><mo>)</mo><mo>)</mo></math></span>. Our results extend as well to power graphs for the coloring and MIS problems.</div></div>","PeriodicalId":49438,"journal":{"name":"Theoretical Computer Science","volume":"1046 ","pages":"Article 115311"},"PeriodicalIF":1.0000,"publicationDate":"2025-05-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Sampling and output estimation in distributed algorithms and LCAs\",\"authors\":\"Leonid Barenboim , Tzalik Maimon\",\"doi\":\"10.1016/j.tcs.2025.115311\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>We consider the distributed message-passing model and the Local Computational Algorithms (LCA) model. In both models a network is represented by an <em>n</em>-vertex graph <span><math><mi>G</mi><mo>=</mo><mo>(</mo><mi>V</mi><mo>,</mo><mi>E</mi><mo>)</mo></math></span>. We focus on labeling problems, such as vertex-coloring, edge-coloring, maximal independent set (MIS) and maximal matching. In the distributed model the vertices of <em>v</em> perform computations in parallel in order to compute their own solution for solving the problem for <em>G</em>. In contrast, in the LCA model probes are performed on certain vertices in order to compute their labels in a solution to a given problem. In this work we study the possibility of estimating a solution produced by an algorithm, much before the algorithm terminates. This estimation not only allows for size approximation of a solution, but also for early detection of failure in randomized algorithms. We do this such that a correcting procedure can be executed. To this end, we propose a sampling technique, in which the labels in the sampling are distributed proportionally to the distribution in the algorithm's output. However, the sampling running time is significantly smaller than that of the algorithm in hand.</div><div>We achieve the following results, in terms of the maximum degree Δ and the arboricity <em>a</em> of the input graph. The running time of our procedures is <span><math><mi>O</mi><mo>(</mo><mi>log</mi><mo></mo><mi>a</mi><mo>+</mo><mi>log</mi><mo></mo><mi>log</mi><mo></mo><mi>n</mi><mo>)</mo></math></span>, for sampling vertex-coloring, edge-coloring, maximal matching and MIS. This significantly improves upon previous sampling techniques, which incur additional dependency on the maximum degree Δ that can be much higher than the arboricity, as well as more significant dependency on <em>n</em>. Not only that, we also show that our technique extends naturally for the power graph <span><math><msup><mrow><mi>G</mi></mrow><mrow><mi>r</mi></mrow></msup></math></span> for any constant integer <span><math><mi>r</mi><mo>></mo><mn>1</mn></math></span> for the problems of MIS and coloring.</div><div>Our techniques for sampling in the distributed model provide a powerful and general tool for estimation in the LCA model. In this setting the goal is estimating the size of a solution to a given problem, by making as few vertex probes as possible. For the above-mentioned problems, we achieve estimations with probe complexity <span><math><msup><mrow><mi>d</mi></mrow><mrow><mi>O</mi><mo>(</mo><mi>log</mi><mo></mo><mi>a</mi><mo>+</mo><mi>log</mi><mo></mo><mi>log</mi><mo></mo><mi>n</mi><mo>)</mo></mrow></msup></math></span>, where <span><math><mi>d</mi><mo>=</mo><mi>m</mi><mi>i</mi><mi>n</mi><mo>(</mo><mi>Δ</mi><mo>,</mo><mi>a</mi><mo>⋅</mo><mi>p</mi><mi>o</mi><mi>l</mi><mi>y</mi><mo>(</mo><mi>log</mi><mo></mo><mo>(</mo><mi>n</mi><mo>)</mo><mo>)</mo></math></span>. Our results extend as well to power graphs for the coloring and MIS problems.</div></div>\",\"PeriodicalId\":49438,\"journal\":{\"name\":\"Theoretical Computer Science\",\"volume\":\"1046 \",\"pages\":\"Article 115311\"},\"PeriodicalIF\":1.0000,\"publicationDate\":\"2025-05-14\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Theoretical Computer Science\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S030439752500249X\",\"RegionNum\":4,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q3\",\"JCRName\":\"COMPUTER SCIENCE, THEORY & METHODS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Theoretical Computer Science","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S030439752500249X","RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"COMPUTER SCIENCE, THEORY & METHODS","Score":null,"Total":0}
Sampling and output estimation in distributed algorithms and LCAs
We consider the distributed message-passing model and the Local Computational Algorithms (LCA) model. In both models a network is represented by an n-vertex graph . We focus on labeling problems, such as vertex-coloring, edge-coloring, maximal independent set (MIS) and maximal matching. In the distributed model the vertices of v perform computations in parallel in order to compute their own solution for solving the problem for G. In contrast, in the LCA model probes are performed on certain vertices in order to compute their labels in a solution to a given problem. In this work we study the possibility of estimating a solution produced by an algorithm, much before the algorithm terminates. This estimation not only allows for size approximation of a solution, but also for early detection of failure in randomized algorithms. We do this such that a correcting procedure can be executed. To this end, we propose a sampling technique, in which the labels in the sampling are distributed proportionally to the distribution in the algorithm's output. However, the sampling running time is significantly smaller than that of the algorithm in hand.
We achieve the following results, in terms of the maximum degree Δ and the arboricity a of the input graph. The running time of our procedures is , for sampling vertex-coloring, edge-coloring, maximal matching and MIS. This significantly improves upon previous sampling techniques, which incur additional dependency on the maximum degree Δ that can be much higher than the arboricity, as well as more significant dependency on n. Not only that, we also show that our technique extends naturally for the power graph for any constant integer for the problems of MIS and coloring.
Our techniques for sampling in the distributed model provide a powerful and general tool for estimation in the LCA model. In this setting the goal is estimating the size of a solution to a given problem, by making as few vertex probes as possible. For the above-mentioned problems, we achieve estimations with probe complexity , where . Our results extend as well to power graphs for the coloring and MIS problems.
期刊介绍:
Theoretical Computer Science is mathematical and abstract in spirit, but it derives its motivation from practical and everyday computation. Its aim is to understand the nature of computation and, as a consequence of this understanding, provide more efficient methodologies. All papers introducing or studying mathematical, logic and formal concepts and methods are welcome, provided that their motivation is clearly drawn from the field of computing.