{"title":"Optimal Sublinear Sampling of Spanning Trees and Determinantal Point Processes via Average-Case Entropic Independence","authors":"Nima Anari, Yang P. Liu, Thuy-Duong Vuong","doi":"10.1137/22m1524321","DOIUrl":null,"url":null,"abstract":"SIAM Journal on Computing, Ahead of Print. <br/> Abstract. We design fast algorithms for repeatedly sampling from strongly Rayleigh distributions, which include as special cases random spanning tree distributions and determinantal point processes. For a graph [math], we show how to approximately sample uniformly random spanning trees from [math] in [math] (Throughout, [math] hides polylogarithmic factors in [math].) time per sample after an initial [math] time preprocessing. This is the first nearly linear runtime in the output size, which is clearly optimal. For a determinantal point process on [math]-sized subsets of a ground set of [math] elements, defined via an [math] kernel matrix, we show how to approximately sample in [math] time after an initial [math] time preprocessing, where [math] is the matrix multiplication exponent. The time to compute just the weight of the output set is simply [math], a natural barrier that suggests our runtime might be optimal for determinantal point processes as well. As a corollary, we even improve the state of the art for obtaining a single sample from a determinantal point process, from the prior runtime of [math] to [math]. In our main technical result, we achieve the optimal limit on domain sparsification for strongly Rayleigh distributions. In domain sparsification, sampling from a distribution [math] on [math] is reduced to sampling from related distributions on [math] for [math]. We show that for strongly Rayleigh distributions, the domain size can be reduced to nearly linear in the output size [math], improving the state of the art from [math] for general strongly Rayleigh distributions and the more specialized [math] for spanning tree distributions. Our reduction involves sampling from [math] domain-sparsified distributions, all of which can be produced efficiently assuming approximate overestimates for marginals of [math] are known and stored in a convenient data structure. Having access to marginals is the discrete analogue of having access to the mean and covariance of a continuous distribution, or equivalently knowing “isotropy” for the distribution, the key behind optimal samplers in the continuous setting based on the famous Kannan–Lovász–Simonovits (KLS) conjecture. We view our result as analogous in spirit to the KLS conjecture and its consequences for sampling, but rather for discrete strongly Rayleigh measures.","PeriodicalId":49532,"journal":{"name":"SIAM Journal on Computing","volume":"673 1","pages":""},"PeriodicalIF":1.2000,"publicationDate":"2024-09-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"SIAM Journal on Computing","FirstCategoryId":"94","ListUrlMain":"https://doi.org/10.1137/22m1524321","RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"COMPUTER SCIENCE, THEORY & METHODS","Score":null,"Total":0}
引用次数: 0
Abstract
SIAM Journal on Computing, Ahead of Print. Abstract. We design fast algorithms for repeatedly sampling from strongly Rayleigh distributions, which include as special cases random spanning tree distributions and determinantal point processes. For a graph [math], we show how to approximately sample uniformly random spanning trees from [math] in [math] (Throughout, [math] hides polylogarithmic factors in [math].) time per sample after an initial [math] time preprocessing. This is the first nearly linear runtime in the output size, which is clearly optimal. For a determinantal point process on [math]-sized subsets of a ground set of [math] elements, defined via an [math] kernel matrix, we show how to approximately sample in [math] time after an initial [math] time preprocessing, where [math] is the matrix multiplication exponent. The time to compute just the weight of the output set is simply [math], a natural barrier that suggests our runtime might be optimal for determinantal point processes as well. As a corollary, we even improve the state of the art for obtaining a single sample from a determinantal point process, from the prior runtime of [math] to [math]. In our main technical result, we achieve the optimal limit on domain sparsification for strongly Rayleigh distributions. In domain sparsification, sampling from a distribution [math] on [math] is reduced to sampling from related distributions on [math] for [math]. We show that for strongly Rayleigh distributions, the domain size can be reduced to nearly linear in the output size [math], improving the state of the art from [math] for general strongly Rayleigh distributions and the more specialized [math] for spanning tree distributions. Our reduction involves sampling from [math] domain-sparsified distributions, all of which can be produced efficiently assuming approximate overestimates for marginals of [math] are known and stored in a convenient data structure. Having access to marginals is the discrete analogue of having access to the mean and covariance of a continuous distribution, or equivalently knowing “isotropy” for the distribution, the key behind optimal samplers in the continuous setting based on the famous Kannan–Lovász–Simonovits (KLS) conjecture. We view our result as analogous in spirit to the KLS conjecture and its consequences for sampling, but rather for discrete strongly Rayleigh measures.
期刊介绍:
The SIAM Journal on Computing aims to provide coverage of the most significant work going on in the mathematical and formal aspects of computer science and nonnumerical computing. Submissions must be clearly written and make a significant technical contribution. Topics include but are not limited to analysis and design of algorithms, algorithmic game theory, data structures, computational complexity, computational algebra, computational aspects of combinatorics and graph theory, computational biology, computational geometry, computational robotics, the mathematical aspects of programming languages, artificial intelligence, computational learning, databases, information retrieval, cryptography, networks, distributed computing, parallel algorithms, and computer architecture.