Abhinandan Dalal, Patrick Blöbaum, Shiva Kasiviswanathan, Aaditya Ramdas
{"title":"Anytime-Valid Inference for Double/Debiased Machine Learning of Causal Parameters","authors":"Abhinandan Dalal, Patrick Blöbaum, Shiva Kasiviswanathan, Aaditya Ramdas","doi":"arxiv-2408.09598","DOIUrl":null,"url":null,"abstract":"Double (debiased) machine learning (DML) has seen widespread use in recent\nyears for learning causal/structural parameters, in part due to its flexibility\nand adaptability to high-dimensional nuisance functions as well as its ability\nto avoid bias from regularization or overfitting. However, the classic\ndouble-debiased framework is only valid asymptotically for a predetermined\nsample size, thus lacking the flexibility of collecting more data if sharper\ninference is needed, or stopping data collection early if useful inferences can\nbe made earlier than expected. This can be of particular concern in large scale\nexperimental studies with huge financial costs or human lives at stake, as well\nas in observational studies where the length of confidence of intervals do not\nshrink to zero even with increasing sample size due to partial identifiability\nof a structural parameter. In this paper, we present time-uniform counterparts\nto the asymptotic DML results, enabling valid inference and confidence\nintervals for structural parameters to be constructed at any arbitrary\n(possibly data-dependent) stopping time. We provide conditions which are only\nslightly stronger than the standard DML conditions, but offer the stronger\nguarantee for anytime-valid inference. This facilitates the transformation of\nany existing DML method to provide anytime-valid guarantees with minimal\nmodifications, making it highly adaptable and easy to use. We illustrate our\nprocedure using two instances: a) local average treatment effect in online\nexperiments with non-compliance, and b) partial identification of average\ntreatment effect in observational studies with potential unmeasured\nconfounding.","PeriodicalId":501293,"journal":{"name":"arXiv - ECON - Econometrics","volume":"24 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-08-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - ECON - Econometrics","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2408.09598","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Double (debiased) machine learning (DML) has seen widespread use in recent
years for learning causal/structural parameters, in part due to its flexibility
and adaptability to high-dimensional nuisance functions as well as its ability
to avoid bias from regularization or overfitting. However, the classic
double-debiased framework is only valid asymptotically for a predetermined
sample size, thus lacking the flexibility of collecting more data if sharper
inference is needed, or stopping data collection early if useful inferences can
be made earlier than expected. This can be of particular concern in large scale
experimental studies with huge financial costs or human lives at stake, as well
as in observational studies where the length of confidence of intervals do not
shrink to zero even with increasing sample size due to partial identifiability
of a structural parameter. In this paper, we present time-uniform counterparts
to the asymptotic DML results, enabling valid inference and confidence
intervals for structural parameters to be constructed at any arbitrary
(possibly data-dependent) stopping time. We provide conditions which are only
slightly stronger than the standard DML conditions, but offer the stronger
guarantee for anytime-valid inference. This facilitates the transformation of
any existing DML method to provide anytime-valid guarantees with minimal
modifications, making it highly adaptable and easy to use. We illustrate our
procedure using two instances: a) local average treatment effect in online
experiments with non-compliance, and b) partial identification of average
treatment effect in observational studies with potential unmeasured
confounding.