{"title":"Identification of Single-Treatment Effects in Factorial Experiments","authors":"Guilherme Duarte","doi":"arxiv-2405.09797","DOIUrl":null,"url":null,"abstract":"Despite their cost, randomized controlled trials (RCTs) are widely regarded\nas gold-standard evidence in disciplines ranging from social science to\nmedicine. In recent decades, researchers have increasingly sought to reduce the\nresource burden of repeated RCTs with factorial designs that simultaneously\ntest multiple hypotheses, e.g. experiments that evaluate the effects of many\nmedications or products simultaneously. Here I show that when multiple\ninterventions are randomized in experiments, the effect any single intervention\nwould have outside the experimental setting is not identified absent heroic\nassumptions, even if otherwise perfectly realistic conditions are achieved.\nThis happens because single-treatment effects involve a counterfactual world\nwith a single focal intervention, allowing other variables to take their\nnatural values (which may be confounded or modified by the focal intervention).\nIn contrast, observational studies and factorial experiments provide\ninformation about potential-outcome distributions with zero and multiple\ninterventions, respectively. In this paper, I formalize sufficient conditions\nfor the identifiability of those isolated quantities. I show that researchers\nwho rely on this type of design have to justify either linearity of functional\nforms or -- in the nonparametric case -- specify with Directed Acyclic Graphs\nhow variables are related in the real world. Finally, I develop nonparametric\nsharp bounds -- i.e., maximally informative best-/worst-case estimates\nconsistent with limited RCT data -- that show when extrapolations about effect\nsigns are empirically justified. These new results are illustrated with\nsimulated data.","PeriodicalId":501323,"journal":{"name":"arXiv - STAT - Other Statistics","volume":"33 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - STAT - Other Statistics","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2405.09797","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Despite their cost, randomized controlled trials (RCTs) are widely regarded
as gold-standard evidence in disciplines ranging from social science to
medicine. In recent decades, researchers have increasingly sought to reduce the
resource burden of repeated RCTs with factorial designs that simultaneously
test multiple hypotheses, e.g. experiments that evaluate the effects of many
medications or products simultaneously. Here I show that when multiple
interventions are randomized in experiments, the effect any single intervention
would have outside the experimental setting is not identified absent heroic
assumptions, even if otherwise perfectly realistic conditions are achieved.
This happens because single-treatment effects involve a counterfactual world
with a single focal intervention, allowing other variables to take their
natural values (which may be confounded or modified by the focal intervention).
In contrast, observational studies and factorial experiments provide
information about potential-outcome distributions with zero and multiple
interventions, respectively. In this paper, I formalize sufficient conditions
for the identifiability of those isolated quantities. I show that researchers
who rely on this type of design have to justify either linearity of functional
forms or -- in the nonparametric case -- specify with Directed Acyclic Graphs
how variables are related in the real world. Finally, I develop nonparametric
sharp bounds -- i.e., maximally informative best-/worst-case estimates
consistent with limited RCT data -- that show when extrapolations about effect
signs are empirically justified. These new results are illustrated with
simulated data.