{"title":"来自有问题的偏倚风险评估的不可靠证据:对Begh等人的评论,“电子烟和随后吸烟的年轻人:系统回顾”。","authors":"Sam Egger, Martin McKee","doi":"10.1111/add.70143","DOIUrl":null,"url":null,"abstract":"<p>It is widely acknowledged that cohort studies consistently find young people who use e-cigarettes are more likely to start smoking compared with non-users [<span>1</span>]. It is also recognised that in many countries, youth smoking prevalence has declined for decades and continues to decline [<span>2</span>]. However, these findings, while not necessarily contradictory, have been portrayed as inconsistent by e-cigarette and tobacco manufacturers, to emphasise uncertainty and doubt. Manufacturers often focus on results from ecological studies to minimise perceived risks to youth and argue against precautionary regulations [<span>3, 4</span>]. Given these complexities, high-quality systematic reviews integrating evidence from cohort and ecological studies are welcome, but they must assess both study types fairly and appropriately. Unfortunately, the recent review ‘Electronic cigarettes and subsequent cigarette smoking in young people: A systematic review’ [<span>5</span>] fails to do so.</p><p>Before addressing specific issues with the review, we must note that there is no established instrument for assessing the risk of bias (ROB) in ecological studies, at least not one recommended by Cochrane, the leading authority on systematic review methodology. The ROB assessment tools endorsed by Cochrane for non-randomised studies, including those by Morgan <i>et al</i>. [<span>6</span>], ROBINS-I (risk of bias in non-randomised studies of interventions) [<span>7</span>] or ROBINS-E (risk of bias in non-randomised studies of exposures) [<span>8</span>], are primarily designed for prospective cohort studies. While these include suggestions for possible adaptations for other individual-level studies, such as case–control studies, they lack guidance for ecological studies. Consequently, the ROB instrument used in this review to assess ecological studies appears largely self-designed and does not align with those tools, despite the authors’ claim that their instrument was adapted from Morgan <i>et al</i>.</p><p>A close examination of the review’s ROB assessment tool (https://osf.io/svgud) reveals a recurring pattern of ‘problematic standards’ in the design and application of ROB criteria. We consider them ‘problematic’ because they lead to unduly harsh ROB assessments of prospective cohort studies (individual-level studies) and/or unduly lenient ROB assessments of ecological studies (population-level studies). In many cases, they take the form of ‘double standards’, as the same or similar criteria could have been applied to both study types but were not. In Appendix S1, we detail 17 examples of ‘problematic standards’. The first two, relating to the ROB domain of ‘bias due to confounding’, are described as follows.</p><p>The first problematic standard concerns the requirement of instrumental variables (IVs) in cohort studies. The ROB criteria used in the review specify that for cohort studies to be classified as being at ‘low’ or ‘moderate’ risk of ‘bias due to confounding’, they must be ‘instrumental variable designs’. All other cohort studies are deemed to be at ‘serious’ or ‘critical’ risk of bias. This is despite prospective cohort studies being regarded as one of the most appropriate non-randomised study designs for determining cause and effect [<span>8-17</span>] when properly conducted, addressing issues such as potential confounding factors, selection bias and missing data. Given this extremely rigid requirement, even a well-adjusted cohort study that carefully controls for multiple confounding factors, including factors related to smoking and risk-taking behaviour, could only be classified as ‘serious risk’ of confounding at best. Given this, it is unsurprising that all 40 cohort studies assessed for ROB were classified as ‘serious’ or ‘critical’ risk of confounding.</p><p>So why do the authors privilege IV designs over all others? Unfortunately, they do not justify their inclusion anywhere in the review, and the term ‘instrumental variable’, or ‘IV’, is mentioned only once throughout the entire article, and even then no explanation is given as to why IVs would result in less biased estimates. This is concerning because IV analysis is a controversial statistical method that is rarely used in prospective cohort studies because of its important limitations, including potential biases [<span>18-20</span>]. Using an IV can sometimes introduce more bias than not adjusting for confounding factors, as it may amplify the impact of unmeasured confounding [<span>20</span>]. Moreover, although the review authors claim their ROB methods were informed by Morgan <i>et al</i>., neither Morgan <i>et al</i>. [<span>6</span>] nor ROBINS-E [<span>8</span>] mentions the use of IVs. While the ROBINS-I guidance article [<span>7</span>] briefly acknowledges the potential use of IVs it does so about an entirely different issue, the risk of ‘bias due to deviations from intended interventions’.</p><p>Worryingly, the insistence on IV designs for cohort studies also represents a ‘double standard’. If IV designs are considered essential for addressing confounding bias in cohort studies, why was this not also applied to ecological studies? If it had, 26 of the 27 included ecological studies would have been classified as having a ‘serious risk’ of confounding bias or worse (instead of the reported 11), as only one used IVs [<span>21</span>].</p><p>The authors might argue that ecological studies should not be held to the same standard as cohort studies, perhaps because of the challenges of identifying or obtaining suitable IVs in ecological research. However, this argument is fundamentally flawed. The assessment of a study’s ROB should be based solely on its inherent ROB. Whether there are practical obstacles to reducing bias is not relevant.</p><p>In summary, the review’s rigid and arbitrary requirement for cohort studies to be IV designs to be classified as ‘low risk’ or even ‘moderate risk’ of confounding is not only inconsistent with standard epidemiological practices but is also unsupported by Morgan <i>et al</i>., ROBINS-I and ROBINS-E. Having introduced this requirement for cohort studies without any apparent justification, their failure to apply it to ecological studies demonstrates how the review’s ROB methods unfairly disadvantage cohort studies.</p><p>This problematic standard is that ecological studies are classified as ‘low risk’ of ‘bias due to confounding’ if they meet the following conditions: ‘Natural experiments OR Parallel trends assumptions are tested and met AND dose response is tested for AND there are no concurrent policy changes or concurrent policy changes are controlled for AND fixed effects for place and time over which exposure varies are included’.</p><p>The review finds that 48% (<i>n</i> = 13) of the 27 ecological studies included are at ‘low risk’ of ‘bias due to confounding’. This seems surprising given the well-established limitations of ecological studies in controlling for confounding bias [<span>9-11, 22, 23</span>]. However, once again, this finding is less surprising when one looks at the criteria for classifying ecological studies as ‘low risk’. They comprise five considerations: four are interconnected and must all be met, while the fifth (‘natural experiments’) is independent and only needs to be satisfied. Despite their seemingly comprehensive nature, the four interconnected criteria: (i) provide only a superficial assessment of confounding; (ii) offer weak control of confounding; or (iii) are unrelated to the issue of confounding altogether.</p><p>The ‘parallel trends assumption’ is necessary for valid difference-in-differences analyses but cannot be empirically verified. Even if trends appear similar before the exposure, this does not guarantee that confounding factors are properly accounted for, as it fails to address unobserved confounding factors that may emerge or change after the exposure begins. ‘Dose–response’ is important for establishing causal exposure–outcome relationships and is part of the GRADE (Grading of Recommendations, Assessment, Development and Evaluations) assessment, but it is not relevant to the domain of confounding. ‘Fixed effects for place and time’ sounds impressive, but it simply refers to including fixed-effect covariates for time and place. While this provides limited control for certain characteristics that are constant in the location of the study and are subject to consistent trends, it does little to control for the many potential confounding factors that change within places over time. Examples here include changes in price, marketing and availability. Additionally, in a simple two-group design (where place defines the exposure), the effects of place and exposure cannot be separated owing to perfect collinearity. ‘Concurrent policy changes’ is the only potential confounding factor specific to smoking and vaping mentioned in the criteria. However, even if we accept that ecological studies could adequately control for concurrent policy changes (which is questionable given the complexity of potential interactions), this approach ignores a wide range of other social, economic and behavioural factors that may shift over time and influence outcomes.</p><p>A more problematic aspect of the ROB criteria is that any study labelled as a ‘natural experiment’ is automatically classified as being at ‘low risk’ of confounding, regardless of whether it takes steps to address confounding. Natural experiments employ various statistical methods, such as interrupted time series analysis or synthetic controls, and can provide valuable insights when certain conditions are met, such as a discrete intervention, rapid implementation, and short lags between implementation and impact. However, as with any study, all assumptions must be tested, sensitivity analyses conducted and falsifiability tests conducted. Just because something is a ‘natural experiment’, it cannot be deemed ‘low risk’, regardless of whether any efforts are made to control for confounding. This is akin to claiming that all cohort studies are inherently ‘low risk’ for confounding bias because they have an exposed and a non-exposed group, and do not need to control for confounding factors.</p><p>Given this standard, the authors’ ROB assessments effectively imply that 48% of the 27 ecological studies are equivalent to high-quality randomised trials in terms of confounding control. This is so implausible that it raises serious concerns about the credibility of the entire review.</p><p>However, there is another concern. Two review group members were credited with providing ‘… expert input on the risk of bias due to confounding associated with different population-level study designs’. Notably, they are co-authors on 10 of the 27 ecological studies included in the review [<span>21, 25-33</span>], and of those 10 studies, eight were classified as being at ‘low risk’ of confounding bias. In other words, the same experts who helped design the ROB criteria benefited from an assessment that deemed 80% of their ecological studies equivalent to high-quality randomised trials.</p><p>In this letter, we described two problematic ROB standards from Begh <i>et al</i>.’s systematic review, with 15 additional issues detailed in Appendix S1. In these two examples, we showed how the review creates a significant methodological imbalance by imposing an unjustified requirement for IV designs in cohort studies while using lenient and inconsistently justified ROB criteria for ecological studies. This leads to cohort studies – one of the most effective non-randomised study designs for determining cause and effect when properly conducted – to be dismissed while ecological studies are elevated to an implausible level of credibility in terms of confounding control.</p><p>The insistence on IV designs for cohort studies is particularly problematic, as IV analyses are rarely used in prospective cohort studies because of their limitations. The authors fail to justify their inclusion, and established ROB assessment tools do not endorse this approach. Consequently, all 40 cohort studies were classified as having a ‘serious’ or ‘critical’ ROB due to confounding.</p><p>Conversely, the review’s treatment of ecological studies is far more permissive. Despite well-documented limitations, nearly half (48%) were classified as being at ‘low risk’ for confounding. The ROB criteria rely on weak or irrelevant factors, such as the ‘parallel trends assumption’ (which cannot be empirically verified) and ‘dose–response’ (which pertains to causal inference rather than confounding). Most concerning is the blanket classification of all ‘natural experiments’ as being at ‘low risk’ for confounding, regardless of whether they take steps to address bias. This creates an illusion of methodological rigour while shielding ecological studies from scrutiny.</p><p>What appears to be a potential conflict of interest raises further concern. Two review group members who helped design the ROB criteria are co-authors of 10 of the included ecological studies, eight of which were rated as being at ‘low risk’ for confounding.</p><p>Because we did not have the capacity to conduct a complete re-analysis of Berg <i>et al</i>.’s ROB assessments with an appropriate ROB instrument, we cannot be certain how their flawed methodology affected their conclusions. Ultimately, however, this review’s selective and inconsistent design and application of ROB criteria has the potential to distort the scientific landscape. Rather than fairly synthesising the evidence, the review systematically disadvantages cohort studies while presenting ecological studies as more reliable than they truly are. Given the public health implications of e-cigarette regulation, these methodological biases risk shaping policy based on unreliable evidence.</p><p><b>Sam Egger</b>: Conceptualization; investigation; writing—original draft. <b>Martin McKee</b>: Conceptualization; writing—review and editing.</p><p>Martin McKee is a past president of the British Medical Association and of the European Public Health Association, both of which have expressed concerns about the public health impact and appropriate regulation of e-cigarettes. Sam Egger is a statistical editor on the editorial board of the Cochrane Breast Cancer Group, serves as a peer reviewer for the Cochrane Database of Systematic Reviews, and has expertise in the relationship between adolescent vaping and future smoking.</p>","PeriodicalId":109,"journal":{"name":"Addiction","volume":"120 11","pages":"2355-2358"},"PeriodicalIF":5.3000,"publicationDate":"2025-07-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1111/add.70143","citationCount":"0","resultStr":"{\"title\":\"Unreliable evidence from problematic risk of bias assessments: Comment on Begh et al., ‘Electronic cigarettes and subsequent cigarette smoking in young people: A systematic review’\",\"authors\":\"Sam Egger, Martin McKee\",\"doi\":\"10.1111/add.70143\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p>It is widely acknowledged that cohort studies consistently find young people who use e-cigarettes are more likely to start smoking compared with non-users [<span>1</span>]. It is also recognised that in many countries, youth smoking prevalence has declined for decades and continues to decline [<span>2</span>]. However, these findings, while not necessarily contradictory, have been portrayed as inconsistent by e-cigarette and tobacco manufacturers, to emphasise uncertainty and doubt. Manufacturers often focus on results from ecological studies to minimise perceived risks to youth and argue against precautionary regulations [<span>3, 4</span>]. Given these complexities, high-quality systematic reviews integrating evidence from cohort and ecological studies are welcome, but they must assess both study types fairly and appropriately. Unfortunately, the recent review ‘Electronic cigarettes and subsequent cigarette smoking in young people: A systematic review’ [<span>5</span>] fails to do so.</p><p>Before addressing specific issues with the review, we must note that there is no established instrument for assessing the risk of bias (ROB) in ecological studies, at least not one recommended by Cochrane, the leading authority on systematic review methodology. The ROB assessment tools endorsed by Cochrane for non-randomised studies, including those by Morgan <i>et al</i>. [<span>6</span>], ROBINS-I (risk of bias in non-randomised studies of interventions) [<span>7</span>] or ROBINS-E (risk of bias in non-randomised studies of exposures) [<span>8</span>], are primarily designed for prospective cohort studies. While these include suggestions for possible adaptations for other individual-level studies, such as case–control studies, they lack guidance for ecological studies. Consequently, the ROB instrument used in this review to assess ecological studies appears largely self-designed and does not align with those tools, despite the authors’ claim that their instrument was adapted from Morgan <i>et al</i>.</p><p>A close examination of the review’s ROB assessment tool (https://osf.io/svgud) reveals a recurring pattern of ‘problematic standards’ in the design and application of ROB criteria. We consider them ‘problematic’ because they lead to unduly harsh ROB assessments of prospective cohort studies (individual-level studies) and/or unduly lenient ROB assessments of ecological studies (population-level studies). In many cases, they take the form of ‘double standards’, as the same or similar criteria could have been applied to both study types but were not. In Appendix S1, we detail 17 examples of ‘problematic standards’. The first two, relating to the ROB domain of ‘bias due to confounding’, are described as follows.</p><p>The first problematic standard concerns the requirement of instrumental variables (IVs) in cohort studies. The ROB criteria used in the review specify that for cohort studies to be classified as being at ‘low’ or ‘moderate’ risk of ‘bias due to confounding’, they must be ‘instrumental variable designs’. All other cohort studies are deemed to be at ‘serious’ or ‘critical’ risk of bias. This is despite prospective cohort studies being regarded as one of the most appropriate non-randomised study designs for determining cause and effect [<span>8-17</span>] when properly conducted, addressing issues such as potential confounding factors, selection bias and missing data. Given this extremely rigid requirement, even a well-adjusted cohort study that carefully controls for multiple confounding factors, including factors related to smoking and risk-taking behaviour, could only be classified as ‘serious risk’ of confounding at best. Given this, it is unsurprising that all 40 cohort studies assessed for ROB were classified as ‘serious’ or ‘critical’ risk of confounding.</p><p>So why do the authors privilege IV designs over all others? Unfortunately, they do not justify their inclusion anywhere in the review, and the term ‘instrumental variable’, or ‘IV’, is mentioned only once throughout the entire article, and even then no explanation is given as to why IVs would result in less biased estimates. This is concerning because IV analysis is a controversial statistical method that is rarely used in prospective cohort studies because of its important limitations, including potential biases [<span>18-20</span>]. Using an IV can sometimes introduce more bias than not adjusting for confounding factors, as it may amplify the impact of unmeasured confounding [<span>20</span>]. Moreover, although the review authors claim their ROB methods were informed by Morgan <i>et al</i>., neither Morgan <i>et al</i>. [<span>6</span>] nor ROBINS-E [<span>8</span>] mentions the use of IVs. While the ROBINS-I guidance article [<span>7</span>] briefly acknowledges the potential use of IVs it does so about an entirely different issue, the risk of ‘bias due to deviations from intended interventions’.</p><p>Worryingly, the insistence on IV designs for cohort studies also represents a ‘double standard’. If IV designs are considered essential for addressing confounding bias in cohort studies, why was this not also applied to ecological studies? If it had, 26 of the 27 included ecological studies would have been classified as having a ‘serious risk’ of confounding bias or worse (instead of the reported 11), as only one used IVs [<span>21</span>].</p><p>The authors might argue that ecological studies should not be held to the same standard as cohort studies, perhaps because of the challenges of identifying or obtaining suitable IVs in ecological research. However, this argument is fundamentally flawed. The assessment of a study’s ROB should be based solely on its inherent ROB. Whether there are practical obstacles to reducing bias is not relevant.</p><p>In summary, the review’s rigid and arbitrary requirement for cohort studies to be IV designs to be classified as ‘low risk’ or even ‘moderate risk’ of confounding is not only inconsistent with standard epidemiological practices but is also unsupported by Morgan <i>et al</i>., ROBINS-I and ROBINS-E. Having introduced this requirement for cohort studies without any apparent justification, their failure to apply it to ecological studies demonstrates how the review’s ROB methods unfairly disadvantage cohort studies.</p><p>This problematic standard is that ecological studies are classified as ‘low risk’ of ‘bias due to confounding’ if they meet the following conditions: ‘Natural experiments OR Parallel trends assumptions are tested and met AND dose response is tested for AND there are no concurrent policy changes or concurrent policy changes are controlled for AND fixed effects for place and time over which exposure varies are included’.</p><p>The review finds that 48% (<i>n</i> = 13) of the 27 ecological studies included are at ‘low risk’ of ‘bias due to confounding’. This seems surprising given the well-established limitations of ecological studies in controlling for confounding bias [<span>9-11, 22, 23</span>]. However, once again, this finding is less surprising when one looks at the criteria for classifying ecological studies as ‘low risk’. They comprise five considerations: four are interconnected and must all be met, while the fifth (‘natural experiments’) is independent and only needs to be satisfied. Despite their seemingly comprehensive nature, the four interconnected criteria: (i) provide only a superficial assessment of confounding; (ii) offer weak control of confounding; or (iii) are unrelated to the issue of confounding altogether.</p><p>The ‘parallel trends assumption’ is necessary for valid difference-in-differences analyses but cannot be empirically verified. Even if trends appear similar before the exposure, this does not guarantee that confounding factors are properly accounted for, as it fails to address unobserved confounding factors that may emerge or change after the exposure begins. ‘Dose–response’ is important for establishing causal exposure–outcome relationships and is part of the GRADE (Grading of Recommendations, Assessment, Development and Evaluations) assessment, but it is not relevant to the domain of confounding. ‘Fixed effects for place and time’ sounds impressive, but it simply refers to including fixed-effect covariates for time and place. While this provides limited control for certain characteristics that are constant in the location of the study and are subject to consistent trends, it does little to control for the many potential confounding factors that change within places over time. Examples here include changes in price, marketing and availability. Additionally, in a simple two-group design (where place defines the exposure), the effects of place and exposure cannot be separated owing to perfect collinearity. ‘Concurrent policy changes’ is the only potential confounding factor specific to smoking and vaping mentioned in the criteria. However, even if we accept that ecological studies could adequately control for concurrent policy changes (which is questionable given the complexity of potential interactions), this approach ignores a wide range of other social, economic and behavioural factors that may shift over time and influence outcomes.</p><p>A more problematic aspect of the ROB criteria is that any study labelled as a ‘natural experiment’ is automatically classified as being at ‘low risk’ of confounding, regardless of whether it takes steps to address confounding. Natural experiments employ various statistical methods, such as interrupted time series analysis or synthetic controls, and can provide valuable insights when certain conditions are met, such as a discrete intervention, rapid implementation, and short lags between implementation and impact. However, as with any study, all assumptions must be tested, sensitivity analyses conducted and falsifiability tests conducted. Just because something is a ‘natural experiment’, it cannot be deemed ‘low risk’, regardless of whether any efforts are made to control for confounding. This is akin to claiming that all cohort studies are inherently ‘low risk’ for confounding bias because they have an exposed and a non-exposed group, and do not need to control for confounding factors.</p><p>Given this standard, the authors’ ROB assessments effectively imply that 48% of the 27 ecological studies are equivalent to high-quality randomised trials in terms of confounding control. This is so implausible that it raises serious concerns about the credibility of the entire review.</p><p>However, there is another concern. Two review group members were credited with providing ‘… expert input on the risk of bias due to confounding associated with different population-level study designs’. Notably, they are co-authors on 10 of the 27 ecological studies included in the review [<span>21, 25-33</span>], and of those 10 studies, eight were classified as being at ‘low risk’ of confounding bias. In other words, the same experts who helped design the ROB criteria benefited from an assessment that deemed 80% of their ecological studies equivalent to high-quality randomised trials.</p><p>In this letter, we described two problematic ROB standards from Begh <i>et al</i>.’s systematic review, with 15 additional issues detailed in Appendix S1. In these two examples, we showed how the review creates a significant methodological imbalance by imposing an unjustified requirement for IV designs in cohort studies while using lenient and inconsistently justified ROB criteria for ecological studies. This leads to cohort studies – one of the most effective non-randomised study designs for determining cause and effect when properly conducted – to be dismissed while ecological studies are elevated to an implausible level of credibility in terms of confounding control.</p><p>The insistence on IV designs for cohort studies is particularly problematic, as IV analyses are rarely used in prospective cohort studies because of their limitations. The authors fail to justify their inclusion, and established ROB assessment tools do not endorse this approach. Consequently, all 40 cohort studies were classified as having a ‘serious’ or ‘critical’ ROB due to confounding.</p><p>Conversely, the review’s treatment of ecological studies is far more permissive. Despite well-documented limitations, nearly half (48%) were classified as being at ‘low risk’ for confounding. The ROB criteria rely on weak or irrelevant factors, such as the ‘parallel trends assumption’ (which cannot be empirically verified) and ‘dose–response’ (which pertains to causal inference rather than confounding). Most concerning is the blanket classification of all ‘natural experiments’ as being at ‘low risk’ for confounding, regardless of whether they take steps to address bias. This creates an illusion of methodological rigour while shielding ecological studies from scrutiny.</p><p>What appears to be a potential conflict of interest raises further concern. Two review group members who helped design the ROB criteria are co-authors of 10 of the included ecological studies, eight of which were rated as being at ‘low risk’ for confounding.</p><p>Because we did not have the capacity to conduct a complete re-analysis of Berg <i>et al</i>.’s ROB assessments with an appropriate ROB instrument, we cannot be certain how their flawed methodology affected their conclusions. Ultimately, however, this review’s selective and inconsistent design and application of ROB criteria has the potential to distort the scientific landscape. Rather than fairly synthesising the evidence, the review systematically disadvantages cohort studies while presenting ecological studies as more reliable than they truly are. Given the public health implications of e-cigarette regulation, these methodological biases risk shaping policy based on unreliable evidence.</p><p><b>Sam Egger</b>: Conceptualization; investigation; writing—original draft. <b>Martin McKee</b>: Conceptualization; writing—review and editing.</p><p>Martin McKee is a past president of the British Medical Association and of the European Public Health Association, both of which have expressed concerns about the public health impact and appropriate regulation of e-cigarettes. Sam Egger is a statistical editor on the editorial board of the Cochrane Breast Cancer Group, serves as a peer reviewer for the Cochrane Database of Systematic Reviews, and has expertise in the relationship between adolescent vaping and future smoking.</p>\",\"PeriodicalId\":109,\"journal\":{\"name\":\"Addiction\",\"volume\":\"120 11\",\"pages\":\"2355-2358\"},\"PeriodicalIF\":5.3000,\"publicationDate\":\"2025-07-05\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://onlinelibrary.wiley.com/doi/epdf/10.1111/add.70143\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Addiction\",\"FirstCategoryId\":\"3\",\"ListUrlMain\":\"https://onlinelibrary.wiley.com/doi/10.1111/add.70143\",\"RegionNum\":1,\"RegionCategory\":\"医学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"PSYCHIATRY\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Addiction","FirstCategoryId":"3","ListUrlMain":"https://onlinelibrary.wiley.com/doi/10.1111/add.70143","RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"PSYCHIATRY","Score":null,"Total":0}
引用次数: 0
摘要
人们普遍认为,队列研究一致发现,使用电子烟的年轻人比不使用电子烟的年轻人更有可能开始吸烟。人们还认识到,在许多国家,青少年吸烟率几十年来一直在下降,并将继续下降。然而,这些发现虽然不一定相互矛盾,但被电子烟和烟草制造商描述为不一致,以强调不确定性和怀疑。制造商经常关注生态研究的结果,以尽量减少对年轻人的感知风险,并反对预防性法规[3,4]。考虑到这些复杂性,整合队列研究和生态学研究证据的高质量系统评价是受欢迎的,但它们必须公平和适当地评估这两种研究类型。不幸的是,最近的评论“电子烟和随后的吸烟在年轻人:一个系统的回顾”b[5]没有做到这一点。在讨论这篇综述的具体问题之前,我们必须注意到,目前还没有确定的工具来评估生态学研究中的偏倚风险(ROB),至少没有一个由Cochrane推荐的工具,Cochrane是系统综述方法的权威。Cochrane认可的用于非随机研究的ROB评估工具,包括Morgan等人的研究。b[6]、ROBINS-I(干预措施的非随机研究的偏倚风险)[7]或ROBINS-E(暴露的非随机研究的偏倚风险)[8]主要是为前瞻性队列研究设计的。虽然这些建议包括对其他个人水平研究(如病例对照研究)可能进行的调整,但它们缺乏对生态研究的指导。因此,本综述中用于评估生态研究的ROB工具似乎主要是自行设计的,与这些工具不一致,尽管作者声称他们的工具改编自Morgan等人。对该综述的ROB评估工具(https://osf.io/svgud)的仔细检查揭示了在设计和应用ROB标准时反复出现的“有问题的标准”模式。我们认为它们是“有问题的”,因为它们导致前瞻性队列研究(个体水平研究)的ROB评估过于苛刻和/或生态研究(种群水平研究)的ROB评估过于宽松。在许多情况下,它们采取“双重标准”的形式,因为相同或类似的标准可以适用于两种研究类型,但却没有。在附录S1中,我们详细介绍了17个“有问题的标准”的例子。前两个与ROB域的“混杂偏差”有关,描述如下。第一个有问题的标准涉及队列研究中工具变量(IVs)的要求。本综述中使用的ROB标准规定,对于被归类为“低”或“中等”“混杂偏倚”风险的队列研究,它们必须是“工具变量设计”。所有其他队列研究被认为存在“严重”或“关键”偏倚风险。尽管前瞻性队列研究被认为是确定因果关系的最合适的非随机研究设计之一[8-17],但如果进行得当,可以解决潜在的混杂因素、选择偏差和缺失数据等问题。考虑到这一极其严格的要求,即使是经过精心调整的队列研究,仔细控制了多种混杂因素,包括与吸烟和冒险行为有关的因素,最多也只能被归类为混杂的“严重风险”。考虑到这一点,对ROB进行评估的所有40项队列研究被归类为“严重”或“关键”混杂风险也就不足为奇了。那么,为什么作者将IV设计置于其他设计之上呢?不幸的是,它们并没有证明它们在综述中的任何地方都是合理的,“工具变量”一词在整篇文章中只被提及一次,即使如此,也没有解释为什么IV会导致更少的偏差估计。这是令人担忧的,因为IV分析是一种有争议的统计方法,由于其重要的局限性,包括潜在的偏差,很少用于前瞻性队列研究[18-20]。使用IV有时会比不调整混杂因素带来更多的偏差,因为它可能会放大未测量的混杂bbb的影响。此外,尽管综述作者声称他们的ROB方法是由Morgan等人提供的,但Morgan等人[8]0和ROBINS-E[8]都没有提到静脉注射的使用。虽然ROBINS-I指南文章[7]简要地承认了静脉注射的潜在用途,但它涉及的是一个完全不同的问题,即“因偏离预期干预措施而产生的偏见”的风险。令人担忧的是,在队列研究中坚持静脉注射设计也代表了一种“双重标准”。 如果IV设计被认为是解决队列研究中混杂偏倚的必要条件,为什么不将其应用于生态学研究?如果有的话,27项纳入的生态学研究中有26项将被归类为具有混杂偏倚的“严重风险”或更严重的风险(而不是报道的11项),因为只有一项使用了IVs bbb。作者可能会争辩说,生态研究不应该与队列研究保持相同的标准,也许是因为在生态研究中识别或获得合适的IVs存在挑战。然而,这种观点从根本上是有缺陷的。评估一项研究的效率效率应完全基于其固有的效率效率。是否存在减少偏见的实际障碍与此无关。综上所述,该综述严格武断地要求将队列研究归类为“低风险”甚至“中等风险”的IV型设计,这不仅与标准流行病学实践不符,而且也没有得到Morgan等人、ROBINS-I和ROBINS-E的支持。在没有任何明显理由的情况下引入队列研究的这一要求,他们未能将其应用于生态学研究,这表明该综述的ROB方法不公平地使队列研究处于不利地位。这个有问题的标准是,如果符合以下条件,生态研究被归类为“低风险”的“混杂偏倚”:“对自然实验或平行趋势假设进行了测试并满足,对剂量反应进行了测试,并且没有同时发生的政策变化,或者同时发生的政策变化被控制,并且固定的影响在暴露变化的地点和时间内被包括在内”。该综述发现,纳入的27项生态学研究中有48% (n = 13)处于“低风险”的“混杂偏倚”。考虑到生态学研究在控制混杂偏差方面的公认局限性,这似乎令人惊讶[9- 11,22,23]。然而,再一次,当人们考虑到将生态学研究归类为“低风险”的标准时,这一发现就不那么令人惊讶了。它们包括五个考虑因素:四个是相互关联的,必须全部满足,而第五个(“自然实验”)是独立的,只需要满足。尽管这四个相互关联的标准看似全面,但它们:(i)只提供了对混淆的肤浅评估;(ii)对混淆的控制较弱;或者(iii)与混淆问题完全无关。“平行趋势假设”对于有效的差异中差异分析是必要的,但不能通过经验验证。即使趋势在暴露前出现相似,这也不能保证混淆因素得到了适当的解释,因为它不能解决暴露开始后可能出现或改变的未观察到的混淆因素。“剂量-反应”对于建立因果暴露-结果关系很重要,是GRADE(建议、评估、发展和评估分级)评估的一部分,但与混淆领域无关。“地点和时间的固定效应”听起来令人印象深刻,但它只是指包括时间和地点的固定效应协变量。虽然这对研究地点的某些特征提供了有限的控制,这些特征在研究地点是恒定的,并且服从于一致的趋势,但它对控制许多潜在的混淆因素几乎没有作用,这些因素随着时间的推移在地方内发生变化。这里的例子包括价格、营销和可用性的变化。此外,在简单的两组设计中(其中位置定义曝光),由于完美的共线性,位置和曝光的效果不能分开。“政策同步变化”是标准中提到的唯一一个与吸烟和电子烟有关的潜在混杂因素。然而,即使我们接受生态研究可以充分控制同时发生的政策变化(考虑到潜在相互作用的复杂性,这是值得怀疑的),这种方法也忽略了广泛的其他社会、经济和行为因素,这些因素可能随着时间的推移而变化并影响结果。ROB标准的一个更有问题的方面是,任何标记为“自然实验”的研究都被自动归类为混淆的“低风险”,而不管它是否采取措施解决混淆。自然实验采用各种统计方法,如中断时间序列分析或综合控制,当满足某些条件时,如离散干预、快速实施和实施与影响之间的短滞后,可以提供有价值的见解。但是,与任何研究一样,必须检验所有假设,进行敏感性分析和可证伪性检验。仅仅因为某件事是“自然实验”,就不能认为它是“低风险”的,不管是否做了任何努力来控制混淆。 这类似于声称所有队列研究天生具有混杂偏倚的“低风险”,因为它们有暴露组和非暴露组,并且不需要控制混杂因素。考虑到这一标准,作者的ROB评估有效地暗示,在混淆控制方面,27项生态学研究中有48%相当于高质量的随机试验。这是如此令人难以置信,以至于引发了对整个审查可信度的严重担忧。然而,还有另一个担忧。两名评审小组成员因提供“关于不同人群水平研究设计相关的混杂导致的偏倚风险的专家意见”而受到赞扬。值得注意的是,他们是综述中包含的27项生态学研究中的10项的共同作者[21,25 -33],在这10项研究中,有8项被归类为混杂偏倚的“低风险”。换句话说,帮助设计ROB标准的专家受益于一项评估,该评估认为他们80%的生态学研究相当于高质量的随机试验。在这封信中,我们从Begh等人的系统综述中描述了两个有问题的ROB标准,并在附录S1中详细说明了另外15个问题。在这两个例子中,我们展示了该综述如何通过在队列研究中对IV设计施加不合理的要求而在生态学研究中使用宽松且不一致的ROB标准,从而造成了显著的方法失衡。这导致队列研究——在正确进行的情况下确定因果关系的最有效的非随机研究设计之一——被驳回,而生态学研究在混淆控制方面被提升到令人难以置信的可信度水平。在队列研究中坚持使用IV设计尤其有问题,因为IV分析由于其局限性而很少用于前瞻性队列研究。作者未能证明他们的纳入,并且已建立的ROB评估工具不支持这种方法。因此,由于混淆,所有40项队列研究都被归类为“严重”或“严重”ROB。相反,该评论对生态学研究的处理要宽容得多。尽管有充分的证据限制,但近一半(48%)被归类为混杂的“低风险”。ROB标准依赖于弱的或不相关的因素,例如“平行趋势假设”(无法通过经验验证)和“剂量-反应”(与因果推断有关,而不是混淆)。最令人担忧的是,所有“自然实验”都被笼统地归类为“低风险”混淆,而不管他们是否采取措施解决偏见。这造成了一种方法论严谨的错觉,同时保护了生态学研究免受审查。这种看似潜在的利益冲突引发了进一步的担忧。帮助设计ROB标准的两名评审小组成员是纳入的10项生态学研究的共同作者,其中8项被评为“低混杂风险”。由于我们没有能力用合适的ROB工具对Berg等人的ROB评估进行完整的重新分析,我们无法确定他们有缺陷的方法如何影响他们的结论。然而,最终,本综述对ROB标准的选择性和不一致的设计和应用有可能扭曲科学景观。这篇综述并没有公平地综合证据,而是系统性地贬低了队列研究,同时将生态学研究呈现得比实际更可靠。考虑到电子烟监管对公共卫生的影响,这些方法上的偏差可能会根据不可靠的证据来制定政策。Sam Egger:概念化;调查;原创作品。马丁·麦基:概念化;写作-审查和编辑。Martin McKee是英国医学协会和欧洲公共卫生协会的前任主席,这两个协会都对电子烟对公共卫生的影响和适当的监管表示担忧。山姆·埃格(Sam Egger)是科克伦乳腺癌小组(Cochrane Breast Cancer Group)编辑委员会的统计编辑,也是科克伦系统评论数据库(Cochrane Database of Systematic Reviews)的同行评议人,在青少年吸电子烟与未来吸烟之间的关系方面拥有专业知识。
Unreliable evidence from problematic risk of bias assessments: Comment on Begh et al., ‘Electronic cigarettes and subsequent cigarette smoking in young people: A systematic review’
It is widely acknowledged that cohort studies consistently find young people who use e-cigarettes are more likely to start smoking compared with non-users [1]. It is also recognised that in many countries, youth smoking prevalence has declined for decades and continues to decline [2]. However, these findings, while not necessarily contradictory, have been portrayed as inconsistent by e-cigarette and tobacco manufacturers, to emphasise uncertainty and doubt. Manufacturers often focus on results from ecological studies to minimise perceived risks to youth and argue against precautionary regulations [3, 4]. Given these complexities, high-quality systematic reviews integrating evidence from cohort and ecological studies are welcome, but they must assess both study types fairly and appropriately. Unfortunately, the recent review ‘Electronic cigarettes and subsequent cigarette smoking in young people: A systematic review’ [5] fails to do so.
Before addressing specific issues with the review, we must note that there is no established instrument for assessing the risk of bias (ROB) in ecological studies, at least not one recommended by Cochrane, the leading authority on systematic review methodology. The ROB assessment tools endorsed by Cochrane for non-randomised studies, including those by Morgan et al. [6], ROBINS-I (risk of bias in non-randomised studies of interventions) [7] or ROBINS-E (risk of bias in non-randomised studies of exposures) [8], are primarily designed for prospective cohort studies. While these include suggestions for possible adaptations for other individual-level studies, such as case–control studies, they lack guidance for ecological studies. Consequently, the ROB instrument used in this review to assess ecological studies appears largely self-designed and does not align with those tools, despite the authors’ claim that their instrument was adapted from Morgan et al.
A close examination of the review’s ROB assessment tool (https://osf.io/svgud) reveals a recurring pattern of ‘problematic standards’ in the design and application of ROB criteria. We consider them ‘problematic’ because they lead to unduly harsh ROB assessments of prospective cohort studies (individual-level studies) and/or unduly lenient ROB assessments of ecological studies (population-level studies). In many cases, they take the form of ‘double standards’, as the same or similar criteria could have been applied to both study types but were not. In Appendix S1, we detail 17 examples of ‘problematic standards’. The first two, relating to the ROB domain of ‘bias due to confounding’, are described as follows.
The first problematic standard concerns the requirement of instrumental variables (IVs) in cohort studies. The ROB criteria used in the review specify that for cohort studies to be classified as being at ‘low’ or ‘moderate’ risk of ‘bias due to confounding’, they must be ‘instrumental variable designs’. All other cohort studies are deemed to be at ‘serious’ or ‘critical’ risk of bias. This is despite prospective cohort studies being regarded as one of the most appropriate non-randomised study designs for determining cause and effect [8-17] when properly conducted, addressing issues such as potential confounding factors, selection bias and missing data. Given this extremely rigid requirement, even a well-adjusted cohort study that carefully controls for multiple confounding factors, including factors related to smoking and risk-taking behaviour, could only be classified as ‘serious risk’ of confounding at best. Given this, it is unsurprising that all 40 cohort studies assessed for ROB were classified as ‘serious’ or ‘critical’ risk of confounding.
So why do the authors privilege IV designs over all others? Unfortunately, they do not justify their inclusion anywhere in the review, and the term ‘instrumental variable’, or ‘IV’, is mentioned only once throughout the entire article, and even then no explanation is given as to why IVs would result in less biased estimates. This is concerning because IV analysis is a controversial statistical method that is rarely used in prospective cohort studies because of its important limitations, including potential biases [18-20]. Using an IV can sometimes introduce more bias than not adjusting for confounding factors, as it may amplify the impact of unmeasured confounding [20]. Moreover, although the review authors claim their ROB methods were informed by Morgan et al., neither Morgan et al. [6] nor ROBINS-E [8] mentions the use of IVs. While the ROBINS-I guidance article [7] briefly acknowledges the potential use of IVs it does so about an entirely different issue, the risk of ‘bias due to deviations from intended interventions’.
Worryingly, the insistence on IV designs for cohort studies also represents a ‘double standard’. If IV designs are considered essential for addressing confounding bias in cohort studies, why was this not also applied to ecological studies? If it had, 26 of the 27 included ecological studies would have been classified as having a ‘serious risk’ of confounding bias or worse (instead of the reported 11), as only one used IVs [21].
The authors might argue that ecological studies should not be held to the same standard as cohort studies, perhaps because of the challenges of identifying or obtaining suitable IVs in ecological research. However, this argument is fundamentally flawed. The assessment of a study’s ROB should be based solely on its inherent ROB. Whether there are practical obstacles to reducing bias is not relevant.
In summary, the review’s rigid and arbitrary requirement for cohort studies to be IV designs to be classified as ‘low risk’ or even ‘moderate risk’ of confounding is not only inconsistent with standard epidemiological practices but is also unsupported by Morgan et al., ROBINS-I and ROBINS-E. Having introduced this requirement for cohort studies without any apparent justification, their failure to apply it to ecological studies demonstrates how the review’s ROB methods unfairly disadvantage cohort studies.
This problematic standard is that ecological studies are classified as ‘low risk’ of ‘bias due to confounding’ if they meet the following conditions: ‘Natural experiments OR Parallel trends assumptions are tested and met AND dose response is tested for AND there are no concurrent policy changes or concurrent policy changes are controlled for AND fixed effects for place and time over which exposure varies are included’.
The review finds that 48% (n = 13) of the 27 ecological studies included are at ‘low risk’ of ‘bias due to confounding’. This seems surprising given the well-established limitations of ecological studies in controlling for confounding bias [9-11, 22, 23]. However, once again, this finding is less surprising when one looks at the criteria for classifying ecological studies as ‘low risk’. They comprise five considerations: four are interconnected and must all be met, while the fifth (‘natural experiments’) is independent and only needs to be satisfied. Despite their seemingly comprehensive nature, the four interconnected criteria: (i) provide only a superficial assessment of confounding; (ii) offer weak control of confounding; or (iii) are unrelated to the issue of confounding altogether.
The ‘parallel trends assumption’ is necessary for valid difference-in-differences analyses but cannot be empirically verified. Even if trends appear similar before the exposure, this does not guarantee that confounding factors are properly accounted for, as it fails to address unobserved confounding factors that may emerge or change after the exposure begins. ‘Dose–response’ is important for establishing causal exposure–outcome relationships and is part of the GRADE (Grading of Recommendations, Assessment, Development and Evaluations) assessment, but it is not relevant to the domain of confounding. ‘Fixed effects for place and time’ sounds impressive, but it simply refers to including fixed-effect covariates for time and place. While this provides limited control for certain characteristics that are constant in the location of the study and are subject to consistent trends, it does little to control for the many potential confounding factors that change within places over time. Examples here include changes in price, marketing and availability. Additionally, in a simple two-group design (where place defines the exposure), the effects of place and exposure cannot be separated owing to perfect collinearity. ‘Concurrent policy changes’ is the only potential confounding factor specific to smoking and vaping mentioned in the criteria. However, even if we accept that ecological studies could adequately control for concurrent policy changes (which is questionable given the complexity of potential interactions), this approach ignores a wide range of other social, economic and behavioural factors that may shift over time and influence outcomes.
A more problematic aspect of the ROB criteria is that any study labelled as a ‘natural experiment’ is automatically classified as being at ‘low risk’ of confounding, regardless of whether it takes steps to address confounding. Natural experiments employ various statistical methods, such as interrupted time series analysis or synthetic controls, and can provide valuable insights when certain conditions are met, such as a discrete intervention, rapid implementation, and short lags between implementation and impact. However, as with any study, all assumptions must be tested, sensitivity analyses conducted and falsifiability tests conducted. Just because something is a ‘natural experiment’, it cannot be deemed ‘low risk’, regardless of whether any efforts are made to control for confounding. This is akin to claiming that all cohort studies are inherently ‘low risk’ for confounding bias because they have an exposed and a non-exposed group, and do not need to control for confounding factors.
Given this standard, the authors’ ROB assessments effectively imply that 48% of the 27 ecological studies are equivalent to high-quality randomised trials in terms of confounding control. This is so implausible that it raises serious concerns about the credibility of the entire review.
However, there is another concern. Two review group members were credited with providing ‘… expert input on the risk of bias due to confounding associated with different population-level study designs’. Notably, they are co-authors on 10 of the 27 ecological studies included in the review [21, 25-33], and of those 10 studies, eight were classified as being at ‘low risk’ of confounding bias. In other words, the same experts who helped design the ROB criteria benefited from an assessment that deemed 80% of their ecological studies equivalent to high-quality randomised trials.
In this letter, we described two problematic ROB standards from Begh et al.’s systematic review, with 15 additional issues detailed in Appendix S1. In these two examples, we showed how the review creates a significant methodological imbalance by imposing an unjustified requirement for IV designs in cohort studies while using lenient and inconsistently justified ROB criteria for ecological studies. This leads to cohort studies – one of the most effective non-randomised study designs for determining cause and effect when properly conducted – to be dismissed while ecological studies are elevated to an implausible level of credibility in terms of confounding control.
The insistence on IV designs for cohort studies is particularly problematic, as IV analyses are rarely used in prospective cohort studies because of their limitations. The authors fail to justify their inclusion, and established ROB assessment tools do not endorse this approach. Consequently, all 40 cohort studies were classified as having a ‘serious’ or ‘critical’ ROB due to confounding.
Conversely, the review’s treatment of ecological studies is far more permissive. Despite well-documented limitations, nearly half (48%) were classified as being at ‘low risk’ for confounding. The ROB criteria rely on weak or irrelevant factors, such as the ‘parallel trends assumption’ (which cannot be empirically verified) and ‘dose–response’ (which pertains to causal inference rather than confounding). Most concerning is the blanket classification of all ‘natural experiments’ as being at ‘low risk’ for confounding, regardless of whether they take steps to address bias. This creates an illusion of methodological rigour while shielding ecological studies from scrutiny.
What appears to be a potential conflict of interest raises further concern. Two review group members who helped design the ROB criteria are co-authors of 10 of the included ecological studies, eight of which were rated as being at ‘low risk’ for confounding.
Because we did not have the capacity to conduct a complete re-analysis of Berg et al.’s ROB assessments with an appropriate ROB instrument, we cannot be certain how their flawed methodology affected their conclusions. Ultimately, however, this review’s selective and inconsistent design and application of ROB criteria has the potential to distort the scientific landscape. Rather than fairly synthesising the evidence, the review systematically disadvantages cohort studies while presenting ecological studies as more reliable than they truly are. Given the public health implications of e-cigarette regulation, these methodological biases risk shaping policy based on unreliable evidence.
Sam Egger: Conceptualization; investigation; writing—original draft. Martin McKee: Conceptualization; writing—review and editing.
Martin McKee is a past president of the British Medical Association and of the European Public Health Association, both of which have expressed concerns about the public health impact and appropriate regulation of e-cigarettes. Sam Egger is a statistical editor on the editorial board of the Cochrane Breast Cancer Group, serves as a peer reviewer for the Cochrane Database of Systematic Reviews, and has expertise in the relationship between adolescent vaping and future smoking.
期刊介绍:
Addiction publishes peer-reviewed research reports on pharmacological and behavioural addictions, bringing together research conducted within many different disciplines.
Its goal is to serve international and interdisciplinary scientific and clinical communication, to strengthen links between science and policy, and to stimulate and enhance the quality of debate. We seek submissions that are not only technically competent but are also original and contain information or ideas of fresh interest to our international readership. We seek to serve low- and middle-income (LAMI) countries as well as more economically developed countries.
Addiction’s scope spans human experimental, epidemiological, social science, historical, clinical and policy research relating to addiction, primarily but not exclusively in the areas of psychoactive substance use and/or gambling. In addition to original research, the journal features editorials, commentaries, reviews, letters, and book reviews.