The problem with data

IF 0.2 Q4 EVOLUTIONARY BIOLOGY
L. Aarssen
{"title":"The problem with data","authors":"L. Aarssen","doi":"10.4033/IEE.2015.8.15.E","DOIUrl":null,"url":null,"abstract":"The progress of science requires inspiration. Some researchers find this only from data. \"Show me the evidence\", they say. Many peer-reviewed publications in science, however, have no data. They involve a different kind of inspiration: proposals for original ideas or new hypothesis development. These are found within the 'Forum', 'Perspectives', 'Opinion' and 'Commentary' sections of many journals, and in some journals, like IEE, devoted entirely to new ideas and commentary. I have always been particularly drawn to the honesty and beauty in this creative brand of enquiry. And so I am puzzled to hear it often dismissed out of hand with pejorative labeling, like ‘hand-waving’ and ‘just-sostories’. Many—especially among the elites and selfappointed guardians of established theory—would have us believe that only ‘evidence-based’ practice and products can be taken seriously as legitimate sources of inspiration and discovery. This is plainly arrogant and wrongheaded. The scientific method means doing whatever is necessary to get good answers to questions worth asking. And so data collection that is not guided by interesting, novel, and important ideas is usually boring at best. At worst, it is a waste of research grant funds. But published data are plagued with an even more serious problem: we never know how much to trust them. A few minutes of Google searching under the terms \"research bias\", \"scientific misconduct\", \"publication bias\", and “retractions” shows that the follies of faith in published data have come sharply and painfully into the public spotlight in recent years. The latest bad news is particularly troubling: most published studies are not reproducible (Baker 2015, Bartlett 2015, Begley et al 2015, Jump 2015). The statistical implication from this is unavoidable: it means that the results of at least half of all empirical research that has ever been published, probably in all fields of study, are inconclusive at best. They may be reliable and useful, but maybe not. Mounting evidence in fact leans toward the latter (Ioannidis 2005, Lehrer 2010,Hayden 2013). Moreover, these inconclusive reports, I suspect, are likely to involve mostly those that had been regarded as especially promising contributions—lauded as particularly novel and ground-breaking. In contrast, the smaller group that passed the reproducibility test is likely to involve mostly esoteric research that few people care about, or so-called ‘safe research’: studies that report merely confirmatory results, designed to generate data that were already categorically expected, i.e. studies that aimed to provide just another example of support for well-established theory—or if not the latter, support for something that was already an obvious bet or easily believable anyway, even without data collection (or theory). A study that anticipates only positive results in advance is pointless. There is no reason for doing the science in the first place; it just confirms what one already knows must be true. This probably accounts for why the majority of published research remains uncited in the literature—or virtually so, attracting only a small handful of citations, many (or most) of which are selfcitations (Bauerlein et al. 2010). Are there any remedies for this reproducibility problem? Undoubtedly some, and researchers are scrambling, ramping up efforts to identify them [see Nature Special (2015) on Challenges in Irreproducible Research, http://www.nature.com/news/reproducibility1.17552]. Addressing them effectively (if it is possible at all) will require nothing short of a complete restructuring of the culture of science, with new and revised manuals of ‘best practice’ (e.g. see Nosek et al.","PeriodicalId":42755,"journal":{"name":"Ideas in Ecology and Evolution","volume":"8 1","pages":""},"PeriodicalIF":0.2000,"publicationDate":"2015-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Ideas in Ecology and Evolution","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.4033/IEE.2015.8.15.E","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q4","JCRName":"EVOLUTIONARY BIOLOGY","Score":null,"Total":0}
引用次数: 0

Abstract

The progress of science requires inspiration. Some researchers find this only from data. "Show me the evidence", they say. Many peer-reviewed publications in science, however, have no data. They involve a different kind of inspiration: proposals for original ideas or new hypothesis development. These are found within the 'Forum', 'Perspectives', 'Opinion' and 'Commentary' sections of many journals, and in some journals, like IEE, devoted entirely to new ideas and commentary. I have always been particularly drawn to the honesty and beauty in this creative brand of enquiry. And so I am puzzled to hear it often dismissed out of hand with pejorative labeling, like ‘hand-waving’ and ‘just-sostories’. Many—especially among the elites and selfappointed guardians of established theory—would have us believe that only ‘evidence-based’ practice and products can be taken seriously as legitimate sources of inspiration and discovery. This is plainly arrogant and wrongheaded. The scientific method means doing whatever is necessary to get good answers to questions worth asking. And so data collection that is not guided by interesting, novel, and important ideas is usually boring at best. At worst, it is a waste of research grant funds. But published data are plagued with an even more serious problem: we never know how much to trust them. A few minutes of Google searching under the terms "research bias", "scientific misconduct", "publication bias", and “retractions” shows that the follies of faith in published data have come sharply and painfully into the public spotlight in recent years. The latest bad news is particularly troubling: most published studies are not reproducible (Baker 2015, Bartlett 2015, Begley et al 2015, Jump 2015). The statistical implication from this is unavoidable: it means that the results of at least half of all empirical research that has ever been published, probably in all fields of study, are inconclusive at best. They may be reliable and useful, but maybe not. Mounting evidence in fact leans toward the latter (Ioannidis 2005, Lehrer 2010,Hayden 2013). Moreover, these inconclusive reports, I suspect, are likely to involve mostly those that had been regarded as especially promising contributions—lauded as particularly novel and ground-breaking. In contrast, the smaller group that passed the reproducibility test is likely to involve mostly esoteric research that few people care about, or so-called ‘safe research’: studies that report merely confirmatory results, designed to generate data that were already categorically expected, i.e. studies that aimed to provide just another example of support for well-established theory—or if not the latter, support for something that was already an obvious bet or easily believable anyway, even without data collection (or theory). A study that anticipates only positive results in advance is pointless. There is no reason for doing the science in the first place; it just confirms what one already knows must be true. This probably accounts for why the majority of published research remains uncited in the literature—or virtually so, attracting only a small handful of citations, many (or most) of which are selfcitations (Bauerlein et al. 2010). Are there any remedies for this reproducibility problem? Undoubtedly some, and researchers are scrambling, ramping up efforts to identify them [see Nature Special (2015) on Challenges in Irreproducible Research, http://www.nature.com/news/reproducibility1.17552]. Addressing them effectively (if it is possible at all) will require nothing short of a complete restructuring of the culture of science, with new and revised manuals of ‘best practice’ (e.g. see Nosek et al.
数据的问题
科学的进步需要灵感。一些研究人员仅从数据中发现了这一点。“出示证据”,他们说。然而,许多同行评议的科学出版物没有数据。它们涉及一种不同的灵感:对原始想法或新假设发展的建议。在许多期刊的“论坛”、“观点”、“观点”和“评论”部分,以及在一些期刊(如IEE)中,它们完全致力于新思想和评论。我一直特别被这种创造性探索中的诚实和美丽所吸引。因此,我很困惑地听到它经常被立即贴上贬义的标签,比如“挥手”和“只是一些故事”。许多人——尤其是精英和自封的既有理论的捍卫者——会让我们相信,只有“基于证据的”实践和产品才能被认真地视为灵感和发现的合法来源。这显然是傲慢和错误的。科学的方法意味着做任何必要的事情来得到值得问的问题的好答案。因此,没有有趣、新颖和重要想法的数据收集通常最多是无聊的。最坏的情况是,这是对研究经费的浪费。但已公布的数据还存在一个更严重的问题:我们永远不知道该在多大程度上信任它们。在谷歌上搜索几分钟“研究偏见”、“科学不端行为”、“发表偏见”和“撤回”就会发现,近年来,对已发表数据的愚蠢信仰已经尖锐而痛苦地成为公众关注的焦点。最新的坏消息尤其令人不安:大多数发表的研究都是不可重复的(Baker 2015, Bartlett 2015, Begley et al 2015, Jump 2015)。由此产生的统计含义是不可避免的:这意味着,至少有一半已发表的实证研究(可能涉及所有研究领域)的结果充其量是不确定的。它们可能是可靠和有用的,但也可能不是。事实上,越来越多的证据倾向于后者(Ioannidis 2005, Lehrer 2010,Hayden 2013)。此外,我怀疑,这些不确定的报告可能主要涉及那些被视为特别有前途的贡献——被称赞为特别新颖和突破性的。相比之下,通过可重复性测试的小群体可能主要涉及很少有人关心的深奥研究,或所谓的“安全研究”:仅报告证实性结果的研究,旨在产生已经明确预期的数据,即旨在为已确立的理论提供支持的另一个例子的研究——或者,如果不是后者,支持已经是一个明显的打赌或很容易相信的东西,即使没有数据收集(或理论)。一项事先只预测积极结果的研究是毫无意义的。一开始就没有理由做科学研究;它只是证实了一个人已经知道的一定是真的。这可能解释了为什么大多数已发表的研究在文献中仍然没有被引用-或者实际上是这样,只吸引了少数引用,其中许多(或大多数)是自我引用(Bauerlein et al. 2010)。对于这种再现性问题有什么补救办法吗?毫无疑问,有一些是不可复制的,研究人员正在争分夺秒地努力识别它们[参见Nature Special (2015) on Challenges in reproducibility Research, http://www.nature.com/news/reproducibility1.17552]。有效地解决这些问题(如果有可能的话)将需要完全重构科学文化,使用新的和修订的“最佳实践”手册(例如见Nosek等人)。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Ideas in Ecology and Evolution
Ideas in Ecology and Evolution EVOLUTIONARY BIOLOGY-
自引率
0.00%
发文量
4
审稿时长
36 weeks
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信