Eliminating accidental deviations to minimize generalization error and maximize replicability: Applications in connectomics and genomics

Author:

Bridgeford Eric W.ORCID,Wang ShangsiORCID,Wang ZeyiORCID,Xu TingORCID,Craddock CameronORCID,Dey JayantaORCID,Kiar GregoryORCID,Gray-Roncal WilliamORCID,Colantuoni Carlo,Douville ChristopherORCID,Noble StephanieORCID,Priebe Carey E.ORCID,Caffo Brian,Milham MichaelORCID,Zuo Xi-NianORCID,Vogelstein Joshua T.ORCID,

Abstract

Replicability, the ability to replicate scientific findings, is a prerequisite for scientific discovery and clinical utility. Troublingly, we are in the midst of a replicability crisis. A key to replicability is that multiple measurements of the same item (e.g., experimental sample or clinical participant) under fixed experimental constraints are relatively similar to one another. Thus, statistics that quantify the relative contributions of accidental deviations—such as measurement error—as compared to systematic deviations—such as individual differences—are critical. We demonstrate that existing replicability statistics, such as intra-class correlation coefficient and fingerprinting, fail to adequately differentiate between accidental and systematic deviations in very simple settings. We therefore propose a novel statistic, discriminability, which quantifies the degree to which an individual’s samples are relatively similar to one another, without restricting the data to be univariate, Gaussian, or even Euclidean. Using this statistic, we introduce the possibility of optimizing experimental design via increasing discriminability and prove that optimizing discriminability improves performance bounds in subsequent inference tasks. In extensive simulated and real datasets (focusing on brain imaging and demonstrating on genomics), only optimizing data discriminability improves performance on all subsequent inference tasks for each dataset. We therefore suggest that designing experiments and analyses to optimize discriminability may be a crucial step in solving the replicability crisis, and more generally, mitigating accidental measurement error.

Funder

National Science Foundation

National Institutes of Health

Defense Advanced Research Projects Agency

Microsoft Research

National Basic Research Program of China

Natural Science Foundation of China

China Netherlands CAS-NWO Programme

Beijing Municipal Science and Tech Commission

Start-up Funds for Leading Talents at Beijing Normal University, the Major Project of National Social Science Foundation of China

National Basic Science Data Center “Chinese Data-sharing Warehouse for In-vivo Imaging Brain”

Guangxi Bagui Scholarship

Publisher

Public Library of Science (PLoS)

Subject

Computational Theory and Mathematics,Cellular and Molecular Neuroscience,Genetics,Molecular Biology,Ecology,Modeling and Simulation,Ecology, Evolution, Behavior and Systematics

Reference59 articles.

1. The Proof and Measurement of Association between Two Things;C Spearman;Am J Psychol,1904

2. Tackling the widespread and critical impact of batch effects in high-throughput data;JT Leek;Nat Rev Genet,2010

3. Statistics: P values are just the tip of the iceberg;JT Leek;Nature,2015

4. National Academies of Sciences E. Reproducibility and Replicability in Science; 2019.

5. What does research reproducibility mean?;SN Goodman;Sci Transl Med,2016

Cited by 24 articles. 订阅此论文施引文献 订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献

全球学者库

1.学者识别学者识别

2.学术分析学术分析

3.人才评估人才评估

"全球学者库"是以全球学者为主线,采集、加工和组织学术论文而形成的新型学术文献查询和分析系统,可以对全球学者进行文献检索和人才价值评估。用户可以通过关注某些学科领域的顶尖人物而持续追踪该领域的学科进展和研究前沿。经过近期的数据扩容,当前全球学者库共收录了国内外主流学术期刊6万余种,收集的期刊论文及会议论文总量共计约1.5亿篇,并以每天添加12000余篇中外论文的速度递增。我们也可以为用户提供个性化、定制化的学者数据。欢迎来电咨询!咨询电话:010-8811{复制后删除}0370

www.globalauthorid.com

TOP

Copyright © 2019-2023 北京同舟云网络信息技术有限公司
京公网安备11010802033243号  京ICP备18003416号-3