Full text

Turn on search term navigation

© 2021 Bridgeford et al. This is an open access article distributed under the terms of the Creative Commons Attribution License: http://creativecommons.org/licenses/by/4.0/ (the “License”), which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited. Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.

Abstract

Replicability, the ability to replicate scientific findings, is a prerequisite for scientific discovery and clinical utility. Troublingly, we are in the midst of a replicability crisis. A key to replicability is that multiple measurements of the same item (e.g., experimental sample or clinical participant) under fixed experimental constraints are relatively similar to one another. Thus, statistics that quantify the relative contributions of accidental deviations—such as measurement error—as compared to systematic deviations—such as individual differences—are critical. We demonstrate that existing replicability statistics, such as intra-class correlation coefficient and fingerprinting, fail to adequately differentiate between accidental and systematic deviations in very simple settings. We therefore propose a novel statistic, discriminability, which quantifies the degree to which an individual’s samples are relatively similar to one another, without restricting the data to be univariate, Gaussian, or even Euclidean. Using this statistic, we introduce the possibility of optimizing experimental design via increasing discriminability and prove that optimizing discriminability improves performance bounds in subsequent inference tasks. In extensive simulated and real datasets (focusing on brain imaging and demonstrating on genomics), only optimizing data discriminability improves performance on all subsequent inference tasks for each dataset. We therefore suggest that designing experiments and analyses to optimize discriminability may be a crucial step in solving the replicability crisis, and more generally, mitigating accidental measurement error.

Details

Title
Eliminating accidental deviations to minimize generalization error and maximize replicability: Applications in connectomics and genomics
Author
Bridgeford, Eric W  VIAFID ORCID Logo  ; Wang, Shangsi  VIAFID ORCID Logo  ; Wang, Zeyi  VIAFID ORCID Logo  ; Xu, Ting  VIAFID ORCID Logo  ; Craddock, Cameron  VIAFID ORCID Logo  ; Dey, Jayanta  VIAFID ORCID Logo  ; Kiar, Gregory  VIAFID ORCID Logo  ; Gray-Roncal, William  VIAFID ORCID Logo  ; Colantuoni, Carlo; Douville, Christopher  VIAFID ORCID Logo  ; Noble, Stephanie  VIAFID ORCID Logo  ; Priebe, Carey E  VIAFID ORCID Logo  ; Caffo, Brian; Milham, Michael  VIAFID ORCID Logo  ; Xi-Nian Zuo  VIAFID ORCID Logo  ; Consortium for Reliability and Reproducibility; Vogelstein, Joshua T  VIAFID ORCID Logo 
First page
e1009279
Section
Research Article
Publication year
2021
Publication date
Sep 2021
Publisher
Public Library of Science
ISSN
1553734X
e-ISSN
15537358
Source type
Scholarly Journal
Language of publication
English
ProQuest document ID
2582586863
Copyright
© 2021 Bridgeford et al. This is an open access article distributed under the terms of the Creative Commons Attribution License: http://creativecommons.org/licenses/by/4.0/ (the “License”), which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited. Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.