Abstract
Scientific inquiry involves observations and measurements, some of which are planned and some of which are not. The most interesting or unusual observations might be regarded as discoveries and therefore particularly worthy of publication. However, the observational process is fraught with inferential land mines, especially if the discoveries are serendipitous. Multiple observations increase the probability of false-positive conclusions and have led to many false and otherwise misleading publications. Statisticians recommend adjustments to final inferences with the goal of reducing the rate of false positives, a strategy that increases the rate of false negatives. Some scientists object to making such adjustments, arguing that it should not be more difficult to determine the validity of a discovery simply because other observations were made. Which tack is right? How does one decide that any particular scientific discovery is real? Unfortunately, there is no panacea, no one-size-fits-all approach. The goal of this commentary is to elucidate the issues and provide recommendations for conducting and reporting results of empirical studies, with emphasis on the problems of multiple comparisons and other types of multiplicities, including what I call "silent multiplicities." Because of the many observations, outcomes, subsets, treatments, etc, that are typically made or addressed in epidemiology and biomarker research, these recommendations may be particularly relevant for such studies. However, the lessons apply quite generally. I consider both frequentist and Bayesian statistical approaches.
Keywords
MeSH Terms
Affiliated Institutions
Related Publications
The advantages of the Matthews correlation coefficient (MCC) over F1 score and accuracy in binary classification evaluation
Abstract Background To evaluate binary classifications and their confusion matrices, scientific researchers can employ several statistical rates, accordingly to the goal of the ...
Enrichment of High-Throughput Screening Data with Increasing Levels of Noise Using Support Vector Machines, Recursive Partitioning, and Laplacian-Modified Naive Bayesian Classifiers
High-throughput screening (HTS) plays a pivotal role in lead discovery for the pharmaceutical industry. In tandem, cheminformatics approaches are employed to increase the probab...
Bayesian Inference for Genomic Data Integration Reduces Misclassification Rate in Predicting Protein-Protein Interactions
Protein-protein interactions (PPIs) are essential to most fundamental cellular processes. There has been increasing interest in reconstructing PPIs networks. However, several cr...
Controlling the False Discovery Rate: A Practical and Powerful Approach to Multiple Testing
SUMMARY The common approach to the multiplicity problem calls for controlling the familywise error rate (FWER). This approach, though, has faults, and we point out a few. A diff...
Operating Characteristics and Extensions of the False Discovery Rate Procedure
Summary We investigate the operating characteristics of the Benjamini–Hochberg false discovery rate procedure for multiple testing. This is a distribution-free method that contr...
Publication Info
- Year
- 2012
- Type
- article
- Volume
- 104
- Issue
- 15
- Pages
- 1125-1133
- Citations
- 51
- Access
- Closed
External Links
Social Impact
Social media, news, blog, policy document mentions
Citation Metrics
Cite This
Identifiers
- DOI
- 10.1093/jnci/djs301
- PMID
- 22859849
- PMCID
- PMC4614276