How to Discount Double-Counting When It Counts: Some Clarifications

British Journal for the Philosophy of Science 59 (4):857-879 (2008)
  Copy   BIBTEX

Abstract

The issues of double-counting, use-constructing, and selection effects have long been the subject of debate in the philosophical as well as statistical literature. I have argued that it is the severity, stringency, or probativeness of the test—or lack of it—that should determine if a double-use of data is admissible. Hitchcock and Sober ([2004]) question whether this ‘severity criterion' can perform its intended job. I argue that their criticisms stem from a flawed interpretation of the severity criterion. Taking their criticism as a springboard, I elucidate some of the central examples that have long been controversial, and clarify how the severity criterion is properly applied to them.1. Severity and Use-Constructing: Four Points (and Some Clarificatory Notes) 1.1. Point 1: Getting beyond ‘all or nothing’ standpoints1.2. Point 2: The rationale for prohibiting double-counting is the requirement that tests be severe1.3. Point 3: Evaluate severity of a test T by its associated construction rule R1.4. Point 4: The ease of passing vs. ease of erroneous passing: Statistical vs. ‘Definitional’ probability2. The False Dilemma: Hitchcock and Sober 2.1. Marsha measures her desk reliably2.2. A false dilemma3. Canonical Errors of Inference 3.1. How construction rules may alter the error-probing performance of tests3.2. Rules for accounting for anomalies3.3. Hunting for statistically significant differences4. Concluding Remarks.

Links

PhilArchive



    Upload a copy of this work     Papers currently archived: 91,164

External links

Setup an account with your affiliations in order to access resources via your University's proxy server

Through your library

Similar books and articles

How to discount double-counting when it counts: Some clarifications.Deborah G. Mayo - 2008 - British Journal for the Philosophy of Science 59 (4):857-879.
Severe testing as a basic concept in a neyman–pearson philosophy of induction.Deborah G. Mayo & Aris Spanos - 2006 - British Journal for the Philosophy of Science 57 (2):323-357.
Some surprising facts about surprising facts.D. Mayo - 2014 - Studies in History and Philosophy of Science Part A 45:79-86.
Climate Models, Calibration, and Confirmation.Katie Steele & Charlotte Werndl - 2013 - British Journal for the Philosophy of Science 64 (3):609-635.
Climate models, calibration, and confirmation.Charlotte Werndl & Katie Steele - 2013 - British Journal for the Philosophy of Science 64 (3):609-635.
Novel evidence and severe tests.Deborah G. Mayo - 1991 - Philosophy of Science 58 (4):523-552.
Model tuning in engineering: uncovering the logic.Katie Steele & Charlotte Werndl - 2015 - Journal of Strain Analysis for Engineering Design 51 (1):63-71.
What Have I Done?Timothy Chappell - 2013 - Diametros 38:86-111.

Analytics

Added to PP
2017-02-23

Downloads
11 (#1,065,379)

6 months
5 (#510,007)

Historical graph of downloads
How can I increase my downloads?

Author's Profile

Deborah Mayo
Virginia Tech

Citations of this work

The Costs of HARKing.Mark Rubin - 2022 - British Journal for the Philosophy of Science 73 (2):535-560.
Some surprising facts about surprising facts.D. Mayo - 2014 - Studies in History and Philosophy of Science Part A 45:79-86.
Some methodological issues in experimental economics.Deborah Mayo - 2008 - Philosophy of Science 75 (5):633-645.

Add more citations

References found in this work

No references found.

Add more references