Behavioristic, evidentialist, and learning models of statistical testing
Philosophy of Science 52 (4):493-516 (1985)
Abstract
While orthodox (Neyman-Pearson) statistical tests enjoy widespread use in science, the philosophical controversy over their appropriateness for obtaining scientific knowledge remains unresolved. I shall suggest an explanation and a resolution of this controversy. The source of the controversy, I argue, is that orthodox tests are typically interpreted as rules for making optimal decisions as to how to behave--where optimality is measured by the frequency of errors the test would commit in a long series of trials. Most philosophers of statistics, however, view the task of statistical methods as providing appropriate measures of the evidential-strength that data affords hypotheses. Since tests appropriate for the behavioral-decision task fail to provide measures of evidential-strength, philosophers of statistics claim the use of orthodox tests in science is misleading and unjustified. What critics of orthodox tests overlook, I argue, is that the primary function of statistical tests in science is neither to decide how to behave nor to assign measures of evidential strength to hypotheses. Rather, tests provide a tool for using incomplete data to learn about the process that generated it. This they do, I show, by providing a standard for distinguishing differences (between observed and hypothesized results) due to accidental or trivial errors from those due to systematic or substantively important discrepancies. I propose a reinterpretation of a commonly used orthodox test to make this learning model of tests explicitAuthor's Profile
DOI
10.1086/289272
My notes
Similar books and articles
Did Pearson reject the Neyman-Pearson philosophy of statistics?Deborah G. Mayo - 1992 - Synthese 90 (2):233 - 262.
Austere Realism and the Worldly Assumptions of Inferential Statistics.J. D. Trout - 1994 - PSA: Proceedings of the Biennial Meeting of the Philosophy of Science Association 1994:190 - 199.
Mathematical statistics and metastatistical analysis.Andrés Rivadulla - 1991 - Erkenntnis 34 (2):211 - 236.
Error statistics and learning from error: Making a virtue of necessity.Deborah G. Mayo - 1997 - Philosophy of Science 64 (4):212.
Of Nulls and Norms.Peter Godfrey-Smith - 1994 - PSA: Proceedings of the Biennial Meeting of the Philosophy of Science Association 1994:280 - 290.
Severe testing as a basic concept in a neyman–pearson philosophy of induction.Deborah G. Mayo & Aris Spanos - 2006 - British Journal for the Philosophy of Science 57 (2):323-357.
Analytics
Added to PP
2009-01-28
Downloads
203 (#63,565)
6 months
4 (#186,020)
2009-01-28
Downloads
203 (#63,565)
6 months
4 (#186,020)
Historical graph of downloads
Author's Profile
Citations of this work
Severe testing as a basic concept in a neyman–pearson philosophy of induction.Deborah G. Mayo & Aris Spanos - 2006 - British Journal for the Philosophy of Science 57 (2):323-357.
A New Proof of the Likelihood Principle.Greg Gandenberger - 2015 - British Journal for the Philosophy of Science 66 (3):475-503.
Duhem's problem, the bayesian way, and error statistics, or "what's belief got to do with it?".Deborah G. Mayo - 1997 - Philosophy of Science 64 (2):222-244.
A Visible Hand in the Marketplace of Ideas: Precision Measurement as Arbitage.Philip Mirowski - 1994 - Science in Context 7 (3):563-589.
References found in this work
The Enterprise of Knowledge: An Essay on Knowledge, Credal Probability, and Chance.Isaac Levi - 1980 - MIT Press.
The Logical Foundations of Probability.Rudolf Carnap - 1950 - Journal of Philosophy 60 (13):362-364.
Logical Foundations of Probability.Ernest H. Hutten - 1950 - Journal of Symbolic Logic 16 (3):205-207.