Synthese 16 (3-4):344 - 380 (1966)
It is argued that current attempts to model human learning behavior commonly fail on one of two counts: either the model assumptions are artificially restricted so as to permit the application of mathematical techniques in deriving their consequences, or else the required complex assumptions are imbedded in computer programs whose technical details obscure the theoretical content of the model. The first failing is characteristic of so-called mathematical models of learning, while the second is characteristic of computer simulation models. An approach to model building which avoids both these failings is presented under the title of a black-box theory of learning. This method permits the statement of assumptions of any desired complexity in a language which clearly exhibits their theoretical content.Section II of the paper is devoted to the problem of testing and comparing alternative learning theories. The policy advocated is to abandon attempts at hypothesis testing. It is argued that, in general, we not only lack sufficient data and sufficiently powerful techniques to test hypotheses, but that the truth of a model is not really the issue of basic interest. A given model may be true in the sense that on the basis of available evidence we cannot statistically reject it, but not interesting in the sense that it provides little information about the processes underlying behavior. Rather, we should accept or reject models on the basis of how much information they provide about the way in which subjects respond to environmental structure. This attitude toward model testing is made precise by introducing a formal measure of the information content of a model. Finally, it is argued that the statistical concept of degrees-of-freedom is misleading when used in the context of model testing and should be replaced by a measure of the information absorbed from the data in estimating parameters.
|Keywords||No keywords specified (fix it)|
|Categories||categorize this paper)|
References found in this work BETA
A Study of Thinking.Jerome S. Bruner, Jacqueline J. Goodnow & George A. Austin - 1958 - Philosophy and Phenomenological Research 19 (1):118-119.
Citations of this work BETA
Explanation, Prediction, Description, and Information Theory.Joseph F. Hanna - 1969 - Synthese 20 (3):308 - 334.
Theory Construction in Psychology: The Interpretation and Integration of Psychological Data.Gordon M. Becker - 1981 - Theory and Decision 13 (3):251.
Statistics, Induction, and Lawlikeness: Comments on Dr. Vetter's Paper.Jaakko Hintikka - 1969 - Synthese 20 (1):72 - 83.
Philosophy of Science (Wissenschaftstheorie) in Finland.Jaakko Hintikka - 1970 - Journal for General Philosophy of Science / Zeitschrift für Allgemeine Wissenschaftstheorie 1 (1):119-132.
Similar books and articles
Computational Models in the Philosophy of Science.Paul Thagard - 1986 - PSA: Proceedings of the Biennial Meeting of the Philosophy of Science Association 1986:329 - 335.
Mechanisms of Implicit Learning: Connectionist Models of Sequence Processing.Axel Cleeremans - 1993 - MIT Press.
Bottoms-Up! A Refreshing Change in Models.Charles T. Snowdon - 2000 - Behavioral and Brain Sciences 23 (2):266-267.
Models in Science.Roman Frigg - 2008 - In Edward N. Zalta (ed.), The Stanford Encyclopedia of Philosophy.
An Integrative Approach to the Modeling of Behavior.William Timberlake, Norman Pecoraro & Matthew Tinsley - 2000 - Behavioral and Brain Sciences 23 (2):268-268.
Models in Science.Stephan Hartmann & Roman Frigg - 2006 - In Ed Zalta (ed.), The Stanford Encyclopedia of Philosophy. Stanford.
Bayesian Model Learning Based on Predictive Entropy.Jukka Corander & Pekka Marttinen - 2006 - Journal of Logic, Language and Information 15 (1-2):5-20.
Added to index2009-01-28
Total downloads23 ( #218,742 of 2,169,644 )
Recent downloads (6 months)2 ( #186,189 of 2,169,644 )
How can I increase my downloads?