Graduate studies at Western
Synthese 16 (3-4):344 - 380 (1966)
|Abstract||It is argued that current attempts to model human learning behavior commonly fail on one of two counts: either the model assumptions are artificially restricted so as to permit the application of mathematical techniques in deriving their consequences, or else the required complex assumptions are imbedded in computer programs whose technical details obscure the theoretical content of the model. The first failing is characteristic of so-called mathematical models of learning, while the second is characteristic of computer simulation models. An approach to model building which avoids both these failings is presented under the title of a black-box theory of learning. This method permits the statement of assumptions of any desired complexity in a language which clearly exhibits their theoretical content.Section II of the paper is devoted to the problem of testing and comparing alternative learning theories. The policy advocated is to abandon attempts at hypothesis testing. It is argued that, in general, we not only lack sufficient data and sufficiently powerful techniques to test hypotheses, but that the truth of a model is not really the issue of basic interest. A given model may be true in the sense that on the basis of available evidence we cannot statistically reject it, but not interesting in the sense that it provides little information about the processes underlying behavior. Rather, we should accept or reject models on the basis of how much information they provide about the way in which subjects respond to environmental structure. This attitude toward model testing is made precise by introducing a formal measure of the information content of a model. Finally, it is argued that the statistical concept of degrees-of-freedom is misleading when used in the context of model testing and should be replaced by a measure of the information absorbed from the data in estimating parameters.|
|Keywords||No keywords specified (fix it)|
No categories specified
(categorize this paper)
|Through your library||Configure|
Similar books and articles
Paul Thagard (1986). Computational Models in the Philosophy of Science. PSA: Proceedings of the Biennial Meeting of the Philosophy of Science Association 1986:329 - 335.
Axel Cleeremans (1993). Mechanisms of Implicit Learning: Connectionist Models of Sequence Processing. MIT Press.
Charles T. Snowdon (2000). Bottoms-Up! A Refreshing Change in Models. Behavioral and Brain Sciences 23 (2):266-267.
Roman Frigg (2008). Models in Science. In Edward N. Zalta (ed.), The Stanford Encyclopedia of Philosophy.
William Timberlake, Norman Pecoraro & Matthew Tinsley (2000). An Integrative Approach to the Modeling of Behavior. Behavioral and Brain Sciences 23 (2):268-268.
Stephan Hartmann & Roman Frigg (2006). Models in Science. In Ed Zalta (ed.), The Stanford Encyclopedia of Philosophy. Stanford.
Jukka Corander & Pekka Marttinen (2006). Bayesian Model Learning Based on Predictive Entropy. Journal of Logic, Language and Information 15 (1-2):5-20.
Sorry, there are not enough data points to plot this chart.
Added to index2009-01-28
Recent downloads (6 months)0
How can I increase my downloads?