Abstract
I argue that John Searle's (1980) influential Chinese room argument (CRA) against computationalism and strong AI survives existing objections, including Block's (1998) internalized systems reply, Fodor's (1991b) deviant causal chain reply, and Hauser's (1997) unconscious content reply. However, a new ``essentialist'' reply I construct shows that the CRA as presented by Searle is an unsound argument that relies on a question-begging appeal to intuition. My diagnosis of the CRA relies on an interpretation of computationalism as a scientific theory about the essential nature of intentional content; such theories often yield non-intuitive results in non-standard cases, and so cannot be judged by such intuitions. However, I further argue that the CRA can be transformed into a potentially valid argument against computationalism simply by reinterpreting it as an indeterminacy argument that shows that computationalism cannot explain the ordinary distinction between semantic content and sheer syntactic manipulation, and thus cannot be an adequate account of content. This conclusion admittedly rests on the arguable but plausible assumption that thought content is interestingly determinate. I conclude that the viability of computationalism and strong AI depends on their addressing the indeterminacy objection, but that it is currently unclear how this objection can be successfully addressed.
- Block, N. (1998), 'The Philosophy of Psychology: Classical Computationalism', in A. C. Grayling, ed., Philosophy 2: Further Through the Subject, New York: Oxford University Press, pp. 5-48.Google Scholar
- Carleton, L. (1984), 'Programs, Language Understanding, and Searle', Synthese 59, pp. 219-233.Google ScholarCross Ref
- Cummins, R. (1989), Meaning and Mental Representation, Cambridge, MA: MIT Press.Google Scholar
- Fodor, J. A. (1991a; orig pub 1980), 'Searle on What Only Brains Can Do', in D. M. Rosenthal, ed., The Nature of Mind, New York: Oxford University Press, pp. 520-521.Google Scholar
- Fodor, J. A. (1991b), 'Afterthoughts: Yin and Yang in the Chinese Room,' in D. M. Rosenthal, ed., The Nature of Mind, New York: Oxford University Press, pp. 524-525.Google Scholar
- Fodor, J. A. (1994), The Elm and the Experat: Mentalese and Its Semantics, Cambridge, MA: MIT Press.Google Scholar
- Hauser, L. (1997), 'Searle's Chinese Box: Debunking the Chinese Room Argument', Minds and Machines 7. pp. 199-226. Google ScholarDigital Library
- Hodges, A. (1983), Alan Turing: The Enigma, New York: Simon and Schuster. Google ScholarDigital Library
- Pinker, S. (1997), How the Mind Works, New York: Norton.Google Scholar
- Quine, W. O. (1960), Word and Object, Cambridge, MA: MIT Press.Google Scholar
- Rapaport, W. J. (1988), ' Semantics: Foundations of Computational Natural Language Understanding'. in J. Fetzer, ed., Aspects of Artificial Intelligence, Dordrecht, The Netherlands: Kluwer Academic Publishers, pp. 81-131.Google ScholarDigital Library
- Searle, J. R. (1980), 'Minds, Brains, and Programs', Behavioral and Brain Sciences 3, pp. 417-424.Google Scholar
- Searle, J. R. (1982), 'The Chinese Room Revisited', Behavioral and Brain Sciences 5, pp. 345-348.Google ScholarCross Ref
- Searle, J. R. (1988), 'Minds and Brains Without Programs.', in C. Blakemore and S. Greenfield, eds., Mindwaves, Oxford: Basil Blackwell, pp. 209-233.Google Scholar
- Searle, J. R. (1990), 'Is the Brain's Mind a Computer Program?', Scientific American 262, pp. 26-31.Google ScholarCross Ref
- Searle, J. R. (1991a; orig pub 1980), 'Minds, Brains, and Programs', in D.M. Rosenthal, ed., The Nature of Mind, New York: Oxford University Press, pp. 509-519.Google Scholar
- Searle, J. R. (1991b; orig pub 1980), 'Author's Response', in D.M. Rosenthal. ed., The Nature of Mind, New York: Oxford University Press, pp. 521-523.Google Scholar
- Searle, J. R. (1991c), 'Yin and Yang Strike Out'. in D. M. Rosenthal, ed., The Nature of Mind, New York: Oxford University Press, (pp. 525-526).Google Scholar
- Searle, J. R. (1992), The Rediscovery of Mind, Cambridge, MA: MIT Press. Google ScholarDigital Library
- Searle, J. R. (1997), The Mystery of Consciousness, New York: New York Review Books.Google Scholar
- Searle, J. R., J. McCarthy, H. Dreyfus, M. Minsky and S. Papert (1984), 'Has Artificial Intelligence Research Illuminated Human Thinking?', Annals of the New York City Academy of Arts and Sciences 426, pp. 138-160.Google Scholar
- Wakefield, J. C. (1992), 'The Concept of Mental Disorder: On the Boundary between Biological Facts and Social Values', American Psychologist 47, pp. 373-388.Google ScholarCross Ref
- Wakefield, J.C. (1995), 'Dysfunction as a Value-free Concept: Reply to Sadler and Agich', Philosophy, Psychiatry, and Psychology 2, pp. 233-246.Google Scholar
- Wakefield, J. C. (1999), 'Disorder as a Black Box Essentialist Concept', Journal of Abnormal Psychology 108, pp. 465-472.Google ScholarCross Ref
- Wakefield, J. C. (2001), Do Unconscious Mental States Exist?: Freud, Searle, and the Conceptual Foundations of Cognitive Science, Berkeley, CA: University of California at Berkeley (Doctoral Dissertation).Google Scholar
- Wakefield, J. C. (in press), 'Fodor on inscrutability', Mind and Language.Google Scholar
- Wilks, Y. (1982), 'Searle's Straw Man', Behavioral and Brain Sciences 5, pp. 344-345.Google ScholarCross Ref
Index Terms
- The Chinese Room Argument Reconsidered: Essentialism, Indeterminacy, and Strong AI
Recommendations
The logic of Searle's Chinese room argument
John Searle's Chinese room argument (CRA) is a celebrated thought experiment designed to refute the hypothesis, popular among artificial intelligence (AI) scientists and philosophers of mind, that "the appropriately programmed computer really is a mind"...
Searle‘s Chinese Box: Debunking the Chinese Room Argument
John Searle‘s Chinese room argument is perhaps the most influential and widely cited argument against artificial intelligence (AI). Understood as targeting AI proper – claims that computers can think or do think – Searle‘s argument, despite its rhetorical ...
The Explanatory Role of Computation in Cognitive Science
Which notion of computation (if any) is essential for explaining cognition? Five answers to this question are discussed in the paper. (1) The classicist answer: symbolic (digital) computation is required for explaining cognition; (2) The broad digital ...
Comments