In the book, I argue that the mind can be explained computationally because it is itself computational—whether it engages in mental arithmetic, parses natural language, or processes the auditory signals that allow us to experience music. All these capacities arise from complex information-processing operations of the mind. By analyzing the state of the art in cognitive science, I develop an account of computational explanation used to explain the capacities in question.
This paper centers around the notion that internal, mental representations are grounded in structural similarity, i.e., that they are so-called S-representations. We show how S-representations may be causally relevant and argue that they are distinct from mere detectors. First, using the neomechanist theory of explanation and the interventionist account of causal relevance, we provide a precise interpretation of the claim that in S-representations, structural similarity serves as a “fuel of success”, i.e., a relation that is exploitable for the representation using (...) system. Then, we discuss crucial differences between S-representations and indicators or detectors, showing that—contrary to claims made in the literature—there is an important theoretical distinction to be drawn between the two. (shrink)
In this article, after presenting the basic idea of causal accounts of implementation and the problems they are supposed to solve, I sketch the model of computation preferred by Chalmers and argue that it is too limited to do full justice to computational theories in cognitive science. I also argue that it does not suffice to replace Chalmers’ favorite model with a better abstract model of computation; it is necessary to acknowledge the causal structure of physical computers that is not (...) accommodated by the models used in computability theory. Additionally, an alternative mechanistic proposal is outlined. (shrink)
The purpose of this paper is to present a general mechanistic framework for analyzing causal representational claims, and offer a way to distinguish genuinely representational explanations from those that invoke representations for honorific purposes. It is usually agreed that rats are capable of navigation because they maintain a cognitive map of their environment. Exactly how and why their neural states give rise to mental representations is a matter of an ongoing debate. I will show that anticipatory mechanisms involved in rats’ (...) evaluation of possible routes give rise to satisfaction conditions of contents, and this is why they are representationally relevant for explaining and predicting rats’ behavior. I argue that a naturalistic account of satisfaction conditions of contents answers the most important objections of antirepresentationalists. (shrink)
Multiple realizability (MR) is traditionally conceived of as the feature of computational systems, and has been used to argue for irreducibility of higher-level theories. I will show that there are several ways a computational system may be seen to display MR. These ways correspond to (at least) five ways one can conceive of the function of the physical computational system. However, they do not match common intuitions about MR. I show that MR is deeply interest-related, and for this reason, difficult (...) to pin down exactly. I claim that MR is of little importance for defending computationalism, and argue that it should rather appeal to organizational invariance or substrate neutrality of computation, which are much more intuitive but cannot support strong antireductionist arguments. (shrink)
In this paper, I argue that even if the Hard Problem of Content, as identified by Hutto and Myin, is important, it was already solved in natu- ralized semantics, and satisfactory solutions to the problem do not rely merely on the notion of information as covariance. I point out that Hutto and Myin have double standards for linguistic and mental representation, which leads to a peculiar inconsistency. Were they to apply the same standards to basic and linguistic minds, they would (...) either have to embrace representationalism or turn to semantic nihilism, which is, as I argue, an unstable and unattractive position. Hence, I conclude, their book does not offer an alternative to representation- alism. At the same time, it reminds us that representational talk in cognitive science cannot be taken for granted and that information is different from men- tal representation. Although this claim is not new, Hutto and Myin defend it forcefully and elegantly. (shrink)
In this paper, I focus on a problem related to teleological theories of content namely, which notion of function makes content causally relevant? It has been claimed that some functional accounts of content make it causally irrelevant, or epiphenomenal; in which case, such notions of function could no longer act as the pillar of naturalized semantics. By looking closer at biological questions about behavior, I argue that past discussion has been oriented towards an ill-posed question. What I defend is a (...) Very Boring Hypothesis: depending on the representational phenomenon and the explanatory question, different aspects might be important, and it is difficult to say a priori which ones these might be. There are multiple facets to biological functionality and causality relevant for explaining representational phenomena, and ignoring them will lead to unmotivated simplifications. In addition, accounting for different facets of functionality helps dispense with intuition-based specifications of cognitive phenomena. (shrink)
The paper defends the claim that the mechanistic explanation of information processing is the fundamental kind of explanation in cognitive science. These mechanisms are complex organized systems whose functioning depends on the orchestrated interaction of their component parts and processes. A constitutive explanation of every mechanism must include both appeal to its environment and to the role it plays in it. This role has been traditionally dubbed competence. To fully explain how this role is played it is necessary to explain (...) the information processing inside the mechanism embedded in the environment. The most usual explanation on this level has a form of a computational model, for example a software program or a trained artificial neural network. However, this is not the end of the explanatory chain. What is left to be explained is how the program is realized (or what processes are responsible for information processing in the artificial neural network). By using two dramatically different examples from the history of cognitive science I show the multi-level structure of explanations in cognitive science. These examples are (1) the explanation of human process solving as proposed by A. Newell & H. Simon; (2) the explanation of cricket phonotaxis via robotic models by B. Webb. (shrink)
Cognitive science is an interdisciplinary conglomerate of various research fields and disciplines, which increases the risk of fragmentation of cognitive theories. However, while most previous work has focused on theoretical integration, some kinds of integration may turn out to be monstrous, or result in superficially lumped and unrelated bodies of knowledge. In this paper, I distinguish theoretical integration from theoretical unification, and propose some analyses of theoretical unification dimensions. Moreover, two research strategies that are supposed to lead to unification are (...) analyzed in terms of the mechanistic account of explanation. Finally, I argue that theoretical unification is not an absolute requirement from the mechanistic perspective, and that strategies aiming at unification may be premature in fields where there are multiple conflicting explanatory models. (shrink)
Nietzsche's treatment of Epicurus is an interesting example of philosophical hermeneutics. Epicurus bas tren notoriously misinterpreted, claims Nietzsche, because bis mask bas been taken for bis true face. Traditionally Epicurus is presented as a utilitarian or hedonist avant la lettre. This is a simplification motivated by a desire to deprecate bis philosophy. To Nietzsche Epicurus was „an idyllic hero”, a teacher with anistocratic predilections aun his own concept of good, critical of the traditional form of religion, and of the „pre-existent (...) form of Cbristianity”. As a hedonist he was much less convincing, as he was afraid of both pain and pleasure. He assumed the mask of an epicure in order to hide bis true self. Nietzsche warn us over and over again not to trust the traditional interpretation of Epicurus and urges us to penetrate beyond the veil of a stylish disguise. (shrink)
In this paper, the role of the environment and physical embodiment of computational systems for explanatory purposes will be analyzed. In particular, the focus will be on cognitive computational systems, understood in terms of mechanisms that manipulate semantic information. It will be argued that the role of the environment has long been appreciated, in particular in the work of Herbert A. Simon, which has inspired the mechanistic view on explanation. From Simon’s perspective, the embodied view on cognition seems natural but (...) it is nowhere near as critical as its proponents suggest. The only point of difference between Simon and embodied cognition is the significance of body-based off-line cognition; however, it will be argued that it is notoriously over-appreciated in the current debate. The new mechanistic view on explanation suggests that even if it is critical to situate a mechanism in its environment and study its physical composition, or realization, it is also stressed that not all detail counts, and that some bodily features of cognitive systems should be left out from explanations. (shrink)
I discuss whether there are some lessons for philosophical inquiry over the nature of simulation to be learnt from the practical methodology of reengineering. I will argue that reengineering serves a similar purpose as simulations in theoretical science such as computational neuroscience or neurorobotics, and that the procedures and heuristics of reengineering help to develop solutions to outstanding problems of simulation.
The paper proposes an empirical method to investigate linguistic prescriptions as inherent corrective behaviors. The behaviors in question may but need not necessarily be supported by any explicit knowledge of rules. It is possible to gain insight into them, for example by extracting information about corrections from revision histories of texts (or by analyzing speech corpora where users correct themselves or one another). One easily available source of such information is the revision history of Wikipedia. As is shown, the most (...) frequent and short corrections are limited to linguistic errors such as typos (and editorial conventions adopted in Wikipedia). By perusing an automatically generated revision corpus, one gains insight into the prescriptive nature of language empirically. At the same time, the prescriptions offered are not reducible to descriptions of the most frequent linguistic use. (shrink)
Artificial models of cognition serve different purposes, and their use determines the way they should be evaluated. There are also models that do not represent any particular biological agents, and there is controversy as to how they should be assessed. At the same time, modelers do evaluate such models as better or worse. There is also a widespread tendency to call for publicly available standards of replicability and benchmarking for such models. In this paper, I argue that proper evaluation ofmodels (...) does not depend on whether they target real biological agents or not; instead, the standards of evaluation depend on the use of models rather than on the reality of their targets. I discuss how models are validated depending on their use and argue that all-encompassing benchmarks for models may be well beyond reach. (shrink)
In most accounts of realization of computational processes by physical mechanisms, it is presupposed that there is one-to-one correspondence between the causally active states of the physical process and the states of the computation. Yet such proposals either stipulate that only one model of computation is implemented, or they do not reflect upon the variety of models that could be implemented physically. -/- In this paper, I claim that mechanistic accounts of computation should allow for a broad variation of models (...) of computation. In particular, some non-standard models should not be excluded a priori. The relationship between mathematical models of computation and mechanistically adequate models is studied in more detail. (shrink)
The claim defended in the paper is that the mechanistic account of explanation can easily embrace idealization in big-scale brain simulations, and that only causally relevant detail should be present in explanatory models. The claim is illustrated with two methodologically different models: Blue Brain, used for particular simulations of the cortical column in hybrid models, and Eliasmith’s SPAUN model that is both biologically realistic and able to explain eight different tasks. By drawing on the mechanistic theory of computational explanation, I (...) argue that large-scale simulations require that the explanandum phenomenon is identified; otherwise, the explanatory value of such explanations is difficult to establish, and testing the model empirically by comparing its behavior with the explanandum remains practically impossible. The completeness of the explanation, and hence of the explanatory value of the explanatory model, is to be assessed vis-à-vis the explanandum phenomenon, which is not to be conflated with raw observational data and may be idealized. I argue that idealizations, which include building models of a single phenomenon displayed by multi-functional mechanisms, lumping together multiple factors in a single causal variable, simplifying the causal structure of the mechanisms, and multi-model integration, are indispensable for complex systems such as brains; otherwise, the model may be as complex as the explanandum phenomenon, which would make it prone to so-called Bonini paradox. I conclude by enumerating dimensions of empirical validation of explanatory models according to new mechanism, which are given in a form of a “checklist” for a modeler. (shrink)
Is there a field of social intelligence? Many various disciplines ap-proach the subject and it may only seem natural to suppose that different fields of study aim at explaining different phenomena; in other words, there is no spe-cial field of study of social intelligence. In this paper, I argue for an opposite claim. Namely, there is a way to integrate research on social intelligence, as long as one accepts the mechanistic account to explanation. Mechanistic inte-gration of different explanations, however, comes (...) at a cost: mechanism requires explanatory models to be fairly complete and realistic, and this does not seem to be the case for many models concerning social intelligence, especially models of economical behavior. Such models need either be made more realistic, or they would not count as contributing to the same field. I stress that the focus on integration does not lead to ruthless reductionism; on the contrary, mechanistic explanations are best understood as explanatorily pluralistic. (shrink)
Many philosophers use “physicalism” and “naturalism” interchangeably. In this paper, I will distinguish ontological naturalism from physicalism. While broad versions of physicalism are compatible with naturalism, naturalism doesn't have to be committed to strong versions of physical reductionism, so it cannot be defined as equivalent to it. Instead of relying on the notion of ideal physics, naturalism can refer to the notion of ideal natural science that doesn't imply unity of science. The notion of ideal natural science, as well as (...) the notion of ideal physics, will be vindicated. I will shortly explicate the notion of ideal natural science, and define ontological naturalism based on it. (shrink)
In most accounts of realization of computational processes by physical mechanisms, it is presupposed that there is one-to-one correspondence between the causally active states of the physical process and the states of the computation. Yet such proposals either stipulate that only one model of computation is implemented, or they do not reflect upon the variety of models that could be implemented physically. In this paper, I claim that mechanistic accounts of computation should allow for a broad variation of models of (...) computation. In particular, some non-standard models should not be excluded a priori. The relationship between mathematical models of computation and mechanistically adequate models is studied in more detail. (shrink)
In Darwin’s Dangerous Idea, Daniel Dennett claims that evolution is algorithmic. On Dennett’s analysis, evolutionary processes are trivially algorithmic because he assumes that all natural processes are algorithmic. I will argue that there are more robust ways to understand algorithmic processes that make the claim that evolution is algorithmic empirical and not conceptual. While laws of nature can be seen as compression algorithms of information about the world, it does not follow logically that they are implemented as algorithms by physical (...) processes. For that to be true, the processes have to be part of computational systems. The basic difference between mere simulation and real computing is having proper causal structure. I will show what kind of requirements this poses for natural evolutionary processes if they are to be computational. (shrink)
Explanations in cognitive science and computational neuroscience rely predominantly on computational modeling. Although the scientific practice is systematic, and there is little doubt about the empirical value of numerous models, the methodological account of computational explanation is not up-to-date. The current chapter offers a systematic account of computational explanation in cognitive science and computational neuroscience within a mechanistic framework. The account is illustrated with a short case study of modeling of the mirror neuron system in terms of predictive coding.
In this paper, I want to deal with the triviality threat to computationalism. On one hand, the controversial and vague claim that cognition involves computation is still denied. On the other, contemporary physicists and philosophers alike claim that all physical processes are indeed computational or algorithmic. This claim would justify the computationalism claim by making it utterly trivial. I will show that even if these two claims were true, computationalism would not have to be trivial.
The standard objection against naturalised epistemology is that it cannot account for normativity in epistemology (Putnam 1982; Kim 1988). There are different ways to deal with it. One of the obvious ways is to say that the objection misses the point: It is not a bug; it is a feature, as there is nothing interesting in normative principles in epistemology. Normative epistemology deals with norms but they are of no use in prac-tice. They are far too general to be guiding (...) principles of research, up to the point that they even seem vacuous (see Knowles 2003). In this chapter, my strategy will be different and more in spirit of the founding father of naturalized epistemology, Quine, though not faithful to the letter. I focus on methodological prescriptions supplied by cogni-tive science in re-engineering of cognitive architectures. Engineering norms based on mechanism design weren’t treated as seriously as they should in epistemology, and that is why I will develop a sketch of a framework for researching them, starting from analysing cognitive sci-ence as engineering in section 3, then showing functional normativity in section 4, to eventually present functional engineering models of cogni-tive mechanisms as normative in section 5. Yet before showing the kind of engineering normativity specific for these prescriptions, it is worth-while to review briefly the role of normative methodology and the levels of norm complexity in it, and show how it follows Quine’s steps. (shrink)
Recent work on skin-brain thesis suggests the possibility of empirical evidence that empiricism is false. It implies that early animals need no traditional sensory receptors to be engaged in cognitive activity. The neural structure required to coordinate extensive sheets of contractile tissue for motility provides the starting point for a new multicellular organized form of sensing. Moving a body by muscle contraction provides the basis for a multicellular organization that is sensitive to external surface structure at the scale of the (...) animal body. In other words, the nervous system first evolved for action, not for receiving sensory input. Thus, sensory input is not required for minimal cognition; only action is. The whole body of an organism, in particular its highly specific animal sensorimotor organization, reflects the bodily and environmental spatiotemporal structure. The skin-brain thesis suggests that, in contrast to empiricist claims that cognition is constituted by sensory systems, cognition may be also constituted by action-oriented feedback mechanisms. Instead of positing the reflex arc as the elementary building block of nervous systems, it proposes that endogenous motor activity is crucial for cognitive processes. In the paper, I discuss the issue whether the skin-brain thesis and its supporting evidence can be really used to overthrow the main tenet of empiricism empirically, by pointing out to cognizing agents that fail to have any sensory apparatus. (shrink)
Deleuze uważa, ze nie można pogodzić Hegla i Nietzschego. Hegel jest wedle niego abstrakcyjny, a Nietzsche - konkretny. Tymczasem pojęcia "konkret" i "abstrakcja" należą do ideologicznego arsenału konserwatyzmu. Rozpatruję nie tyle prawdziwość tezy Deleuza, co jej genealogię. Hegel i Nietzsche kontynuują oświeceniowe poszukiwania "człowieka konkretnego". "Człowiek konkretny" to wytwór drugiej fazy oświecenia (rodzaj "kompensacji" w znaczeniu Marquarda): przekształcenie parenetyki w filozofię historii i kultury (wzgl. społeczną). "Wielki bohater historii" i "nadczłowiek" są próbami ujęcia konkretu społeczno-historycznego. Rzut oka na strukturalną pozycję (...) kategorii Hegla w jego systemie oraz ideału Nietzschego w jego myśli (mediatyzacja a problem wątpliwej konkretności nadczłowieka) uprawdopodabnia następujący wniosek: Nadczłowiek stanowi abstrakcyjny projekt uchwycenia konkretu. Nie można ukryć trudności w ocenie stosunku Hegla i Nietzschego wobec konserwatyzmu romantycznego. Przyczyna leży m.in. w wykorzystaniu arsenału myśli konserwatywnej przy jednoczesnym jej przekształceniu. Stąd trudno orzec, czy "konserwatyzm romantyczny" zniesiony (Hegel) lub przekształcony w ekstatyczno-chiliastyczną filozofię przyszłości (Nietzsche) jest bardziej konserwatywny, czy raczej liberalny itp. Jeden przykład z "warsztatu filozofowania" obu myślicieli wykazuje swoiste odniesienie się do konkretu, a także problematyczność tego odniesienia. Co więcej, opozycja konkret-abstrakt przy zmieniającym się jej nacechowaniu ideologicznym obejmuje również "empiryzm transcendentalny" samego Deleuze'a, a także - jak się zdaje - sporą część filozofii kontynuującej Nietzschego w intencji sprzeciwienia się Heglowi i jego metanarracjom. Wieloznaczność problemu konkretu ukrywa to, że jest pseudoproblemem lub - w najlepszym wypadku - mylącym wyróżnikiem stanowiska filozoficznego. (shrink)
It would be hard to find a more fervent advocate of the position that computers are of profound significance to philosophy than Aaron Sloman. Yet, he is not a stereotypical proponent of Artificial Intelligence (AI). Far from it; in his writings, he undermines several popular convictions of functionalists. Through his drafts and polemics, Sloman definitely exerts quite substantial influence on the philosophy of Artificial Intelligence. Sloman's paper “Evolution: The Computer Systems Engineer Designing Minds” presents a bold hypothesis that the evolution (...) of the human mind actually involved the development of a several dozen of virtual machines that support various forms of self-monitoring. This, in turn, helps explain different features of our cognitive functioning. (shrink)
I argue that influential purely syntactic views of computation, shared by such philosophers as John Searle and Hilary Putnam, are mistaken. First, I discuss common objections, and during the discussion I mention additional necessary conditions of implementation of computations in physical processes that are neglected in classical philosophical accounts of computation. Then I try to show why realism in regards of physical computations is more plausible, and more coherent with any realistic attitude towards natural science than the received view, and (...) distinguish computational simulation, implementation, and re-engineering. I also point out the sources of confusion about what computation is that seem to stem from disregarding the use/mention distinction. (shrink)
In this chapter, I argue that some aspects of cognitive phenomena cannot be explained computationally. In the first part, I sketch a mechanistic account of computational explanation that spans multiple levels of organization of cognitive systems. In the second part, I turn my attention to what cannot be explained about cognitive systems in this way. I argue that information-processing mechanisms are indispensable in explanations of cognitive phenomena, and this vindicates the computational explanation of cognition. At the same time, it has (...) to be supplemented with other explanations to make the mechanistic explanation complete, and that naturally leads to explanatory pluralism in cognitive science. The price to pay for pluralism, however, is the abandonment of the traditional autonomy thesis asserting that cognition is independent of implementation details. (shrink)
In this paper, I suggest that the notion of module explicitly defined by Peter Carruthers in The Architecture of The Mind (Carruthers 2006) is not really In use in the book. Instead, a more robust notion seems to be actually in play. The more robust notion, albeit implicitly assumed, seems to be far more useful for making claims about the modularity of mind. Otherwise, the claims would become trivial. This robust notion will be reconstructed and improved upon by putting it (...) into a more general framework of mental architecture. I defend the view that modules are the outcome of structural rather than functional decomposition and that they should be conceived as near decomposable systems. (shrink)
The goal of the article is to show that a complete answer to the title question can be given only in the context of the natural sciences. We believe that the group of cognitive sciences are the most reliable source of information about cognitive mental processes is. Making use of their achievements we present a series of criteria for possessing a mind. We distinguish between many kinds of minds . We attempt to outline the conditions that must be fulfilled by (...) an adequate model of the mind. In our opinion such a model must make use of all available empirical data and of scientific theories constructed on the basis of such data. From the point of view of philosophy the requirements placed upon such theories by ontology are especially important. Their reconstruction can be a prolegomena to a future integrated ontology of the mind. We emphasize that the mind is not an independent thing . In speaking about the mind we have in mind states, events, processes, functions, and dispositions that are derivative with respect to processes of a lower order. We assume that an adequate model of the mind is multi-dimensional, taking into account several mutually interacting levels of organization . We interpret the psychophysical problem as one of the relation between levels of organization, a relation that is constitutive for the actualization of mental states. Psychophysical relations turn out to be a particular case of the broader issue of relations between levels. In carrying out a preliminary conceptualization we make use of the notion of emergence; this is why our position, which is mainly in opposition to substantial dualism, may be termed emergent monism or naturalism. (shrink)
Autor artykułu broni tezy, że niektóre systemy obliczeniowe mogą mieć własności semantyczne. Wskazana została klasa systemów obliczeniowych, w których reprezentacje mogą mieć przynajmniej dwie własności: własność odnoszenia się do obiektów (desygnowanie) i własność wspomagania rozpoznawania obiektów oznaczanych przez daną reprezentację (konotowanie). Autor argumentuje także, że własności semantyczne reprezentacji nie zależą wyłącznie od architektury systemów obliczeniowych, w których te reprezentacje występują. Konkretna architektura obliczeniowa nie jest czynnikiem kluczowym, a bodaj najmniej istotne są same rodzaje struktur danych, które mają mieć własności desygnowania (...) czy konotowania. Własność desygnowania czy konotowania nie musi być zlokalizowana w samych reprezentacjach, może być własnością wyższego rzędu, powstającą w mechanizmie wyższego poziomu. Własności semantyczne reprezentacji mogą być wielorako realizowane. Systemy klasyczne, koneksjonistyczne czy też hybrydowe mogą równie dobrze mieć własności semantyczne, jak ich nie mieć. (shrink)
Herbert A. Simon is well known for his account of bounded rationality. Whereas classical economics idealized economic agency and framed rational choice in terms of the decision theory, Simon insisted that agents need not be optimal in their choices. They might be mere satispcers, i.e., attain good enough goals rather than optimal ones. At the same time, behaviorally as well as computationally, bounded rationality is much more realistic.
W artykule przedstawiono argumenty, że konfirmacja tezy, iż istnieją moduły umysłowe wyjaśniające cechy umysłu, jest z kilku powodów kłopotliwa. Po pierwsze, istnieje kilka konkurencyjnych teorii modularności, które zresztą nie zawsze się wykluczają, przez co nie można między nimi rozstrzygać eksperymentalnie. Po drugie, tezy na temat modularności często oparte są na bezzasadnym założeniu, iż wyróżnianie specyficznych dziedzin (semantycznych lub składniowych) działania modułów nie jest problematyczne. Po trzecie, analizując znany z literatury moduł wykrywania oszustów, postulowany przez Cosmides w celu wyjaśnienia rzekomej irracjonalności (...) objawiającej się w tzw. zadaniu Wasona, pokazuję, że wyjaśniane zjawisko nie zostało zdefiniowane dostatecznie dokładnie, a przez to nieostra jest funkcjonalna charakterystyka modułów je wyjaśniających. Co więcej, nie ma powodów sądzić, że zjawisko, które ten moduł miał wyjaśniać, w ogóle istnieje. Wskazuję też kilka problemów metodologicznych związanych ze zbieraniem eksperymentalnych świadectw na rzecz modularności, takich jak zaburzanie wyników eksperymentów przez uśrednianie i brak kontroli nad kluczowymi czynnikami wpływającymi na rezultaty uzyskiwane przez uczestników badania. (shrink)
Naturalism is currently the most vibrantly developing approach to philosophy, with naturalised methodologies being applied across all the philosophical disciplines. One of the areas naturalism has been focussing upon is the mind, traditionally viewed as a topic hard to reconcile with the naturalistic worldview. A number of questions have been pursued in this context. What is the place of the mind in the world? How should we study the mind as a natural phenomenon? What is the significance of cognitive science (...) research for philosophical debates? In this book, philosophical questions about the mind are asked in the context of recent developments in cognitive science, evolutionary theory, psychology, and the project of the naturalisation. Much of the focus is upon what we have learned by studying natural mental mechanisms as well as designing artificial ones. In the case of natural mental mechanisms, this includes consideration of such issues as the significance of deficits in these mechanisms for psychiatry. The significance of the evolutionary context for mental mechanisms as well as questions regarding rationality and wisdom is also explored. Mechanistic and functional models of the mind are used to throw new light on discussions regarding issues of explanation, reduction and the realisation of mental phenomena. Finally, naturalistic approaches are used to look anew at such traditional philosophical issues as the correspondence of mind to world and presuppositions of scientific research. (shrink)
The contributors to this volume engage with issues of normativity within naturalised philosophy. The issues are critical to naturalism as most traditional notions in philosophy, such as knowledge, justification or representation, are said to involve normativity. Some of the contributors pursue the question of the correct place of normativity within a naturalised ontology, with emergentist and eliminativist answers offered on neighbouring pages. Others seek to justify particular norms within a naturalised framework, the more surprising ones including naturalist takes on the (...) a priori and intuitions. Finally, yet others examine concrete examples of the application of norms within particular epistemic endeavours, such as psychopathology and design. The overall picture is that of an intimate engagement with issues of normativity on the part of naturalist philosophers – questioning some of the fundamentals at the same time as they try to work out many of the details. (shrink)