In this paper we seek to make the case for a teaching and learning strategy that integrates business ethics in the curriculum, whilst not precluding a disciplines based approach to this subject. We do this in the context of specific work experience modules at undergraduate level which are offered by Middlesex University Business School, part of a modern university based in North West London. We firstly outline our educative values and then the modules that form the basis of our research. (...) We then identify and elaborate what we believe are the five dimensions which distinguish an integrated approach based on work experience from a disciplines-based approach, namely: process and content, internal and external, facilitation and teaching, covert and overt, and living wisdom and established wisdom. The last dimension draws on the practical relevance of the Aristotelian notion of phronesis inherent in our approach. We go on to provide two case examples of our practice to illustrate our perspective and in support of our conclusions. These are that reflection integrated into the Business Studies curriculum, using the ASKE typology of learning [Frame, 2001, Proceedings of the 9th Annual Teaching and Learning Conference (Nottingham: Nottingham Business School, Nottingham Trent University), p. 80], in respect of personal and group process in a work experience context, provides a useful heuristic for the development of moral sensibility and ethical practice. (shrink)
The frame problem is the problem of how we selectively apply relevant knowledge to particular situations in order to generate practical solutions. Some philosophers have thought that the frame problem can be used to rule out, or argue in favor of, a particular theory of belief states. But this is a mistake. Sentential theories of belief are no better or worse off with respect to the frame problem than are alternative theories of belief, most notably, the “map” (...) theory of belief. (shrink)
The frame problem is the difficulty of explaining how non-magical systems think and act in ways that are adaptively sensitive to context-dependent relevance. Influenced centrally by Heideggerian phenomenology, Hubert Dreyfus has argued that the frame problem is, in part, a consequence of the assumption (made by mainstream cognitive science and artificial intelligence) that intelligent behaviour is representation-guided behaviour. Dreyfus' Heideggerian analysis suggests that the frame problem dissolves if we reject representationalism about intelligence and recognize that human agents (...) realize the property of thrownness (the property of being always already embedded in a context). I argue that this positive proposal is incomplete until we understand exactly how the properties in question may be instantiated in machines like us. So, working within a broadly Heideggerian conceptual framework, I pursue the character of a representation-shunning thrown machine. As part of this analysis, I suggest that the frame problem is, in truth, a two-headed beast. The intra-context frame problem challenges us to say how a purely mechanistic system may achieve appropriate, flexible and fluid action within a context. The inter-context frame problem challenges us to say how a purely mechanistic system may achieve appropriate, flexible and fluid action in worlds in which adaptation to new contexts is open-ended and in which the number of potential contexts is indeterminate. Drawing on the field of situated robotics, I suggest that the intra-context frame problem may be neutralized by systems of special-purpose adaptive couplings, while the inter-context frame problem may be neutralized by systems that exhibit the phenomenon of continuous reciprocal causation. I also defend the view that while continuous reciprocal causation is in conflict with representational explanation, special-purpose adaptive coupling, as well as its associated agential phenomenology, may feature representations. My proposal has been criticized recently by Dreyfus, who accuses me of propagating a cognitivist misreading of Heidegger, one that, because it maintains a role for representation, leads me seriously astray in my handling of the frame problem. I close by responding to Dreyfus' concerns. (shrink)
Within cognitive science, mental processing is often construed as computation over mental representations—i.e., as the manipulation and transformation of mental representations in accordance with rules of the kind expressible in the form of a computer program. This foundational approach has encountered a long-standing, persistently recalcitrant, problem often called the frame problem; it is sometimes called the relevance problem. In this paper we describe the frame problem and certain of its apparent morals concerning human cognition, and we argue that (...) these morals have significant import regarding both the nature of moral normativity and the human capacity for mastering moral normativity. The morals of the frame problem bode well, we argue, for the claim that moral normativity is not fully systematizable by exceptionless general principles, and for the correlative claim that such systematizability is not required in order for humans to master moral normativity. (shrink)
It is shown that the Fodor's interpretation of the frame problem is the central indication that his version of the Modularity Thesis is incompatible with computationalism. Since computationalism is far more plausible than this thesis, the latter should be rejected.
This paper investigates connectionism's potential to solve the frame problem. The frame problem arises in the context of modelling the human ability to see the relevant consequences of events in a situation. It has been claimed to be unsolvable for classical cognitive science, but easily manageable for connectionism. We will focus on a representational approach to the frame problem which advocates the use of intrinsic representations. We argue that although connectionism's distributed representations may look promising from this (...) perspective, doubts can be raised about the potential of distributed representations to allow large amounts of complexly structured information to be adequately encoded and processed. It is questionable whether connectionist models that are claimed to effectively represent structured information can be scaled up to a realistic extent. We conclude that the frame problem provides a difficulty to connectionism that is no less serious than the obstacle it constitutes for classical cognitive science. (shrink)
The theory-theory claims that the explanation and prediction of behavior works via the application of a theory, while the simulation theory claims that explanation works by putting ourselves in others' places and noting what we would do. On either account, in order to develop a prediction or explanation of another person's behavior, one first needs to have a characterization of that person's current or recent actions. Simulation requires that I have some grasp of the other person's behavior to project myself (...) upon; whereas theorizing requires a subject matter to theorize about. The frame problem shows that multiple, true characterizations are possible for any behavior or situation. However, only one or a few of these characterizations are relevant to explaining or predicting behavior. Since different characterizations of a behavior lead to different predictions or explanations, much of the work of interpersonal interpretation is done in the process of finding this characterization - that is, prior to either theorizing or simulating. Moreover, finding this characterization involves extensive knowledge of the physical, cultural, and social worlds of the persons involved. (shrink)
The frame problem is widely reputed among philosophers to be one of the deepest and most difficult problems of cognitive science. This paper discusses three recent attempts to display this problem: Dennett's problem of ignoring obviously irrelevant knowledge, Haugeland's problem of efficiently keeping track of salient side effects, and Fodor's problem of avoiding the use of kooky concepts. In a negative vein, it is argued that these problems bear nothing but a superficial similarity to the frame problem of (...) AI, so that they do not provide reasons to disparage standard attempts to solve it. More positively, it is argued that these problems are easily solved by slight variations on familiar AI themes. Finally, some discussion is devoted to more difficult problems confronting AI. (shrink)
As many philosophers agree, the frame problem is concerned with how an agent may efficiently filter out irrelevant information in the process of problem-solving. Hence, how to solve this problem hinges on how to properly handle semantic relevance in cognitive modeling, which is an area of cognitive science that deals with simulating human's cognitive processes in a computerized model. By "semantic relevance", we mean certain inferential relations among acquired beliefs which may facilitate information retrieval and practical reasoning under certain (...) epistemic constraints, e. g., the insufficiency of knowledge, the limitation of time budget, etc. However, traditional approaches to relevance—as for example, relevance logic, the Bayesian approach, as well as Description Logic—have failed to do justice to the foregoing constraints, and in this sense, they are not proper tools for solving the frame problem/relevance problem. As we will argue in this paper, Non-Axiomatic Reasoning System (NARS) can handle the frame problem in a more proper manner, because the resulting solution seriously takes epistemic constraints on cognition as a fundamental theoretical principle. (shrink)
Taking Löb's Axiom in modal provability logic as a running thread, we discuss some general methods for extending modal frame correspondences, mainly by adding fixed-point operators to modal languages as well as their correspondence languages. Our suggestions are backed up by some new results – while we also refer to relevant work by earlier authors. But our main aim is advertizing the perspective, showing how modal languages with fixed-point operators are a natural medium to work with.
This paper aims at analyzing Philip Kitcher's naturalistic epistemology, particularly its normative features, which are viewed as a sort of response to negative assessments made by radical naturalists on the plurality of epistemic values. According to them such values are ineffective for normative ends, e.g. theory choice. Differently from that quite excessive evaluation, Kitcher argues rather for explanatory unity as the most important and universal epistemic value. Even though Kitcher's arguments are sound, there remains some serious gaps as regards (...) his attempts; there are also serious doubts about the desirability of achieving such a value. (shrink)
With calls for (business) leaders to contribute to greater global fairness and social justice (BAWB 2006; Maak and Pless Journal of Business Ethics, 88, 537–550, 2009), this paper considers gender equality on University home web page images as one means of communicating equal access to leadership roles for both men and women. Although there are many paths for leadership development, one important purpose of Universities is to create people who will potentially become leaders in our society (Shapiro 2005). We analyzed (...) the home web pages at 24 leading universities to identify implicit messages about gender roles, building on implicit leadership theory and leadership prototypes. Using an adapted version of Goffman’s frame analysis (1979), our results suggest depiction of gender equality in university home web pages, in contrast to studies of print advertisements (cf., Kang Sex Roles, 37(11/12), 979–996, 1997; Lindner Sex Roles, 51(7/8), 409–421, 2004). Our results also identify specific opportunities to depict greater equity and to continue to expand the potential for both women and men to be seen as being capable and belonging on this leadership path. (shrink)
Architects design building structures, although structural design is the profession of structural engineers. Thus, it is better for architects and structural engineers to collaborate starting from the initial phases of the architectural design. However, this is not very common because of the contradictory design processes and value systems held within the two professions. This article provides a platform upon which architects and structural engineers can resolve the value conflicts between them by analysing phases of the structural design of reinforced concrete (...)frame systems in architecture, the criteria of the structural design for each phase and determining the conflicting values for each criterion. The results shown in the article demonstrate that the architectural design of structures is a complex process, which is based on contradictory values and value systems. Finally, the article suggests to architects and structural engineers to use Value Sensitive Design and to choose an appropriate team leader in order to resolve the unethical conflict between them and to avoid any unreasonable decision making. (shrink)
For many of the authors in this volume, this is the second attempt to explore what McCarthy and Hayes (1969) ﬁrst called the “Frame Problem”. Since the ﬁrst compendium (Pylyshyn, 1987), nicely summarized here by Ronald Loui, there have been several conferences and books on the topic. Their goals range from providing a clariﬁcation of the problem by breaking it down into subproblems (and sometimes declaring the hard subproblems to not be the_ real_ Frame Problem), to providing formal (...) “solutions” to certain aspects of the problem. But more often the message has been that the problem is not solvable except in a piecemeal way in special circumstances by some sort of heuristic approximations. It has sometimes also been said that solving the Frame Problem is not only an unachievable goal, but it is also an unnecessary one since_ humans_ do not solve it either; we simply get along as best we can and deal with the problem of planning in ways that, to use Dennett’s phrase, is “good enough for government work”. (shrink)
The simplest combination of unimodal logics L1 and L2 into a bimodal logic is their fusion, L1 ⊗ L2, axiomatized by the theorems of L1 for 1 and of L2 for 2. Shehtman introduced combinations that are not only bimodal, but two-dimensional: he defined 2-d Cartesian products of 1-d Kripke frames, using these Cartesian products to define the frame product L1 × L2 of L1 and L2. Van Benthem, Bezhanishvili, ten Cate and Sarenac generalized Shehtman’s idea and introduced the (...) topological product L1 ×t L2, using Cartesian products of topological spaces rather than of Kripke frames. Frame products have been extensively studied, but much less is known about topological products. The goal of the current paper is to give necessary and conditions for the topological product to match the frame product, for Kripke complete extensions of S4: L1×tL2 = L1× L2 iff L1 S5 or L2 S5 or L1, L2 = S5. (shrink)
This brief opening for a special issue of Tradition and Discovery: The Polanyi Society Periodical on Philip Clayton’s thought and its connection with that of Michael Polany introduces Clayton’s essay and the responses by Martinez Hewlett, Gregory R. Peterson, Andy F. Sanders and Waler B. Gulick.
Semantic Leaps explores how people combine knowledge from different domains in order to understand and express new ideas. Concentrating on dynamic aspects of on-line meaning construction, Coulson identifies two related sets of processes: frame-shifting and conceptual blending. Frame-shifting is semantic reanalysis in which existing elements in the contextual representation are reorganized into a new frame. Conceptual blending is a set of cognitive operations for combining partial cognitive models. By addressing linguistic phenomena often ignored in traditional meaning research, (...) Coulson explains how processes of cross-domain mapping, frame-shifting, and conceptual blending enhance the explanatory adequacy of traditional frame-based systems for natural language processing. The focus is on how the constructive processes speakers use to assemble, link, and adapt simple cognitive models underlie a broad range of productive language behavior. (shrink)
Prior research flags the inherent incompatibilities between for-profit and nonprofit partners and cautions that clashing value creation logics and conflicting identities can stall social innovation in cross sector partnerships. Process narratives of successful versus unsuccessful cross sector partnerships paint a more optimistic picture, whereby the frequency, intensity, breadth, and depth of interactions may afford frame alignment despite partners’ divergent value creation approaches. However, little is known about how cross sector partners come to recognize and reconcile their divergent value creation (...) frames in order to co-construct social value. Using longitudinal narratives of four dyads, we show that partners initially contrast their sector-embedded diagnostic frames and then work together to deliberately develop partnership-specific prognostic frames. We extend the literature on framing by developing a four-stage grounded model of frame negotiation, elasticity, plasticity, and fusion which unpacks the relational process of value creation in cross sector partnerships. Our qualitative analyses further show how partners orchestrate multilevel coordination that helps scaffold and calibrate this relational process of frame fusion. (shrink)
In Science, Truth, and Democracy, Philip Kitcher develops the notion of well-ordered science: scientific inquiry whose research agenda and applications (but not methods) are subject to public control guided by democratic deliberation. Kitcher's primary departure from his earlier views involves rejecting the idea that there is any single standard of scientific significance. The context-dependence of scientific significance opens up many normative issues to philosophical investigation and to resolution through democratic processes. Although some readers will feel Kitcher has not (...) moved far enough from earlier epistemological positions, the book does represent an important addition to literature on science, society, and values. (shrink)
Philosophy is often conceived in the Anglophone world today as a subject that focuses on questions in particular ‘‘core areas,’’ pre-eminently epistemology and metaphysics. This article argues that the contemporary conception is a new version of the scholastic ‘‘self-indulgence for the few’’ of which Dewey complained nearly a century ago. Philosophical questions evolve, and a first task for philosophers is to address issues that arise for their own times. The article suggests that a renewal of philosophy today should turn the (...) contemporary conception inside out, attending to and developing further the valuable work being done on the supposed ‘‘periphery’’ and attending to the ‘‘core areas’’ only insofar as is necessary to address genuinely significant questions. (shrink)
Chiappe and Kukla argue that relevance theory fails to solve the frame problem as defined by Fodor. They are right. They are wrong, however, to take Fodors frame problem too seriously. Fodors concerns, on the other hand, even though they are wrongly framed, are worth addressing. We argue that Relevance thoery helps address them.
The frame problem is a problem in artificial intelligence that a number of philosophers have claimed has philosophical relevance. The structure of this paper is as follows: (1) An account of the frame problem is given; (2) The frame problem is distinguished from related problems; (3) The main strategies for dealing with the frame problem are outlined; (4) A difference between commonsense reasoning and prediction using a scientific theory is argued for; (5) Some implications for the..
From its humble origins labeling a technical annoyance for a particular AI formalism, the term "frame problem" has grown to cover issues confronting broader research programs in AI. In philosophy, the term has come to encompass allegedly fundamental, but merely superficially related, objections to computational models of mind in AI and beyond.
I analyze the frame problem and its relation to other epistemological problems for artificial intelligence, such as the problem of induction, the qualification problem and the "general" AI problem. I dispute the claim that extensions to logic (default logic and circumscriptive logic) will ever offer a viable way out of the problem. In the discussion it will become clear that the original frame problem is really a fairy tale: as originally presented, and as tools for its solution are (...) circumscribed by Pat Hayes, the problem is entertaining, but incapable of resolution. The solution to the frame problem becomes available, and even apparent, when we remove artificial restrictions on its treatment and understand the interrelation between the frame problem and the many other problems for artificial epistemology. I present the solution to the frame problem: an adequate theory and method for the machine induction of causal structure. Whereas this solution is clearly satisfactory in principle, and in practice real progress has been made in recent years in its application, its ultimate implementation is in prospect only for future generations of AI researchers. (shrink)
In their article entitled “Ethical Particularism and Patterns”, Frank Jackson, Philip Pettit, and Michael Smith (JPS henceforth) argue that moral particularism is a cognitively implausible theory since it appears to entail the view that one might have a skill that is not grounded in an ability to recognise and represent natural patterns in the world. This charge echoes the complaints of computational theorists of cognition against their embodied cognition counterparts, namely that, theories of cognition that eschew talk of mental (...) representation are implausible qua theories of cognition. In both debates, the cognitive role of generalisation is central to the discussion; however, contrary to the received wisdom, I want to suggest that the dispute is not between generalisation or mental representation on the one hand and no generalisation or mental representation on the other, but rather between what I will call global and local generalisation. Using the dialogue between JPS and Dancy (our paradigm particularist) to frame this discussion, I show that by replacing Dancy's connectionist model for particularist reasoning with a case-based one, we not only vindicate his response to JPS, but we also gain insight into how it is the global/local distinction rather than the generalisation/no generalisation distinction that divides the two views. (shrink)
In this paper we examine the pattern of conceptual change during scientific revolutions by using methods from cognitive psychology. We show that the changes characteristic of scientific revolutions, especially taxonomic changes, can occur in a continuous manner. Using the frame model of concept representation to capture structural relations within concepts and the direct links between concept and taxonomy, we develop an account of conceptual change in science that more adequately reflects the current understanding that episodes like the Copernican revolution (...) are not always abrupt. When concepts are represented by frames, the transformation from one taxonomy to another can be achieved in a piecemeal fashion not preconditioned by a crisis stage, and a new taxonomy can arise naturally out of the old frame instead of emerging separately from the existing conceptual system. This cognitive mechanism of continuous change demonstrates the constructive roles of anomaly and incommensurability in promoting the progress of science. (shrink)
The aim of this paper is to give a systematic account of the so-called “measurement problem” in the frame of the standard interpretation of quantum mechanics. It is argued that there is not one but five distinct formulations of this problem. Each of them depends on what is assumed to be a “satisfactory” description of the measurement process in the frame of the standard interpretation. Moreover, the paper points out that each of these formulations refers not to a (...) unique problem, but to a set of sub-problems. (shrink)
Genetic determinism is the idea that many significant human characteristics are rendered inevitable by the presence of certain genes. The psychologist Susan Oyama has famously compared arguing against genetic determinism to battling the undead. Oyama suggests that genetic determinism is inherent in the way we currently represent genes and what genes do. As long as genes are represented as containing information about how the organism will develop, they will continue to be regarded as determining causes no matter how much evidence (...) exists to the contrary. Philip Kitcher has strongly disputed Oyama’s diagnosis, arguing that the conventional ‘interactionist’ perspective on development is the correct framework for understanding the role of the genes in development. While acknowledging the legitimacy of many of Kitcher’s observations, I believe that Oyama’s view is substantially correct. In this paper I provide several lines of support for support the Oyama diagnosis. (shrink)
En este artículo me propongo analizar el punto de partida epistemológico de un reciente libro de Philip Kitcher (The Advancement of Science) a través de su discusión con las concepciónes ‘escépticas’. Podemos distinguir entre dos tipos de escepticismo en Ia trama deI libro de Kitcher: uno débil y otro radical. Intentamos difinir el tipo de realismo que Kitcher defiende, para finalmente mostrar que tal tipo de realismo es posible para Kitcher en Ia medida que no toma en cuenta el (...) escepticismo en su versión radical. En efecto, Kitcher sólo se enfrenta al escepticismo débil. Y es precisamente debido a esta restricción que es capaz de mantenerse al margen de una alternativa que sigue siendo crucial: realismo fuerte o realismo “de espíritu kantiano”.The purpose of this article is to carry out an analysis of the epistemologic standpoint on a recent book by Philip Kitcher (The Advancement of Science) by discussing the sceptic ideas which are dealt with there. We can discriminate between two kinds of scepticism appearing on Kitcher’s book: a weak and a radical one. Then we work towards a definition of the kind of realism held by this author and, finally, we try to show that such a viewpoint as Kitcher’s is possible to hold provided that we do not take the radical scepticism into account for that question. Kitcher only objects by means of the weak scepticism. And it is precisely because of that restriction that he is capable of not giving a definition of a crucial alternative: strong realism or realism in “Kantian spirit”. (shrink)
Dispositions can combine as vector sums. Recent authors on dispositions, such as George Molnar and Stephen Mumford, have responded to this feature of dispositions by introducing a distinction between effects and contributions to effects, and by identifying disposition-manifestations with the latter. But some have been sceptical of the reality or knowability of component vectors; Jennifer McKitrick (forthcoming) presses these concerns against the conception of manifestations as contributions to effects. In this paper, I aim to respond to McKitrick's arguments and to (...) defend the metaphysical and epistemological propriety of component vectors. My strategy appeals to varying kinematic frames of reference. By transforming to the appropriate non-inertial frame, component acceleration vectors can be transformed into resultant acceleration vectors, and in such frames they become directly observable. Being a component acceleration vector and being a resultant acceleration vector are both frame-dependent properties of properties; they are not to be thought of as intrinsic or fundamental properties of an acceleration vector, but as artefacts of our frame-dependent notation for representing vector quantities. To conclude the paper, I defend the view proposed against two styles of objection. The first objection resurrects scepticism about component vectors as scepticism about fundamental component vectors. The second objection questions the need for reference frames in the explanation by invoking a 'counterfactual' theory of contributions. (shrink)
In Nietzsche and the Horror of Existence, Philip J. Kain makes a compelling case for taking Nietzsche’s concern with the subject of horror seriously and then challenges his conclusions about it. A corollary of existence, horror is an ineliminable part of being human. Our experience of horror prompts reflection on life and the act of philosophizing. Arguing it is a formative yet often overlooked theme in Nietzsche’s oeuvre, Kain recognizes that the experience of horror is central to “Nietzsche’s vision” (...) of life, truth, beauty, and knowledge (1). Kain examines Nietzsche’s interrogation of philosophical responses to horror, tracing his approach from his innovative reinterpretation of the function of tragic .. (shrink)
Kleinberg (1999) describes a novel procedure for efficient search in a dense hyper-linked environment, such as the world wide web. The procedure exploits information implicit in the links between pages so as to identify patterns of connectivity indicative of “authorative sources”. At a more general level, the trick is to use this second-order link-structure information to rapidly and cheaply identify the knowledge-structures most likely to be relevant given a specific input. I shall argue that Kleinberg’s procedure is suggestive of a (...) new, viable, and neuroscientifically plausible solution to at least (one incarnation of) the so-called “Frame Problem” in cognitive science viz the problem of explaining global abductive inference. More accurately, I shall argue that Kleinberg’s procedure suggests a new variety of “fast and frugal heuristic” (Gigerenzer and Todd (1999)) capable of pressing maximum utility from the vast bodies of information and associations commanded by the biological brain. The paper thus takes up the challenge laid down by Fodor ((1983)(Ms)). Fodor depicts the problem of global knowledge-based reason as the point source of many paradigmatic failings of contemporary computational theories of mind. These failings, Fodor goes on to argue, cannot be remedied by any simple appeal to alternative (e.g. connectionist) modes of encoding and processing. I shall show, however, that connectionist models can provide for one neurologically plausible incarnation of Kleinberg’s procedure. The paper ends by noting that current commercial applications increasingly confront the kinds of challenge (such as managing complexity and making efficient use of vast data-bases) initially posed to biological thought and reason. (shrink)
In the first part of this paper we investigate how scientific theories can be represented by frames. Different kinds of scientific theories can be distinguished in terms of the systematic power of their frames. In the second part we outline the central questions and goals of our research project. In the third and final part of this paper we show that frame-representation is a useful tool in the comparison of the theories of phlogiston and oxygen, despite those theories being (...) traditionally conceived as incommensurable. The frame-theoretic representation reveals common attributes, values and ultimately structural correspondence relations between the two theories. In our view this outcome lends credence to a structural realist view of science. (shrink)
This article proposes that Goffman's "Frame Analysis" can be interpreted as a step toward unpacking the idea of context. His analysis implies a recursive model involving frames within frames. The key problem is that neither Goffman nor anyone else has clearly defined what is meant by a frame. I propose that it can be represented by a word, phrase, or proposition. A subjective context can be represented as an assembly of these items, joined together by operators such as (...) and, since, if, not, and then. Furthermore, this model can be combined with the recursive levels of mutual awareness in earlier approaches to consensus. The combination would represent the inter subjective context: it can be used to find the minimum amount of background that would allow consensual interpretations of discourse. It could also construct a chain that links discourse to the institutional level, the micro-macro pathway from word and gesture to social structure. Goffman hinted that mathematical notation might be used to represent a frame assembly. By adding levels of awareness to such notation, it could represent social facts. Because the use of vernacular words rather than concepts is a problem in social science, Goffman's approach has a general as well as a particular significance. (shrink)
The species-specific organizational property of speech is a continual mouth open-close alternation, the two phases of which are subject to continual articulatory modulation. The cycle constitutes the syllable, and the open and closed phases are segments framescontent displays that are prominent in many nonhuman primates. The new role of Broca's area and its surround in human vocal communication may have derived from its evolutionary history as the main cortical center for the control of ingestive processes. The frame and content (...) components of speech may have subsequently evolved separate realizations within two general purpose primate motor control systems: (1) a motivation-related medial system, including anterior cingulate cortex and the supplementary motor area, for self-generated behavior, formerly responsible for ancestral vocalization control and now also responsible for frames, and (2) a lateral system, including Broca's area and surround, and Wernicke's area, specialized for response to external input (and therefore the emergent vocal learning capacity) and more responsible for content. (shrink)
This paper is concerned with the possibility and nature of relativistic hidden-variable formulations of quantum mechanics. Both ad hoc teleological constructions of spacetime maps and frame-dependent constructions of spacetime maps are considered. While frame-dependent constructions are clearly preferable, they provide neither mechanical nor causal explanations for local quantum events. Rather, the hiddenvariable dynamics used in such constructions is just a rule that helps to characterize the set of all possible spacetime maps. But while having neither mechanical nor causal (...) explanations of the values of quantummechanical measurement records is a signiﬁcant cost, it may simply prove too much to ask for such explanations in relativistic quantum mechanics. (shrink)
Classifying spatial frames of references have placed egocentric/body-based representations on muddy grounds. The traditional taxonomy places it under the deictic distinction while the Levinson’s terminology does not provide a special status for it but classifies it along with the relative frame of reference. Research from other areas of cognition has come up with other implied classifications that are motivated by the special role played by these egocentric representation(s). Tangled among such issues is the fuzzy distinction between egocentric and body (...) based representations. The current paper takes up exactly this issue and proposes to sub classify egocentric representations into two different subtypes namely the first- and the second-order representations. The proposed distinction serves an essential purpose for understanding important cognitive processes like spatial transformation, mental perspective taking, and so on. (shrink)