In an interesting experimental study, Bonini et al. (1999) present partial support for truth-gap theories of vagueness. We say this despite their claim to find theoretical and empirical reasons to dismiss gap theories and despite the fact that they favor an alternative, epistemic account, which they call ‘vagueness as ignorance’. We present yet more experimental evidence that supports gap theories, and argue for a semantic/pragmatic alternative that unifies the gappy supervaluationary approach together with its glutty relative, the subvaluationary approach.
previous theories and the relevance of those criticisms to the new accounts. Additionally, we have included a new section at the end, which gives some directions to literature outside of formal semantics in which the notion of mass has been employed. We looked at work on mass expressions in psycholinguistics and computational linguistics here, and we discussed some research in the history of philosophy and in metaphysics that makes use of the notion of mass.
1: Linguistic and Epistemological Background 1 . 1 : Generic Reference vs. Generic Predication 1 . 2 : Why are there any Generic Sentences at all? 1 . 3 : Generics and Exceptions, Two Bad Attitudes 1 . 4 : Exceptions and Generics, Some Other Attitudes 1 . 5 : Generics and Intensionality 1 . 6 : Goals of an Analysis of Generic Sentences 1 . 7 : A Little Notation 1 . 8 : Generics vs. Explicit Statements of Regularities..
The Principle of Semantic Compositionality (sometimes called Frege''s Principle) is the principle that the meaning of a (syntactically complex) whole is a function only of the meanings of its (syntactic) parts together with the manner in which these parts were combined. This principle has been extremely influential throughout the history of formal semantics; it has had a tremendous impact upon modern linguistics ever since Montague Grammars became known; and it has more recently shown up as a guiding principle for a (...) certain direction in cognitive science.Despite the fact that The Principle is vague or underspecified at a number of points — such as what meaning is, what counts as a part, what counts as a syntactic complex, what counts as combination — this has not stopped some people from viewing The Principle as obviously true, true almost by definition. And it has not stopped other people from viewing The Principle as false, almost pernicious in its effect. And some of these latter theorists think that it is an empirically false principle while others think of it as a methodologically wrong-headed way to proceed. (shrink)
In (1991), Meinwald initiated a major change of direction in the study of Plato’s Parmenides and the Third Man Argument. On her conception of the Parmenides , Plato’s language systematically distinguishes two types or kinds of predication, namely, predications of the kind ‘x is F pros ta alla’ and ‘x is F pros heauto’. Intuitively speaking, the former is the common, everyday variety of predication, which holds when x is any object (perceptible object or Form) and F is a property (...) which x exempliﬁes or instantiates in the traditional sense. The latter is a special mode of predication which holds when x is a Form and F is a property which is, in some sense, part of the nature of that Form. Meinwald (1991, p. 75, footnote 18) traces the discovery of this distinction in Plato’s work to Frede (1967), who marks the distinction between pros allo and kath’ hauto predications by placing subscripts on the copula ‘is’. (shrink)
Natural deduction is the type of logic most familiar to current philosophers, and indeed is all that many modern philosophers know about logic. Yet natural deduction is a fairly recent innovation in logic, dating from Gentzen and Ja?kowski in 1934. This article traces the development of natural deduction from the view that these founders embraced to the widespread acceptance of the method in the 1960s. I focus especially on the different choices made by writers of elementary textbooks?the standard conduits of (...) the method to a generation of philosophers?with an eye to determining what the ?essential characteristics? of natural deduction are. (shrink)
This paper discusses the general problem of translation functions between logics, given in axiomatic form, and in particular, the problem of determining when two such logics are "synonymous" or "translationally equivalent." We discuss a proposed formal definition of translational equivalence, show why it is reasonable, and also discuss its relation to earlier definitions in the literature. We also give a simple criterion for showing that two modal logics are not translationally equivalent, and apply this to well-known examples. Some philosophical morals (...) are drawn concerning the possibility of having two logical systems that are "empirically distinct" but are both translationally equivalent to a common logic. (shrink)
Gentzen’s and Jaśkowski’s formulations of natural deduction are logically equivalent in the normal sense of those words. However, Gentzen’s formulation more straightforwardly lends itself both to a normalization theorem and to a theory of “meaning” for connectives . The present paper investigates cases where Jaskowski’s formulation seems better suited. These cases range from the phenomenology and epistemology of proof construction to the ways to incorporate novel logical connectives into the language. We close with a demonstration of this latter aspect by (...) considering a Sheffer function for intuitionistic logic. (shrink)
In this essay I will consider two theses that are associated with Frege,and will investigate the extent to which Frege really believed them.Much of what I have to say will come as no surprise to scholars of thehistorical Frege. But Frege is not only a historical figure; he alsooccupies a site on the philosophical landscape that has allowed hisdoctrines to seep into the subconscious water table. And scholars in a widevariety of different scholarly establishments then sip from thesedoctrines. I believe (...) that some Frege-interested philosophers at various ofthese establishments might find my conclusions surprising.Some of these philosophical establishments have arisen from an educationalmilieu in which Frege is associated with some specific doctrine at theexpense of not even being aware of other milieux where other specificdoctrines are given sole prominence. The two theses which I will discussillustrate this point. Each of them is called Frege''s Principle, but byphilosophers from different milieux. By calling them milieux I do not want to convey the idea that they are each located at some specificsocio-politico-geographico-temporal location. Rather, it is a matter oftheir each being located at different places on the intellectuallandscape. For this reason one might (and I sometimes will) call them(interpretative) traditions. (shrink)
Default reasoning occurs whenever the truth of the evidence available to the reasoner does not guarantee the truth of the conclusion being drawn. Despite this, one is entitled to draw the conclusion “by default” on the grounds that we have no information which would make us doubt that the inference should be drawn. It is the type of conclusion we draw in the ordinary world and ordinary situations in which we find ourselves. Formally speaking, ‘nonmonotonic reasoning’ refers to argumentation in (...) which one uses certain information to reach a conclusion, but where it is possible that adding some further information to those very same premises could make one want to retract the original conclusion. It is easily seen that the informal notion of default reasoning manifests a type of nonmonotonic reasoning. Generally speaking, default statements are said to be true about the class of objects they describe, despite the acknowledged existence of “exceptional instances” of the class. In the absence of explicit information that an object is one of the exceptions we are enjoined to apply the default statement to the object. But further information may later tell us that the object is in fact one of the exceptions. So this is one of the points where nonmonotonicity resides in default reasoning. (shrink)
Psychologism in logic is the doctrine that the semantic content of logical terms is in some way a feature of human psychology. We consider the historically influential version of the doctrine, Psychological Individualism, and the many counter-arguments to it. We then propose and assess various modifications to the doctrine that might allow it to avoid the classical objections. We call these Psychological Descriptivism, Teleological Cognitive Architecture, and Ideal Cognizers. These characterizations give some order to the wide range of modern views (...) that are seen as psychologistic because of one or another feature. Although these can avoid some of the classic objections to psychologism, some still hold. (shrink)
In 1934 a most singular event occurred. Two papers were published on a topic that had (apparently) never before been written about, the authors had never been in contact with one another, and they had (apparently) no common intellectual background that would otherwise account for their mutual interest in this topic.1 These two papers formed the basis for a movement in logic which is by now the most common way of teaching elementary logic by far, and indeed is perhaps all (...) that is known in any detail about logic by a number of philosophers (especially in North America). This manner of proceeding in logic is called ‘natural deduction’. And in its own way the instigation of this style of logical proof is as important to the history of logic as the discovery of resolution by Robinson in 1965, or the discovery of the logistical method by Frege in 1879, or even the discovery of the syllogistic by Aristotle in the fourth century BC. (shrink)
Strawson described ‘descriptive metaphysics’, Bach described ‘natural language metaphysics’, Sapir and Whorf describe, well, Sapir-Whorﬁanism. And there are other views concerning the relation between correct semantic analysis of linguistic phenomena and the “reality” that is supposed to be thereby described. I think some considerations from the analyses of the mass-count distinction can shed some light on that very dark topic.
This volume showcases an interplay between leading philosophical and linguistic semanticists on the one side, and leading cognitive and developmental psychologists on the other side. The topic is a class of outstanding questions in the semanticists on the one side, and leading cognitive and developmental psychologists on the other side. The topic is a class of outstanding questions in the semantic and logical theories of generic statements and statements that employ mass terms by looking to the cognitive abilities of speakers (...) and of child language-learners. (shrink)
This paper investigates the strange case of an argument that was directed against a positivist verification principle. We find an early occurrence of the argument in a talk by the phenomenologist Roman Ingarden at the 1934 International Congress of Philosophy in Prague, where Carnap and Neurath were present and contributed short rejoinders. We discuss the underlying presuppositons of the argument, and we evaluate whether the attempts by Carnap actually succeed in answering this argument. We think they don’t, and offer instead (...) a few sociological thoughts about why the argument seems to have disappeared from the profession’s evaluaton of the positivist criterion of verifiability. (shrink)
Average‐NPs, such as the one in the title of this paper, have been claimed to be ‘linguistically identical’ to any other definite‐NPs but at the same time to be ‘semantically inconsistent’ with these other definite‐NPs. To some this is an ironclad proof of the irrelevance of semantics to linguistics. We argue that both of the initial claims are wrong: average‐NPs are not ‘linguistically identical’ to other definite‐NPs but instead show a number of interesting divergences, and we provide a plausible semantic (...) account for them that is not ‘semantically inconsistent’ with the account afforded other definite‐NPs but in fact blends quite nicely with one standard account of the semantics for NPs. (shrink)
The logic of paradox, LP, is a first-order, three-valued logic that has been advocated by Graham Priest as an appropriate way to represent the possibility of acceptable contradictory statements. Second-order LP is that logic augmented with quantification over predicates. As with classical second-order logic, there are different ways to give the semantic interpretation of sentences of the logic. The different ways give rise to different logical advantages and disadvantages, and we canvass several of these, concluding that it will be extremely (...) difficult to appeal to second-order LP for the purposes that its proponents advocate, until some deep, intricate, and hitherto unarticulated metaphysical advances are made. (shrink)
In ‘On Denoting’ and to some extent in ‘Review of Meinong and Others, Untersuchungen zur Gegenstandstheorie und Psychologie’, published in the same issue of Mind (Russell, 1905a,b), Russell presents not only his famous elimination (or contextual deﬁ nition) of deﬁ nite descriptions, but also a series of considerations against understanding deﬁ nite descriptions as singular terms. At the end of ‘On Denoting’, Russell believes he has shown that all the theories that do treat deﬁ nite descriptions as singular terms fall (...) logically short: Meinong’s, Mally’s, his own earlier (1903) theory, and Frege’s. (He also believes that at least some of them fall short on other grounds—epistemological and metaphysical—but we do not discuss these criticisms except in passing). Our aim in the present paper is to discuss whether his criticisms actually refute Frege’s theory. We ﬁ rst attempt to specify just what Frege’s theory is and present the evidence that has moved scholars to attribute one of three different theories to Frege in this area. We think that each of these theories has some claim to be Fregean, even though they are logically quite different from each other. This raises the issue of determining Frege’s attitude towards these three theories. We consider whether he changed his mind and came to replace one theory with another, or whether he perhaps thought that the different theories applied to different realms, for example, to natural language versus a language for formal logic and arithmetic. We do not come to any hard and fast conclusion here, but instead just note that all these theories treat deﬁ nite descriptions as singular terms, and that Russell proceeds as if he has refuted them all. After taking a brief look at the formal properties of the Fregean theories (particularly the logical status of various sentences containing nonproper deﬁ - nite descriptions) and comparing them to Russell’s theory in this regard, we turn to Russell’s actual criticisms in the above-mentioned articles to examine the extent to which the criticisms hold.. (shrink)
Simple mass nouns are words like ‘water’, ‘furniture’ and ‘gold’. We can form complex mass noun phrases such as ‘dirty water’, ‘leaded gold’ and ‘green grass’. I do not propose to discuss the problems in giving a characterization of the words that are mass versus those that are not. For the purposes of this paper I shall make the following decrees: (a) nothing that is not a noun or noun phrase can be mass, (b) no abstract noun phrases are considered (...) mass, (c) words like ‘thing’, ‘entity’ and ‘object’ are not mass, (d) I shall not consider such words as ‘stuff’, ‘substance or ‘matter’, (e) measures on mass nouns (like ‘gallon of gasoline’, ‘blade of grass’, etc.) are not considered, (f) plurals of count terms are not considered mass. Within these limitations, we can say generally that mass noun phrases are those phrases that ‘much’ can be prefexed to, by ‘many’ cannot be prefexed to, without an0maly.l Semantically, such phrases usually have the property of collectiveness- they are true of any sum of things of which they are true ; and of divisiveness - they are true of any part (down to a certain limit) of things of which they are true. All of this, however, is only ‘generally speaking’ - I shall mostly use only the simple examples given above and ignore the problems in giving a complete characterization of mass nouns. In the paper I want to discuss some problems involved in casting English sentences containing mass nouns into some artificial language; but in order to do this we should have some anchoring framework on which to justify or reject a given proposal. The problem of finding an adequate language can be viewed as a case of translation (from English to the artificial language), where the translation relation must meet certain requirements. I shall suggest five such requirements; others could be added. (shrink)
Fuzzy logics are systems of logic with infinitely many truth values. Such logics have been claimed to have an extremely wide range of applications in linguistics, computer technology, psychology, etc. In this note, we canvass the known results concerning infinitely many valued logics; make some suggestions for alterations of the known systems in order to accommodate what modern devotees of fuzzy logic claim to desire; and we prove some theorems to the effect that there can be no fuzzy logic which (...) will do what its advocates want. Finally, we suggest ways to accommodate these desires in finitely many valued logics. (shrink)
Logics that have many truth values—more than just True and False—have been argued to be useful in the analysis of very many philosophical and linguistic puzzles. In this paper, which is a followup to, we will start with a particularly well-motivated four-valued logic that has been studied mainly in its propositional and first-order versions. And we will then investigate its second-order version. This four-valued logic has two natural three-valued extensions: what is called a “gap logic”, and what is called a (...) “glut logic”. We mention various results about the second-order version of these logics as well. And we then follow our earlier papers, where we had added a specific conditional connective to the three valued logics, and now add that connective to the four-valued logic under consideration. We then show that, although this addition is “conservative” in the sense that no new theorems are generated in the four-valued logic unless they employ this new conditional in their statement, nevertheless the resulting second-order versions of these logics with and without the conditional are quite different in important ways. We close with a moral for logical investigations in this realm. (shrink)
We investigate the notion of relevance as it pertains to ‘commonsense’, subjunctive conditionals. Relevance is taken here as a relation between a property (such as having a broken wing) and a conditional (such as birds typically fly). Specifically, we explore a notion of ‘causative’ relevance, distinct from ‘evidential’ relevance found, for example, in probabilistic approaches. A series of postulates characterising a minimal, parsimonious concept of relevance is developed. Along the way we argue that no purely logical account of relevance (even (...) at the metalevel) is possible. Finally, and with minimal restrictions, an explicit definition that agrees with the postulates is given. (shrink)
Vagueness: an expression is vague if and only if it is possible that it give rise to a “borderline case.” A borderline case is a situation in which the application of a particular expression to a (name of) a particular object does not generate an expression with a definite TRUTH-VALUE. That is, the piece of language in question neither applies to the object nor fails to apply. Although such a formulation leaves it open what the pieces of language might be (...) (whole sentences, individual words, NAMES or SINGULAR TERMS, PREDICATES or GENERAL TERMS), most discussions have focussed on vague general terms and have considered other types of terms to be non-vague. (Exceptions to this have called attention to the possibility of vague objects, thereby making the designation relation for singular terms be vague). The formulation also leaves open the possible causes for the expression not to have a definite truth value. If this indeterminacy is due to there being insufficient information available to determine applicability or non-applicability of the term (that is, we’re convinced the term either does or doesn’t apply, but we just don’t have enough information to determine which), then this is sometimes called “epistemic vagueness.” It is somewhat misleading to call this vagueness, for unlike true vagueness, this epistemic vagueness disappears if more information is brought into the situation. (‘There are 1.89∞106 stars in the sky’ epistemically vague but is not vague in the generally accepted sense of the term). ‘Vagueness’ may also be used to characterize non-linguistic items such as CONCEPTS, MEMORIES, and OBJECTS ... as well as such semi-linguistic items as STATEMENTS and PROPOSITIONS. Many of the issues involved in discussing the topic of vagueness impinge upon other philosophical topics, such as the existence of TRUTH-VALUE GAPS (declarative sentences which are neither TRUE nor FALSE) and the plausibility of MANY-VALUED LOGIC.. (shrink)
The Law of Non-Contradiction - that no contradiction can be true - has been a seemingly unassailable dogma since the work of Aristotle. It is an assumption challenged from a variety of angles in this collection of original papers. Twenty-three of the world's leading experts investigate the 'law', considering arguments for and against it and discuss methodological issues that arise. The result is a balanced inquiry into a venerable principle of logic, one that raises questions at the very centre of (...) logic itself. (shrink)
In an earlier paper entitled Synonymous Logics, the authors attempted to show that there are two modal logics so that each is exactly translatable into the other, but they are not translationally equivalent. Unfortunately, there is an error in the proof of this result. The present paper provides a new example of two such logics, and a proof of the result claimed in the earlier paper.
The Society tor Exact Philosop-hy was founded :in·l97D at a meeting held at McGill University in Montreal on 4-5 November at which was organised iby Mario Bunge. Funding for the meeting iwas provided by SDiii the International Union of Hsistory and Philosophy of cience (vson..
Although resolution-based inference is perhaps the industry standard in automated theorem proving, there have always been systems that employed a different format. For example, the Logic Theorist of 1957 produced proofs by using an axiomatic system, and the proofs it generated would be considered legitimate axiomatic proofs; Wang’s systems of the late 1950’s employed a Gentzen-sequent proof strategy; Beth’s systems written about the same time employed his semantic tableaux method; and Prawitz’s systems of again about the same time are often (...) said to employ a natural deduction format. [See Newell, et al (1957), Beth (1958), Wang (1960), and Prawitz et al (1960)]. Like sequent proof systems and tableaux proof systems, natural deduction systems retain.. (shrink)
THINKER is an automated natural deduction first-order theorem proving program. This paper reports on how it was adapted so as to prove theorems in modal logic. The method employed is an indirect semantic method, obtained by considering the semantic conditions involved in being a valid argument in these modal logics. The method is extended from propositional modal logic to predicate modal logic, and issues concerning the domain of quantification and existence in a world's domain are discussed. Finally, we look at (...) the very interesting issues involved with adding identity to the theorem prover in the realm of modal predicate logic. Various alternatives are discussed. (shrink)
In the last ten years or so, talk of “possible worlds” has become decidedly more fashionable from a logical point of view. And here fashion is well-justified: both from a logical and from a metaphysical point of view, the work of Saul Kripke on the concept of a possible world is as challenging as any contributions since the time of Leibniz himself. It is only fair, then, and indirectly flattering, to complain that Kripke is limiting himself as a major philosopher (...) in ways which Leibniz saw to be essential to balanced philosophy. We confine ourselves in this paper to just two crucial dimensions of philosophical depth when it comes to the concept of possible worlds. On the one hand, Leibniz saw the concept as one which raises very usefully many traditional problems of ethics, including those of free will and theodicy. On the other hand, Leibniz saw that questions about possible worlds may call for a radical approach to sane ontology. Such an approach, “revisionary metaphysics,” systematically reinterprets rather than dogmatically regurgitates the convictions of Plain Men and Theoretical Physicists about—to take a favorite example of Kripke’s—this single table’s being made of many distinct molecules. (shrink)
Contrary to the claims of owen, frede, and many other platonic scholars, there is a straight forward way to explicate plato's "sophist" as having 'heteron' first be understood as "non-identical" and after 257b or so be understood as "incompatible." this should encourage scholars who prefer the "incompatibility" reading but don't see how to get the required change of meaning.