Two kinds of theoretical framework for syntax are encountered in current linguistics. One emerged from the mathematization of proof theory, and is referred to here as generative-enumerative syntax (GES). A less explored alternative stems from the semantic side of logic, and is here called model-theoretic syntax (MTS). I sketch the outlines of each, and give a capsule summary of some mathematical results pertaining to the latter. I then briefly survey some diverse types of evidence suggesting that in some ways MTS (...) seems better suited to theorizing about the relevant linguistic phenomena. (shrink)
Notice that this paper has not claimed that all natural languages are CFL's. What it has shown is that every published argument purporting to demonstrate the non-context-freeness of some natural language is invalid, either formally or empirically or both.18 Whether non-context-free characteristics can be found in the stringset of some natural language remains an open question, just as it was a quarter century ago.Whether the question is ultimately answered in the negative or the affirmative, there will be interesting further questions (...) to ask. If it turns out that natural languages are indeed always CFL's, it will be reasonable to ask whether this helps to explain why speakers apparently recognize so quickly whether a presented utterance corresponds to a grammatical sentence or not, and to associate structural and semantic details with it. It might also be reasonable to speculate about the explanation for the universally context-free character of the languages used by humans, and to wonder whether evolutionary biological factors are implicated in some way (Sampson (1979) could be read in this light). And naturally, it will be reasonable to pursue the program put forward by Gazdar (1981, in press) to see to what extent CFL-inducing grammatical devices can be exploited to yield insightful descriptions of natural languages that capture generalizations in revealing ways.If a human language that is not a CFL is proved to exist, on the other hand, a different question will be raised: give the non-context-free character of human languages in general, why has this property been so hard to demonstrate that it has taken over twenty-five years to bring it to light since the issue was first explicitly posed? If human languages do not have to be CFL's, why do so many (most?) of them come so close to having the property of context-freeness? And, since the CFL's certainly constitute a very broad class of mathematically natural and computationally tractable languages, what property of human beings or their communicative or cognitive needs is it that has caused some linguistic communities to reach beyond the boundaries of this class in the course of evolving a linguistic system?Either way, we shall be interested to see our initial question resolved, and further questions raised. One cautionary word should be said, however, about the implications (or lack of them) that the answer will have for grammatical studies. Chomsky has repeatedly stated that he does not see weak generative capacity as a theme of central importance in the theory of grammar, and we agree. It is very far from being the case that the recent resurgence of interest in exploring the potential of CF-PSG or equivalent systems will, or should, be halted dead in its tracks by the discovery (if it is ever forthcoming) that some natural language is not a CFL. In the area of parsing, for instance, it seems possible that natural languages are not only parsed on the basis of constituent structure such as a CF-PSG would assign, but are parsed as if they were finite state languages (see Langendoen (1975) and Church (1980) for discussion along these lines). That is, precisely those construction-types that figure in the various proofs that English is not an FSL appear to cause massive difficulty in the human processing system; the sentences crucial to the proofs are for the most part unprocessable unless they are extremely short (yet the arguments for English not being and FSL only go through if length is not an issue). This means that in practice properties of finite state grammars are still of great potential importance to linguistic theory despite the fact that they do not provide the framework for defining the total class of grammatical sentences. The same would almost certannly be true of CF-PSG's if they were shown to be inadequate in a similar sense. It is highly unlikely that the advances made so far in far in phrase structure description could be nullified by a discovery about weak generative capacity. Moreover, there are known to be numerous ways in which the power of CF-PSG's can be marginally enhanced to permit, for example, xx languages to be generated without allowing anything like the full class of recursively enumerable or even context-sensitive languages (see Hopcroft and Ullmann (1979), Chapter 14) for an introduction to this topic, noting especially Figure 14.7 on p. 393. The obvious thing to do if natural languages were ever shown not to be CFL's in the general case would be to start exploring such minimal enhancements of expressive power to determine exactly what natural languages call for in this regard and how it could be effectively but parsimoniously provided in a way that closely modelled human linguistic capacities.In the meantime, it seems reasonable to assume that the natural languages are a proper subset of the infinite-cardinality CFL's until such time as they are validly shown not to be. (shrink)
We explore the formal foundations of recent studies comparing aural pattern recognition capabilities of populations of human and non-human animals. To date, these experiments have focused on the boundary between the Regular and Context-Free stringsets. We argue that experiments directed at distinguishing capabilities with respect to the Subregular Hierarchy, which subdivides the class of Regular stringsets, are likely to provide better evidence about the distinctions between the cognitive mechanisms of humans and those of other species. Moreover, the classes of the (...) Subregular Hierarchy have the advantage of fully abstract descriptive (model-theoretic) characterizations in addition to characterizations in more familiar grammar- and automata-theoretic terms. Because the descriptive characterizations make no assumptions about implementation, they provide a sound basis for drawing conclusions about potential cognitive mechanisms from the experimental results. We review the Subregular Hierarchy and provide a concrete set of principles for the design and interpretation of these experiments. (shrink)
Chomsky’s highly influential Syntactic Structures ( SS ) has been much praised its originality, explicitness, and relevance for subsequent cognitive science. Such claims are greatly overstated. SS contains no proof that English is beyond the power of finite state description (it is not clear that Chomsky ever gave a sound mathematical argument for that claim). The approach advocated by SS springs directly out of the work of the mathematical logician Emil Post on formalizing proof, but few linguists are aware of (...) this, because Post’s papers are not cited. Chomsky’s extensions to Post’s systems are not clearly defined, and the arguments for their necessity are weak. Linguists have also overlooked Post’s proofs of the first two theorems about effects of rule format restrictions on generative capacity, published more than ten years before SS was published. (shrink)
We examine an argument for the non-context-freeness of English that has received virtually no discussion in the literature. It is based on adjuncts of the form 'X or no X', where X is a nominal. The construction has been held to exemplify unbounded syntactic reduplication. We argue that although the argument can be made in a mathematically valid form, its empirical basis is not secure. First, the claimed unbounded syntactic identity between nominals does not always hold in attested cases, and (...) second, an understanding of the semantics of the construction removes the necessity of making reference to any syntactic reduplication. (shrink)
Evans & Levinson's (E&L's) major point is that human languages are intriguingly diverse rather than (like animal communication systems) uniform within the species. This does not establish a about language universals, or advance the ill-framed pseudo-debate over universal grammar. The target article does, however, repeat a troublesome myth about Fitch and Hauser's (2004) work on pattern learning in cotton-top tamarins.
Modern weaponry is often too complex for unaided human operation, and is largely or totally controlled by computers. But modern software, particularly artificial intelligence software, exhibits such complexity and inscrutability that there are grave dangers associated with its use in non-benign applications. Recent efforts to make computer systems more accessible to military personnel through natural language processing systems, as proposed in the Strategic Computing Initiative of the Defense Advanced Research Projects Agency, increases rather than decreases the dangers of unpredictable behavior. (...) Defense systems constitute, in fact, a paradigm case of the wrong kind of application for this technology. This cannot be expected to change, since the unpredictability stems from inherent properties of computer systems and of natural languages. (shrink)