Syntax
In linguistics, syntax (from Ancient Greek συν- syn-, “together”, and τάξις táxis, “arrangement”) is the study of the rules that govern the structure of sentences, and which determine their relative grammaticality. The term syntax can also be used to refer to these rules themselves, as in “the syntax of a language”. Modern research in syntax attempts to describe languages in terms of such rules, and, for many practitioners, to find general rules that apply to all languages. Since the field of syntax attempts to explain grammaticality judgments, and not provide them, it is unconcerned with linguistic prescription.
Though all theories of syntax takes human as their object of study, there are some significant differences in outlook. Modern linguists see syntax as a branch of biology, since they conceive syntax as the study of linguistic knowledge as embodied in the human mind/brain. Others (e.g. Gerald Gazdar) take a more Platonistic view, regarding syntax as the study of an abstract formal system. [1]; others also (e.g. Joseph Greenberg) consider grammar as a taxonomical device to reach broad generalizations among languages.
Contents |
Early history
Works on grammar were of course being written long before modern syntax came about; the Aṣṭādhyāyī of Pāṇini is often cited as an example of a pre-modern work that approaches the sophistication of a modern syntactic theory.[1] In the West, the school of thought that came to be known as ‘traditional grammar’ began with the work of Dionysius Thrax.
For centuries, work in syntax was dominated by a framework known as grammaire générale, first expounded in 1660 by Antoine Arnauld in a book of the same title. This system took as its basic premise the assumption that language is a direct reflection of thought processes, and that hence there is a single most natural way to express a thought (which, coincidentally, was exactly the way it was expressed in French).
However, in the 19th century, with the development of historical-comparative linguistics, linguists began to realize the sheer diversity of human language, and to question fundamental assumptions about the relation between language and logic. It became apparent that there was no such thing as a most natural way to express a thought, and logic could no longer be relied upon as a base for studying the structure of language.
The Port-Royal grammar modelled the study of syntax on that of logic (indeed, large parts of the Port-Royal Logic were copied or adapted from the Grammaire générale[2]). Syntactic categories were identified with logical ones, and all sentences were analysed into the form "Subject-Copula-Predicate". Initially, this view was adopted even by the early comparative linguists (e.g., Bopp).
The central role of syntax within theoretical linguistics became clear only in the last century which could reasonably called the "century of syntactic theory" as far as linguistics is concerned. For a detailed and critical survey of the history of syntax in the last two centuries see the monumental work by Graffi 2001.
Modern theories
Generative grammar constitutes one of the most innovative ideas in linguistics since its origin. There are two features shared by most theories of formal syntax. First, they hierarchically group subunits into constituent units (usually referred to as phrases). Second, they provide a system of rules to explain why certain utterances seem more acceptable or grammatical than others. Most formal theories of syntax also offer explanations of the systematic relationships between syntax and semantics, in other words, between form and meaning.
In the framework of transformational-generative grammar (of which government and binding theory and minimalism are recent developments), the structure of a sentence is represented by phrase structure trees, otherwise known as phrase markers or tree diagrams. Such trees provide information about the sentences they represent by showing the hierarchical relations between their component parts. Dependency grammar is a different type of generative grammar in which structure is determined by the relation between a word (a head) and its dependents. One difference from phrase structure grammar is that dependency grammar does not have phrasal categories. Algebraic syntax is a type of dependency grammar.
A modern approach to combining accurate descriptions of the grammatical patterns of language with their function in context is that of systemic functional grammar, an approach originally developed by Michael A.K. Halliday in the 1960s. Systemic-functional grammar is related both to feature-based approaches such as Head-driven phrase structure grammar and to the older functional traditions of European schools of linguistics such as British Contextualism and the Prague School.
Tree-adjoining grammar is a grammar formalism with interesting mathematical properties which has sometimes been used as the basis for the syntactic description of natural language. In monotonic and monostratal frameworks, variants of are often preferred formalisms.
With the publication of Gold's Theorem[3] 1967 it was claimed that grammars for natural languages governed by deterministic rules could not be learned based on positive instances alone. This was part of the argument from the poverty of stimulus, presented in 1980[4] and implicit since the early works by Chomsky of the 1950s. This led to the nativist view, that a form of grammar (including a complete conceptual lexicon in certain versions) were hardwired from birth.
A grammar is a description of the syntax of a language. Theoretical models rarely consider the language in use, as revealed by corpus linguistics, but focus on a mental language or i-language as its "proper" object of study. In contrast, the "empirically responsible"[5] approach to syntax seeks to construct grammars that will explain language in use. A key class of grammars in the latter tradition are the stochastic context-free grammars.
A problem faced in any formal syntax is that often more than one production rule may apply to a structure, thus resulting in a conflict. The greater the coverage, the higher this conflict, and all grammarians (starting with Panini) have spent considerable effort devising a prioritization for the rules, which usually turn out to be defeasible. Another difficulty is overgeneration, where unlicensed structures are also generated. Probabilistic grammars circumvent these problems by using the frequency of various productions to order them, resulting in a "most likely" (winner-take-all) interpretation, which by definition, is defeasible given additional data. As usage patterns are altered in shifts, these probabilistic rules can be re-learned, thus upgrading the grammar.
One may construct a probabilistic grammar from a traditional formal syntax by assigning each non-terminal a probability taken from some distribution, to be eventually estimated from usage data. On most samples of broad language, probabilistic grammars that tune these probabilities from data typically outperform hand-crafted grammars (although some rule-based grammars are now approaching the accuracies of PCFG).
Recently, probabilistic grammars appear to have gained some cognitive plausibility. It is well known that there are degrees of difficulty in accessing different syntactic structures (e.g. the Accessibility Hierarchy for relative clauses). Probabilistic versions of minimalist grammars have been used to compute information-theoretic entropy values which appear to correlate well with psycholinguistic data on understandability and production difficulty.[6]
Statistical grammars are not subject to Gold's theorem since the learning is incremental.
See also
Syntactic terms
- Adjective
- Adjunct
- Adverb
- Antecedent-contained deletion
- Appositive
- Article
- Aspect
- Auxiliary verb
- Case
- Clause
- Closed class word
- Comparative
- Complement
- Compound noun and adjective
- Conjugation
- Conjunction
- Dangling modifier
- Declension
- Determiner
- Dual (form for two)
- Expletive
- Function word
- Gender
- Gerund
- Infinitive
- Measure word (classifier)
- Modal particle
- Movement paradox
- Modifier
- Mood
- Noun
- Number
- Object
- Open class word
- Parasitic gap
- Part of speech
- Particle
- Person
- Phrase
- Phrasal verb
- Plural
- Predicate (also verb phrase)
- Predicative (adjectival or nominal)
- Preposition
- Personal pronoun
- Pronoun
- Restrictiveness
- Sandhi
- Sentence (linguistics)
- Singular
- Subject
- Superlative
- Tense
- Uninflected word
- Verb
- Voice
- Wh-movement
- Word order
Notes
- ^ Fortson IV, Benjamin W. (2004). Indo-European Language and Culture: An Introduction. Blackwell, 186. ISBN 1-4051-0315-9 (hb); 1-4051-0316-7 (pb). “[The Aṣṭādhyāyī] is a highly precise and thorough description of the structure of Sanskrit somewhat resembling modern generative grammar…[it] remained the most advanced linguistic analysis of any kind until the twentieth century.”
- ^ Arnauld, Antoine (1683). La logique, 5th ed., Paris: G. Desprez, 137. “Nous avons emprunté…ce que nous avons dit…d'un petit Livre…sous le titre de Grammaire générale.”
- ^ Gold, E. (1967). Language identification in the limit. Information and Control 10, 447-474.
- ^ Chomsky, N. (1980). Rules and representations Oxford: Basil Blackwell.
- ^ George Lakoff and Mark Johnson (1999). Philosophy in the Flesh: The embodied mind and its challenge to Western thought. Part IV.. New York: Basic Books..
- ^ John Hale (2006). "Uncertainty About the Rest of the Sentence". Cognitive Science 30: 643-672. DOI:doi:10.1207/s15516709cog0000_64.
jual.
References
- Brown, Keith; Jim Miller (eds.) (1996). Concise Encyclopedia of Syntactic Theories. New York: Elsevier Science. ISBN 0-08-042711-1.
- Freidin, Robert; Howard Lasnik (eds.) (2006). Syntax, Critical Concepts in Linguistics. New York: Routledge. ISBN 0-415-24672-5.
- Graffi, Giorgio (2001). 200 Years of Syntax. A Critical Survey, Studies in the History of the Language Sciences 98. Amsterdam: Benjamins. ISBN 90-272-4587-8.
External links
- The syntax of natural language (Beatrice Santorini & Anthony Kroch, University of Pennsylvania)
- Learn Programming Language Syntax
- Various syntactic constructs used in computer programming languages