Jump to content

Phrase structure grammar

From Wikipedia, the free encyclopedia

The term phrase structure grammar was originally introduced by Noam Chomsky as the term for grammar studied previously by Emil Post and Axel Thue (Post canonical systems). Some authors, however, reserve the term for more restricted grammars in the Chomsky hierarchy: context-sensitive grammars or context-free grammars. In a broader sense, phrase structure grammars are also known as constituency grammars. The defining character of phrase structure grammars is thus their adherence to the constituency relation, as opposed to the dependency relation of dependency grammars.

History

[edit]

In 1956, Chomsky wrote, "A phrase-structure grammar is defined by a finite vocabulary (alphabet) Vp, and a finite set Σ of initial strings in Vp, and a finite set F of rules of the form: X → Y, where X and Y are strings in Vp."[1]

Constituency relation

[edit]

In linguistics, phrase structure grammars are all those grammars that are based on the constituency relation, as opposed to the dependency relation associated with dependency grammars; hence, phrase structure grammars are also known as constituency grammars.[2] Any of several related theories for the parsing of natural language qualify as constituency grammars, and most of them have been developed from Chomsky's work, including

Further grammar frameworks and formalisms also qualify as constituency-based, although they may not think of themselves as having spawned from Chomsky's work, e.g.

The fundamental trait that these frameworks all share is that they view sentence structure in terms of the constituency relation. The constituency relation derives from the subject-predicate division of Latin and Greek grammars[citation needed] that is based on term logic and reaches back to Aristotle[3] in antiquity. Basic clause structure is understood in terms of a binary division of the clause into subject (noun phrase NP) and predicate (verb phrase VP).

The binary division of the clause results in a one-to-one-or-more correspondence. For each element in a sentence, there are one or more nodes in the tree structure that one assumes for that sentence. A two word sentence such as Luke laughed necessarily implies three (or more) nodes in the syntactic structure: one for the noun Luke (subject NP), one for the verb laughed (predicate VP), and one for the entirety Luke laughed (sentence S). The constituency grammars listed above all view sentence structure in terms of this one-to-one-or-more correspondence.

Constituency and dependency relations
Constituency and dependency relations

Dependency relation

[edit]

By the time of Gottlob Frege, a competing understanding of the logic of sentences had arisen. Frege rejected the binary division of the sentence and replaced it with an understanding of sentence logic in terms of logical predicates and their arguments. On this alternative conception of sentence logic, the binary division of the clause into subject and predicate was not possible. It therefore opened the door to the dependency relation (although the dependency relation had also existed in a less obvious form in traditional grammars long before Frege). The dependency relation was first acknowledged concretely and developed as the basis for a comprehensive theory of syntax and grammar by Lucien Tesnière in his posthumously published work Éléments de syntaxe structurale (Elements of Structural Syntax).[4]

The dependency relation is a one-to-one correspondence: for every element (word or morph) in a sentence, there is just one node in the syntactic structure. The distinction is thus a graph-theoretical distinction. The dependency relation restricts the number of nodes in the syntactic structure of a sentence to the exact number of syntactic units (usually words) that that sentence contains. Thus the two word sentence Luke laughed implies just two syntactic nodes, one for Luke and one for laughed. Some prominent dependency grammars are listed here:

Since these grammars are all based on the dependency relation, they are by definition NOT phrase structure grammars.

Non-descript grammars

[edit]

Other grammars generally avoid attempts to group syntactic units into clusters in a manner that would allow classification in terms of the constituency vs. dependency distinction. In this respect, the following grammar frameworks do not come down solidly on either side of the dividing line:

See also

[edit]

Notes

[edit]
  1. ^ Chomsky, Noam (1956). "Three models for the description of language" (PDF). IRE Transactions on Information Theory: 117. Archived from the original (PDF) on 2023-05-03.
  2. ^ Matthews (1981:71ff.) provides an insightful discussion of the distinction between constituency- and dependency-based grammars. See also Allerton (1979:238f.), McCawley (1988:13), Mel'cuk (1988:12-14), Borsley (1991:30f.), Sag and Wasow (1999:421f.), van Valin (2001:86ff.).
  3. ^ Bobzien, Susanne. "Ancient Logic". Stanford Encyclopedia of Philosophy. Stanford. Retrieved 4 March 2018.
  4. ^ See Tesnière (1959).

References

[edit]
  • Allerton, D. 1979. Essentials of grammatical theory. London: Routledge & Kegan Paul.
  • Borsley, R. 1991. Syntactic theory: A unified approach. London: Edward Arnold.
  • Chomsky, Noam 1957. Syntactic structures. The Hague/Paris: Mouton.
  • Matthews, P. Syntax. 1981. Cambridge, UK: Cambridge University Press, ISBN 978-0521297097.
  • McCawley, T. 1988. The syntactic phenomena of English, Vol. 1. Chicago: The University of Chicago Press.
  • Mel'cuk, I. 1988. Dependency syntax: Theory and practice. Albany: SUNY Press.
  • Sag, I. and T. Wasow. 1999. Syntactic theory: A formal introduction. Stanford, CA: CSLI Publications.
  • Tesnière, Lucien 1959. Éleménts de syntaxe structurale. Paris: Klincksieck.
  • van Valin, R. 2001. An introduction to syntax. Cambridge, UK: Cambridge University Press.