File Information

File: 05-lr/acl_arc_1_sum/cleansed_text/xml_by_section/metho/06/w06-1503_metho.xml

Size: 15,030 bytes

Last Modified: 2025-10-06 14:10:42

<?xml version="1.0" standalone="yes"?>
<Paper uid="W06-1503">
  <Title>CAT V I +</Title>
  <Section position="3" start_page="0" end_page="17" type="metho">
    <SectionTitle>
1 An Overview of Metagrammars
</SectionTitle>
    <Paragraph position="0"> A metagrammar (MG) factors common properties of TAG elementary trees to avoid redundancy, ease grammar development, and expand coverage with minimal effort: typically, from a compact manually encoded MG of a few dozen classes, one or more TAGs with several hundreds of elementary trees are automatically generated. This is appealing from a grammar engineering point of view, and also from a linguistic point of view: cross-linguistic generalizations are expressed directly in the MG. In this paper, we extend some earlier work on multilingual MGs (Candito, 1998; Kinyon and Rambow, 2003) by proposing cross-linguistic and framework-neutral syntactic invariants, which we apply to TAG. We focus on the verb-second phenomenon as a prototypical example of cross-language variation.</Paragraph>
    <Paragraph position="1"> The notion of Metagrammar Metagrammars were first introduced by Candito (1996) to manually encode syntactic knowledge in a compact and abstract class hierarchy which supports multiple inheritance, and from which a TAG is automatically generated offline. Candito's class hierarchy imposes a general organization of syntax into three dimensions: a0 Dimension 1: to encode initial subcategorization frames i.e. TAG tree families a0 Dimension 2: to encode valency alternations / redistribution of syntactic functions a0 Dimension 3: to encode the surface realization of arguments.</Paragraph>
    <Paragraph position="2"> Each class in the MG hierarchy is associated with a partial tree description The tool computes a set of well-formed classes by combining exactly one terminal class from dimension 1, one terminal class from dimension 2, and a1 terminal classes from dimensions 3 (a1 being the number of arguments subcategorized by the lexical head anchoring the elementary tree(s) generated). The conjunction of the tree descriptions associated with each well-formed class in the set yields a minimal satisfying description, which results in the generation of one or more elementary trees. Candito's tool was used to develop a large TAG for French as well as a medium-size TAG for Italian Candito (1999), so multilinguality was addressed from the start, but each language had its dedicated hierarchy, with no sharing of classes despite the obvious similarities between Italian and French. A related approach was proposed by (Xia, 2001); the work of Evans, Gazdar, and Weir (2000) also has some common elements with MG.</Paragraph>
    <Paragraph position="3"> Framework- and language-neutral syntactic invariants Using a MG, and following Candito, we can postulate cross-linguistic and crossframework syntactic invariants such as:  tions, of syntactic phenomena which do not alter valency, such as wh-movement (Candito's dimension 3).</Paragraph>
    <Paragraph position="4"> These invariants -- unlike other framework-specific syntactic assumptions such as the existence of &amp;quot;movement&amp;quot; or &amp;quot;wh-traces&amp;quot; -- are accepted by most if not all existing frameworks, even though the machinery of a given framework may not necessarily account explicitly for each invariant. For instance, TAG does not have an explicit notion of syntactic function: although by convention node indices tend to reflect a function, it is not enforced by the framework's machinery.1 Hypertags Based on such framework- and language-neutral syntactic properties, Kinyon (2000) defined the notion of Hypertag (HT), a combination of Supertags (ST) Srinivas (1997) and of the MG. A ST is a TAG elementary tree, which provides richer information than standard POS tagging, but in a framework-specific manner (TAG), and also in a grammar-specific manner since a ST tagset can't be ported from one TAG to another TAG. A HT is an abstraction of STs, where the main syntactic properties of any given ST is encoded in a general readable Feature Structure (FS), by recording which MG classes a ST inherited from when it was generated. Figure 1 illustrates the a0ST, HTa1 pair for Par qui sera accompagn'ee Marie 'By whom will Mary be accompanied'. We see that a HT feature structure directly reflects the MG organization, by having 3 features &amp;quot;Dimension 1&amp;quot;, &amp;quot;Dimension 2&amp;quot; and &amp;quot;Dimension 3&amp;quot;, where each feature takes its value from the MG terminal classes used to generate a given ST.</Paragraph>
    <Paragraph position="5"> The XMG Tool Candito's tool brought a significant linguistic insight, therefore we essentially retain the above-mentioned syntactic invariants.</Paragraph>
    <Paragraph position="6"> However, more recent MG implementations have been developed since, each adding its significant contribution to the underlying metagrammatical hypothesis.</Paragraph>
    <Paragraph position="7"> In this paper, we use the eXtensible MetaGrammar (XMG) tool which was developed by Crabb'e 1But several attempts have been made to explicitly add functions to TAG, e.g. by Kameyama (1986) to retain the benefits of both TAG and LFG, or by Prolo (2006) to account for the coordination of constituents of different categories, yet sharing the same function.</Paragraph>
    <Paragraph position="8">  dito's MetaGrammar compiler (2005). In XMG, an MG consists of a set of classes similar to those in object-oriented programming, which are structured into a multiple inheritance hierarchy. Each class specifies a partial tree description (expressed by dominance and precedence constraints). The nodes of these tree fragment descriptions may be annotated with features. Classes may instantiate each other, and they may be parametrized (e.g., to hand down features like the grammatical function of a substitution node). The compiler unifies the instantiations of tree descriptions that are called. This unification is additionally guided by node colors, constraints that specify that a node must not be unified with any other node (red), must be unified (white), or may be unified, but only with a white node (black).</Paragraph>
    <Paragraph position="9"> XMG allows us to implement a hierarchy similar to that of Candito, but it also allows us to modify and extend it, as no structural assumptions about the class hierarchy are hard-coded.</Paragraph>
  </Section>
  <Section position="4" start_page="17" end_page="17" type="metho">
    <SectionTitle>
2 The V2 Phenomenon
</SectionTitle>
    <Paragraph position="0"> The Verb-Second (V2) phenomenon is a well-known set of data that demonstrates small-scale cross-linguistic variation. The examples in (1) show German, a language with a V2-constraint: (1a) is completely grammatical, while (1b) is not.</Paragraph>
    <Paragraph position="1"> This is considered to be due to the fact that the finite verb is required to be located in &amp;quot;second position&amp;quot; (V2) in German. Other languages with a V2 constraint include Dutch, Yiddish, Frisian, Icelandic, Mainland Scandinavian, and Kashmiri.</Paragraph>
    <Paragraph position="2">  Int.: 'On the path, the boy sees a duck.' Interestingly, these languages differ with respect to how exactly the constraint is realized. Rambow and Santorini (1995) present data from the mentioned languages and provide a set of parameters that account for the exhibited variation. In the following, for the sake of brevity, we will confine the discussion to two languages: German, and Yiddish. The German data is as follows (we do not repeat (1a) from above):  While main clauses exhibit V2 in German, embedded clauses with complementizers are verb-final (2b). In contrast, Yiddish embedded clauses must also be V2 (3c).</Paragraph>
  </Section>
  <Section position="5" start_page="17" end_page="17" type="metho">
    <SectionTitle>
3 Handling V2 in the Metagrammar
</SectionTitle>
    <Paragraph position="0"> It is striking that the basic V2 phenomenon is the same in all of these languages: the verb can appear in either its underlying position, or in second position (or, in some cases, third). We claim that what governs the appearance of the verb in these different positions (and thus the cross-linguistic differences) is that the heads--the verbal head and functional heads such as auxiliaries and complementizers--interact in specific ways. For example, in German a complementizer is not compatible with a verbal V2 head, while in Yiddish it is. We express the interaction among heads by assigning the heads different values for a set of features. Which heads can carry which feature values is a language-specific parameter. Our implementation is based on the previous pen-and-pencil analysis of Rambow and Santorini (1995), which we have modified and extended.</Paragraph>
    <Paragraph position="1"> The work we present in this paper thus has a threefold interest: (1) we show how to handle an important syntactic phenomenon cross-linguistically in a MG framework; (2) we partially validate, correct, and extend a previously proposed linguistically-motivated analysis; and (3) we provide an initial fragment of a MG implementation from which we generate TAGs for languages which are relatively less-studied and for which no TAG currently exists (Yiddish).</Paragraph>
  </Section>
  <Section position="6" start_page="17" end_page="19" type="metho">
    <SectionTitle>
4 Elements of Our Implementation
</SectionTitle>
    <Paragraph position="0"> In this paper, we only address verbal elementary trees. We define a verbal realization to be a combination of three classes (or &amp;quot;dimensions&amp;quot; in Candito's terminology): a subcategorization frame, a redistribution of arguments/valency alternation (in our case, voice, which we do not further discuss), and a topology, which encodes the position and characteristics of the verbal head. Thus, we reinterpret Candito's &amp;quot;Dimension 3&amp;quot; to concentrate on the position of the verbal heads, with the different argument realizations (topicalized, base position) depending on the available heads, rather than defined as first-class citizens. The subcat and argument redistributions result in a set of structures for arguments which are left- or right-branching (depending on language and grammatical function). Figure 2 shows some argument structures for German. The topology reflects the basic clause structure, that is, the distribution of arguments and adjuncts, and the position of the verb (initial, V2, final, etc.). Our notion of sentence topology is thus similar to the notion formalized by Gerdes (2002). Specifically, we see positions of arguments and adjuncts as defined by the positions of their verbal heads. However, while Gerdes (2002) assumes as basic underlying notions the fields created by the heads (the traditional Vorfeld for the topicalized element and the Mittelfeld between the verb in second position and the verb in clause-final position), we only use properties of the heads. The fields are epiphenomenal for us.As mentioned above, we use the following set of features to define our MG topology: a0 I (finite tense and subject-verb agreement): creates a specifier position for agreement which must be filled in a derivation, but allows recursion (i.e., adjunction at IP).</Paragraph>
    <Paragraph position="1"> a0 Top (topic): a feature which creates a specifier position for the topic (semantically represented in a lambda abstraction) which must be filled in a derivation, and which does not allow recursion.</Paragraph>
    <Paragraph position="2"> a0 M (mood): a feature with semantic content (to be defined), but no specifier.</Paragraph>
    <Paragraph position="3"> a0 C (complementizer): a lexical feature introduced only by complementizers.</Paragraph>
    <Paragraph position="4"> We can now define our topology in more detail.</Paragraph>
    <Paragraph position="5"> It consists of two main parts:</Paragraph>
    <Paragraph position="7"/>
    <Paragraph position="9"> can be filled in at the top feature structure to control the derivation.</Paragraph>
    <Paragraph position="10"> a0 The projection includes the origin of the verb in the phrase structure (with an empty head since we assume it is no longer there) and its maximal projection. It is shown in  the expected feature content. For example, if we want to model non-finite clauses, the maximal projection will have [a1I], while root V2 clauses will have [+Top], and embedded finite clauses with complementizers will have [+I,+C].</Paragraph>
    <Paragraph position="11"> a0 Structures for heads, which can be head-initial or head-final. They introduce categorial features. Languages differ in what sort of heads they have. Which heads are available for a given language is captured in a head inventory, i.e., a list of possible heads for that language (which use the head structure just mentioned). Two such lists are shown in Figure 4, for German and Yiddish. The corresponding head structures are shown in Figures 5 and 6.</Paragraph>
    <Paragraph position="12"> A topology is a combination of the projection and any combination of heads allowed by the language-specific head inventory. This is hard to express in XMG, so instead we list the specific combinations allowed. One might ask how we derive trees for language without the V2 phenomenon. Languages without V2 will usually have a smaller set of possible heads. We are working on a metagrammar for Korean in parallel with our work on the V2 languages. Korean is very much like German without the V2 phenomenon: the verbal head can only be in clause-final position (i.e., head 1 from Figure 5. However, passivization and scrambling can be treated the same way in Korean and German, since these phenomena are independent of V2.</Paragraph>
  </Section>
  <Section position="7" start_page="19" end_page="21" type="metho">
    <SectionTitle>
5 Sample Derivation
</SectionTitle>
    <Paragraph position="0"> Given a feature ordering (C a1 M a1 Top a1 I) and language-specific head inventories as in Figure 4, we compile out MGs for German (Figure 5) and Yiddish (Figure 6).2 The projection and the argument realizations do not differ between the two languages: thus, these parts of the MG can be reused. The features, which were introduced for descriptive reasons, now guide the TAG compilation: only certain heads can be combined. Furthermore, subjects and non-subjects are distinguished, as well as topicalized and non-topicalized NPs (producing 4 kinds of arguments so far). The compiler picks out any number of compatible elements from the Metagrammar and performs the unifications of nodes that are permitted (or required) by 2All terminal nodes are &amp;quot;red&amp;quot;; spine nodes have been annotated with their color.</Paragraph>
    <Paragraph position="1">  the node descriptions and the colors. By way of example, the derivations of elementary trees which can be used in a TAG analysis of German (2c) and Yiddish (3c) are shown in Figures 7 and 8, respectively. null</Paragraph>
  </Section>
class="xml-element"></Paper>
Download Original XML