File Information

File: 05-lr/acl_arc_1_sum/cleansed_text/xml_by_section/metho/99/e99-1031_metho.xml

Size: 11,071 bytes

Last Modified: 2025-10-06 14:15:21

<?xml version="1.0" standalone="yes"?>
<Paper uid="E99-1031">
  <Title>A Flexible Architecture for Reference Resolution</Title>
  <Section position="3" start_page="229" end_page="230" type="metho">
    <SectionTitle>
2 The Architecture
</SectionTitle>
    <Paragraph position="0"/>
    <Section position="1" start_page="229" end_page="230" type="sub_section">
      <SectionTitle>
2.1 Encapsulation of layers
</SectionTitle>
      <Paragraph position="0"> Figure 1 depicts the organization of the architecture. Each of the three layers have different responsibilities: null  * Layer 1: The supervisor controls which modules in Layers 2 and 3 execute, In our implementation, the supervisor sets a run-time switch for each module in layer 2 and 3, and the first instruction of each of those modules checks its runtime flag to see if it is active for the current experiment.</Paragraph>
      <Paragraph position="1"> * Layer 2: Translation reads the input text and creates the main data structure used for reference resolution, called the discourse context (DC). The DC consists of discourse entities (DEs) introduced in the text, some of which are anaphoric. This layer contains all syntactic and semantic analysis components and all interaction with the surrounding system, such as access to a gender database or a lexicon for semantic restrictions. All features that need to be available to reference resolution are posted to the DC. This layer is also responsible for deciding which input constituents create DEs.</Paragraph>
      <Paragraph position="2"> * Layer 3: Anaphora resolution contains a variety of functions for resolving different types of anaphora. Responsibilities of this layer include determining what anaphoric phenomena are to be resolved in the current experiment, determining what anaphora resolution technique(s) will be used, and determining what updates to make to the DC.</Paragraph>
      <Paragraph position="3"> Even though the modules are independent of the input format, they are still somewhat dependent on the availability of DE features.</Paragraph>
      <Paragraph position="4"> If a feature needed by a particular resolution module was not created in a particular experiment, the module must either do without it or give up and exit. This layer's output is an updated DC with anaphoric elements re- null solved to their referents. If labeled training data is available, this layer is also responsible for calculating the accuracy of anaphora resolution.</Paragraph>
    </Section>
    <Section position="2" start_page="230" end_page="230" type="sub_section">
      <SectionTitle>
2.2 Benefits of this design
</SectionTitle>
      <Paragraph position="0"> This strict delineation of responsibilities between layers provides the following advantages:  * Once a translation layer is written for a specific type of input, all the implemented anaphora resolution techniques are immediately available and can be compared. * Different models of DC construction can be compared using the same underlying reference resolution modules.</Paragraph>
      <Paragraph position="1"> * It is simple to activate or deactivate each component of the system for a particular experiment. null</Paragraph>
    </Section>
  </Section>
  <Section position="4" start_page="230" end_page="231" type="metho">
    <SectionTitle>
3 Implementation
</SectionTitle>
    <Paragraph position="0"> We used this architecture to implement a testing platform for pronoun resolution. Several experiments were run to demonstrate the flexibility of the architecture. The purpose of this paper is not to compare the pronoun resolution results for the techniques we implemented, so pronoun resolution accuracy of particular techniques will not be discussed here.l Instead, our implementation is described to provide some examples of how the architecture can be put to use.</Paragraph>
    <Section position="1" start_page="230" end_page="230" type="sub_section">
      <SectionTitle>
3.1 Supervisor layer
</SectionTitle>
      <Paragraph position="0"> The supervisor layer controls which modules within layers 2 and 3 execute for a particular experiment. We created two different supervisor t See (Byron and Allen. 1999; Tetreault, 1999) for results of pronoun resolution experiments run within the testbed.  modules in the testbed. One of them simply reads a configuration file with runtime flags hard-coded by the user. This allows the user to explicitly control which parts of the system execute, and will be used when a final reference resolution techniques is chosen for integration into the TRIPS system parser (Ferguson and Allen, 1998).</Paragraph>
      <Paragraph position="1"> The second supervisor layer was coded as a genetic algorithm (Byron and Allen, 1999). In this module, the selection of translation layer modules to execute was hard-coded for the evaluation corpus, but pronoun resolution modules and methods for combining their results were activated and de-activated by the genetic algorithm. Using pronoun resolution accuracy as the fitness function, the algorithm learned an optimal combination of pronoun resolution modules.</Paragraph>
    </Section>
    <Section position="2" start_page="230" end_page="230" type="sub_section">
      <SectionTitle>
3.2 Translation layer
</SectionTitle>
      <Paragraph position="0"> Translation layer modules are responsible for all syntactic and semantic analysis of the input text.</Paragraph>
      <Paragraph position="1"> There are a number of design features that must be controlled in this layer, such as how the discourse structure affects antecedent accessibility and which surface constituents trigger DEs. All these design decisions should be implemented as independent modules so that they can be turned on or off for particular experiments.</Paragraph>
      <Paragraph position="2"> Our experiments created translation modules for two evaluation corpora: written news stories from the Penn Treebank corpus (Marcus et al., 1993) and spoken task-oriented dialogues from the TRAINS93 corpus (Heeman and Allen, 1995). The input format and features added onto DEs from these two corpora are very different, but by encapsulating the translation layer, the same pronoun resolution code can be used for both domains. In both of our experiments only simple noun phrases in the surface form triggered DEs.</Paragraph>
      <Paragraph position="3"> Treebank texts contain complete structural parsers, POS tags, and annotation of the antecedents of definite pronouns (added by Ge et al. 1998). Because of the thorough syntactic information, DEs can be attributed with explicit phrase structure information. This corpus contains unconstrained news stories, so semantic type information is not available. The Treebank translator module adds the following features to each  1.</Paragraph>
      <Paragraph position="4"> DE: Whether its surface constituent is contained in reported speech; 2. A list of parent nodes containing its surface constituent in the parse tree. Each node's unique identifier encodes the phrase type (i.e. VB, NP, ADJP); 3. Whether the surface constituent is in the second half of a compound sentence; 4. The referent's animacy and gender from a  hand-coded agreement-feature database.</Paragraph>
      <Paragraph position="5"> A second translation module was created for a selection of TRAINS93 dialogue transcripts. The input was POS-tagged words with no structural analysis. Other information, such as basic punctuation and whether each pronoun was in a main or subordinate clause, had previously been hand-annotated onto the transcripts. We also created an interface to the semantic type hierarchy within the Trains system and added semantic information to the DEs.</Paragraph>
      <Paragraph position="6"> Common DE attributes for both corpora:  I. Plural or singular numeric agreement; 2. Whether&amp;quot; the entity is contained in the subject of the matrix clause; 3. Linear position of the surface constituent; 4. Whether its surface constituent is definite or indefinite; 5. Whether its surface constituent is contained in quoted speech; 6. For pronoun DEs, the id of the correct antecedent (used for evaluation).</Paragraph>
    </Section>
    <Section position="3" start_page="230" end_page="231" type="sub_section">
      <SectionTitle>
3.3 Anaphora resolution layer
</SectionTitle>
      <Paragraph position="0"> Modules within this layer can be coded to resolve a variety of anaphoric phenomena in a variety of ways. For example, a particular experiment may be concerned only with resolving pronouns or it might also require determination of coreference between definite noun phrases. This layer is reminiscent of the independent anaphora resolution modules in the Lucy system (Rich and LuperFoy, 1988), except that modules in that system were not designed to be easily turned on or off.</Paragraph>
      <Paragraph position="1"> For our testbed, we implemented a variety of pronoun resolution techniques. Each technique  Boost salience for the first entity in each sentence Decrease salience for entities in prepositional phrases or relative clauses Increase the salience for non-subject entities for demonstrative pronoun resolution (Schiffman, 1985) Decrease salience for indefinite entities Decrease salience for entities in reported speech Increase the salience of entities in the subject of the previous sentence Increase the salience of entities whose surface form is pronominal  can run in isolation or with the addition of meta-modules that combine the output of multiple techniques. We implemented meta-modules to interface to the genetic algorithm driver and to combine different salience factors into an over-all score (similar to (Carbonell and Brown, 1988; Mitkov, 1998)). Table 1 describes the pronoun resolution techniques implemented at this point, and shows whether they are activated for the Treebank and the TRAINS93 experiments. Although each module could run for both experiments without error, if the features a particular module uses in the DE were not available, we simply de-activated the module. When we migrate the TRIPS system to a new domain this year, all these pronoun resolution methods will be available for comparison.</Paragraph>
    </Section>
  </Section>
  <Section position="5" start_page="231" end_page="231" type="metho">
    <SectionTitle>
4 Summary
</SectionTitle>
    <Paragraph position="0"> This paper has described a framework for reference resolution that separates details of the syntactic/semantic interpretation process from anaphora resolution in a plug-and-play architecture. The approach is not revolutionary, it simply demonstrates how to apply known software engineering techniques to the reference resolution component of a natural language understanding system. The framework enables comparison of baseline techniques across corpora and allows for easy modification of an implemented system when the sources of information available to anaphora resolution change. The architecture facilitates experimentation on different mixtures of discourse context and anaphora resolution algorithms. Modules written within this framework are portable across domains and language genres. null</Paragraph>
  </Section>
class="xml-element"></Paper>
Download Original XML