File Information

File: 05-lr/acl_arc_1_sum/cleansed_text/xml_by_section/metho/91/m91-1021_metho.xml

Size: 16,772 bytes

Last Modified: 2025-10-06 14:12:42

<?xml version="1.0" standalone="yes"?>
<Paper uid="M91-1021">
  <Title>PRC AND THE SOVIET UNION. THE BOMBS CAUSED DAMAGE BUTNOINJURIES .&amp;quot; &amp;quot;A CAR-BOMB EXPLODED IN FRONT OF THE PRC EMBASSY, WHICH IS IN THE LIM ARESIDENTIAL DISTRICT OF SANISIDRO. MEANWHILE, TWO BOMBS WERE THROWN ATA USSR EMBASSY VEHICLE THAT WAS PARKED IN FRONT OF THE EMBASSY LOCATE D</Title>
  <Section position="3" start_page="0" end_page="137" type="metho">
    <SectionTitle>
SYSTEM ARCHITECTURE
</SectionTitle>
    <Paragraph position="0"> The PLUM architecture is presented in Figure 1 .</Paragraph>
    <Paragraph position="1"> Preprocessin g The input to the system is a file containing one or more messages. The sectioning module determines message boundaries, identifies the header, and determines paragraph and sentence boundaries. In addition, we have built a preprocessor which classifies text according to its relevance and topic . We expect this component to allow th e system to ignore paragraphs that are irrelevant and to focus on those that contain relevant information, greatl y increasing the efficiency of the overall system . Time constraints did not permit us to integrate this approach with the rest of our system, however; it was therefore not used for the MUC-3 task.</Paragraph>
    <Section position="1" start_page="0" end_page="137" type="sub_section">
      <SectionTitle>
Morphological Analysis
</SectionTitle>
      <Paragraph position="0"> The first phase of the text processing is assignment of part-of-speech information . In our current system, we use the MIT Fast Parser [4]. In the MITFP, a bi-gram probability model, frequency models for known words (derived from large corpora) and heuristics based on word endings for unknown words, assign part of speech to the highly ambiguous words of the corpus .l Since the MITFP predictions for unknown words were very inaccurate fo r input that is all upper case, we augmented this part-of-speech tagging with probabilistic models (automaticall y l We are now in the process of integrating BBN's POST probabilistic part-of-speech tagger [8] for the tagger i n MITFP.</Paragraph>
      <Paragraph position="1">  trained) for recognizing words of Spanish origin and words of English origin . This allowed us to tag new words tha t were actually Latin American names highly reliably . The Spanish classifier uses a 5 character hidden Markov model, trained on about 30,000 words of Spanish text . The five-gram model of words of English was derived from text from the Wall Street Journal .</Paragraph>
    </Section>
    <Section position="2" start_page="137" end_page="137" type="sub_section">
      <SectionTitle>
Parsing
</SectionTitle>
      <Paragraph position="0"> Each sentence identified by the sectioning module is passed to the parsing component . The MITFP is a deterministic stochastic parser which does not attempt to generate a single syntactic interpretation of the whol e sentence, rather, it generates one or more parse fragments spanning the input sentence, deferring difficult decision s on attachment ambiguities . Consequently, every sentence is assigned some (set of) syntactic interpretations , producing an average of seven fragments for sentences of the complexity seen in the MUC-3 corpus .</Paragraph>
      <Paragraph position="1"> Here are the parse fragments generated by MITFP for the second sentence of message 99 in the TST1 corpus ,</Paragraph>
    </Section>
  </Section>
  <Section position="4" start_page="137" end_page="141" type="metho">
    <SectionTitle>
&amp;quot;THE BOMBS CAUSED DAMAGE BUT NO INJURIES&amp;quot; (the full text of the message is in Appendix H) :
</SectionTitle>
    <Paragraph position="0"/>
    <Paragraph position="2"> The semantic interpreter operates on each fragment produced by MITFP in a bottom-up, compositional fashion .</Paragraph>
    <Paragraph position="3"> Throughout the system, defaults are provided so that missing semantic information or rules do not produce errors , but simply mark semantic elements or relationships as unknown. This is consistent with our belief that partial understanding has to be a key element of text processing systems, and missing data has to be regarded as a norma l event.</Paragraph>
    <Paragraph position="4"> The semantic component encompasses both lexical semantics and semantic rules. The semantic lexicon is separate from the parser's lexicon and has much less coverage . At present it contains the following numbers o f entries:  Lexical semantic entries typically include a domain model concept, as well as predicates pertaining to it. For example, here is the lexical semantics for the verb BOMB:  (defverb BOMB-V-i &amp;quot;BOMB&amp;quot; BOMBING (:case (subject PEOPLE TI-PERP-OF ) (object ANYTYPE OBJECT-OF)))  This entry indicates that the domain model concept is BOMBING, that a subject argument whose type is PEOPLE should be given the role TI-PERP-OF, and that an object argument of any type should be given the role OBJECT -OF. BOMB-V-1 is the unique identifier of this word sense .</Paragraph>
    <Paragraph position="5"> The semantic rules are based on general syntactic patterns, using wildcards and similar mechanisms to provid e an extra measure of robustness . The basic elements of our semantic representation are &amp;quot;semantic forms&amp;quot;, each o f which introduces a variable (e .g. ? 13) with a type taken from the domain model, and a collection of predicate s pertaining to that variable .</Paragraph>
    <Paragraph position="6"> There are three basic types of semantic forms : entities of the domain, events, and states of affairs . Each of these three can be further categorized as known, unknown, and referential . Entities correspond to the people, places , things, and time intervals of the domain. These are related in important ways, such as through events (who did wha t to whom) and states of affairs (properties of the entities) . Entity descriptions typically arise from noun phrases ; events and states of affairs may be described in clauses .</Paragraph>
    <Paragraph position="7"> Not everything that is represented in the semantics has actually been understood . For example, the predicate PP-MODIFIER indicates that two entities (expressed as noun phrases) are connected via a certain preposition . In this way, we have a &amp;quot;placeholder&amp;quot; for the information that a certain structural relation holds between these two items , even though we do not know what the actual semantic relation is . Sometimes understanding the relation more fully is of no consequence, since the information does not contribute to the template-filling task . The information is maintained, however, so that later expectation-driven processing can use it if necessary .  Here is a semantic rule which handles, for example, &amp;quot;group of businessmen&amp;quot;, &amp;quot;murder of a man&amp;quot;, and &amp;quot;terrorists of the FMLN&amp;quot; : For an NP dominating an NP1, and a PP whose PREP is &amp;quot;OF&amp;quot; and which dominates NP2: If NP1 is in (&amp;quot;GROUP, &amp;quot;BAND&amp;quot;) ; return semantics of NP2 If NP1 is an EVENT of type TERRORIST ; make NP2 the OBJECT-OF NP1 and return result If type of NP1 is PEOPLE and type of NP2 is ORGANIZATION, merge semantics, showing that NP ! BELONGS-TO NP2 otherwise use a more general NP =&gt; NP PP rule An important consequence of the fragmentation produced by MITFP is that top-level constituents are typically more shallow and less varied than full sentence parses. As a result, more semantics coverage was obtained early on in the development process with few semantic rules than would have been expected if the system had had to cove r widely varied syntactic structures before producing any semantic structures . In this way, semantic coverage was added gradually, while the rest of the system was progressing in parallel .</Paragraph>
    <Paragraph position="8"> Another novel aspect of our use of the MITFP was in combining its output fragments . After having assigned semantic representations to the fragments, it is often possible to make some of the attachment decisions deferred b y the MITFP. For example, it is possible to combine two NPs of compatible semantic types that are conjoined, o r attach prepositional phrases preferentially, using information automatically derived from a corpus [7] . While we lacked sufficient time to pursue this as fully as we would have liked, we did use this for certain proper nam e constructions, and anticipate using further fragment combining strategies as our semantic coverage increases . Figure 2 shows a graphical version of the semantics generated for the first fragment of sentence 1 in message 99 :  In this example note that the prepositional phrase in &amp;quot;embassies of the PRC&amp;quot; was not connected properl y semantically, as evidenced by the use of the general &amp;quot;pp-modifier&amp;quot; relation . This is because we had no case frame rule for &lt;diplomatic building&gt; of &lt;country&gt; .</Paragraph>
    <Section position="1" start_page="140" end_page="141" type="sub_section">
      <SectionTitle>
Discourse Processing
</SectionTitle>
      <Paragraph position="0"> The discourse component of PLUM performs the operations necessary to derive, from the semanti c representation of the fragments in the input message, a high level &amp;quot;discourse event structure&amp;quot;, or a representation o f the events of interest that occurred in the message . Each event in the discourse event structure is similar in principl e to the notion of a &amp;quot;frame&amp;quot;, with its corresponding &amp;quot;slots&amp;quot; or fields . There is a correspondence between a discours e event and the semantics that the semantic interpreter assigns to an event in the text . However, the semantic representation assigned by the interpreter can only include relations contained locally in a fragment (after fragmen t combination); the discourse module must infer other long-distance or indirect relations not explicitly found by th e interpreter. The template generator then uses the structures created by the discourse component to generate the fina l templates. Currently only terrorist incidents (and &amp;quot;possible terrorist incidents&amp;quot;) generate discourse events, since thes e are the core events for MUC-3 template generation . The discourse component is further discussed in the pape r &amp;quot;Computational Aspects of Discourse in the Context of MUC-3&amp;quot; in these proceedings .</Paragraph>
      <Paragraph position="1"> Two primary structures are created by the discourse processor which are used by the template generator : the discourse predicate-database and the discourse event structure. The database contains all the predicates mentioned in the semantic representation of the message (e .g., that some entity is the object of an event). It supports unification of semantic variables, so that all the information can be easily retrieved when references in the text are resolved .</Paragraph>
      <Paragraph position="2"> Any other inferences done by the discourse component also get added to the database . While only one database is produced at present, ideally there should be several, to handle multiple inference paths .</Paragraph>
      <Paragraph position="3"> To create the discourse event structure, the discourse component processes each semantic form produced by the interpreter, adding its information to the database and performing reference resolution (currently only pronouns an d proper name references) when needed . When a semantic form for an event of interest is encountered, a discourse event is generated, and any slots already found by the interpreter are filled in the event . This event is then merged with a previous event if they are compatible. This heuristic assumes that the events were derived from repeate d references to a single real event in the text .</Paragraph>
      <Paragraph position="4"> Once all the semantic forms have been processed, heuristic rules are applied to fill in any unfilled slots b y looking at text surrounding the forms which triggered a given event. Each filler found is assigned a score based o n where it was found in relation to an event trigger, indicating a higher confidence for fillers found closer to a trigger .</Paragraph>
      <Paragraph position="5"> This will not always be a valid assumption, but has proved to be a good approximation .</Paragraph>
      <Paragraph position="6"> Following is the discourse event structure created by using information in the first three sentences (spanning 2 paragraphs) of message 99:</Paragraph>
    </Section>
  </Section>
  <Section position="5" start_page="141" end_page="141" type="metho">
    <SectionTitle>
OBJECT-OF : &amp;quot;THE EMBASSIES&amp;quot; (?22, score=0)
</SectionTitle>
    <Paragraph position="0"> In the example above, a score of 0 indicates the filler was found directly by the semantics ; 4 indicates it wa s found in the same paragraph; and 6 that it was found in an adjacent paragraph . Note that El Salvador, though not i n the text, was introduced by the defmition of San Isidro in the lexicon, which had only been seen previously as a tow n of El Salvador.</Paragraph>
    <Paragraph position="1"> Template Generatio n The template generator takes the event structure produced by discourse processing and fills out the application specific templates . Clearly much of this process is governed by the specific requirements of the application , considerations which have little to do with linguistic processing . For example, in our domain model, all terroris t incidents have a result, but the MUC-3 task description states that, if the incident type is MURDER, the RESUL T slot is to be left unspecified. The template generator must incorporate these kinds of arbitrary constraints, as well a s deal with the basic details of formatting .</Paragraph>
    <Paragraph position="2"> The template generator uses a combination of data-driven and expectation-driven strategies . First the information in the event structure is used to produce initial values. At this point, values which should be filled in but are not available in the event structure are supplied from defaults, either from the header (e .g., date and location information) or from reasonable guesses (e.g. that the object of a murder is usually a suitable filler for the human target slot when the semantic type of the object is unknown) .</Paragraph>
    <Paragraph position="3"> We expect to eventually use a classifier at this stage of processing . This is especially appropriate for template slots with a set list of possible fillers, e.g. perpetrator confidence, category of incident, etc .</Paragraph>
  </Section>
  <Section position="6" start_page="141" end_page="141" type="metho">
    <SectionTitle>
EXAMPLE
</SectionTitle>
    <Paragraph position="0"> Here is the first template generated by PLUM for message 99 in the TST1 corpus :  0. MESSAGE ID TST1-MUC3-0099 1. TEMPLATE ID 1 2. DATE OF INCIDENT - 25 OCT 89 3. TYPE OF INCIDENT BOMBING 4. CATEGORY OF INCIDENT TERRORIST ACT 5. PERPETRATOR : ID OF INDIV(S) &amp;quot;TERRORISTS &amp;quot; 6. PERPETRATOR : ID OR ORG(S) 7. PERPETRATOR CONFIDENC E 8. PHYSICAL TARGET : ID(S) &amp;quot;THE EMBASSIES &amp;quot; 9. R1JMSICAL TARGET : TOTAL PLURAL 10. PHYSICAL TARGET: TYPE(S) DIPLOMAT OFFICE OR RESIDENCE : &amp;quot;THE EMBASSIES &amp;quot; 11. HUMAN TARGET: ID(S) 12. HUMAN TARGET: TOTAL NU M 13. HUMAN TARGET: TYPE(S) 14. TARGET : FOREIGN NATION S 15. INSTRUMENT: TYPE(S) * 16. LOCATION OF INCIDENT EL SALVADOR: SAN ISIDRO (TOWN ) 17. EFFECT ON PHYSICAL TARGET SOME DAMAGE : &amp;quot;THE EMBASSIES &amp;quot; 18. EFFECT ON HUMAN TARGET NO INJURY : &amp;quot;-&amp;quot;  Several things were processed correctly here: * we correctly identified the nature of the attack, the identity of the attacking individuals, and the identity and type of the target, an d * we correctly determined the nature of the damage, including the negation in &amp;quot;NO INJURIES&amp;quot; . However, several points were missed : * we failed to understand &amp;quot;TONIGHT&amp;quot;, and so filled in the default of some time before the header date; * the identity of the terrorist organization was missed because our strategy for looking for perpetrators was to o inflexible and did not keep looking once &amp;quot;TERRORISTS&amp;quot; was found ; 14 2 * our system does not yet attempt to fill the foreign target slot, so naturally we missed that filler ; and * our semantics for locations are too limited, listing only the town of San Isidro (which is in El Salvador) and not the neighborhood of San Isidro (which is in Lima, Peru). There is a reference to Lima ; the syntactic structure assigned, however, does not permit the proper semantics to identify it as a location .</Paragraph>
  </Section>
class="xml-element"></Paper>
Download Original XML