File Information
File: 05-lr/acl_arc_1_sum/cleansed_text/xml_by_section/metho/89/j89-3002_metho.xml
Size: 107,161 bytes
Last Modified: 2025-10-06 14:12:16
<?xml version="1.0" standalone="yes"?> <Paper uid="J89-3002"> <Title>KNOWLEDGE REPRESENTATION FOR COMMONSENSE REASONING WITH TEXT</Title> <Section position="3" start_page="0" end_page="0" type="metho"> <SectionTitle> 1.2 THE KT SYSTEM </SectionTitle> <Paragraph position="0"> Naive semantics is the theoretical motivation for the KT system under development at the IBM Los Angeles Scientiific Center by Dahlgren, McDowell, and others. 3 The heart of the system is a commonsense knowledge base with two components, a commonsense ontology and databases of generic knowledge associated with lexical items. The first phase of the project, which is nearly complete, is a text understanding and query system. In this phase, text is read into the system and parsed by the MODL parser (McCord 1987), which has very wide coverage. The parse is submitted to a module (DISAMBIG) that outputs a logical structure which reflects the scope properties of operators and quantifiers, correct attachment of post-verbal adjuncts, and selects, word senses. This is passed to a semantic translator whose output (a DRS) is then converted to first-order logic (FOL). We then have the text in two different semantic forms (DRS and FOL), each of which has its advantages and each of which is utilized in the system in different ways. Queries are handled in the same way as text. Answers to the queries are obtained either by matching to the FOL textual database or to the commonsense databases. However, the commonsense knowledge is accessed at many other stages in the processing of text and queries, namely in parse disambiguafion, in lexical retrieval, anaphora resolution, and in the construction of the discourse structure of the entire text.</Paragraph> <Paragraph position="1"> 150 Computational Linguistics, Volume 15, Number 3, September 1989 Kathleen Dahlgren, Joyce McDowell, and Edward P. Stabler, Jr. Knowledge Representation for Commonsense Reasoning with Text The second phase of the project, which is at present in the research stage, will be to use the commonsense knowledge representations and the textual database to guide text selection. We anticipate a system, NewSelector, which, given a set of user profiles, will distribute textual material in accordance with user interests, thus, in effect, acting as an automatic clipping service. Our target text is the Wall Street Journal. The inferencing capabilities provided by the commonsense knowledge will allow us to go well beyond simple keyword search.</Paragraph> <Paragraph position="2"> The theoretical underpinnings and practical work on the KT system have been reported extensively elsewhere, in conference papers (Dahlgren and McDowell 1986a, 1986b; McDowell and Dahlgren 1987) and in a book (Dahlgren 1988a). Since the publication of those works, a number of significant additions and modifications have been made to the system. The intended focus of this paper is this new work. However, in order to make this accessible to readers unfamiliar with our previous reports, we present in Section 2 an overview of the components of the present system. (Readers familiar with our system can skip Section 2). In the remaining sections on new work we have emphasized implementation, because this paper is addressed to the computational linguistics community: in Section 3, the details of disambiguation procedures that use the NS representations and in Section 4 the details of the query system. Finally, Section 5 discusses work in progress regarding discourse and naive semantics.</Paragraph> </Section> <Section position="4" start_page="0" end_page="0" type="metho"> <SectionTitle> 2 OVERVIEW or THE KT SYSTEM 2.1 KNOWLEDGE REPRESENTATION 2.1.1 THE ONTOLOGY </SectionTitle> <Paragraph position="0"> Naive theories associated with words include beliefs concerning the structure of the actual world and the significant &quot;joints&quot; in that structure. People have the environment classified, and the classification scheme of a culture is reflected in its language. Since naive semantics is intended as a cognitive model, we constructed the naive semantic ontology empirically, rather than intuitix;ely. We studied the behavior of hundreds of verbs and determined selectional restrictions, which are constraints reflecting the naive ontology embodied in English. We also took into account psychological studies of classification (Keil 1979), and philosophical studies of epistemology (Strawson 1953).</Paragraph> </Section> <Section position="5" start_page="0" end_page="0" type="metho"> <SectionTitle> 2.1.1.1 MATHEMATICAL PROPERTIES OF THE ONTOLOGY </SectionTitle> <Paragraph position="0"> The ontology has several properties which distinguish it from classical taxonomies. It is a directed acyclic graph, rather than a binary tree, because many concepts have more than two subordinate concepts (Rosch et al. 1976).</Paragraph> <Paragraph position="1"> FISH, BIRD, MAMMAL, and so on, are subordinates of VERTEBRATE. It is a directed graph rather than a tree, because it handles cross-classification. Cross-classification is justified by contrasts between individual and collective nouns such as &quot;cow&quot; and &quot;herd.&quot; This</Paragraph> </Section> <Section position="6" start_page="0" end_page="0" type="metho"> <SectionTitle> ENTITY ABSTRACT NUMERICAL REAL PHYSICAL --* NON-STATIONARY --* COLLECTIVE --* TEMPORAL ---> RELATIONAL --~ EVENT --* </SectionTitle> <Paragraph position="0"/> <Section position="1" start_page="0" end_page="0" type="sub_section"> <SectionTitle> The Ontological Schema </SectionTitle> <Paragraph position="0"> implies that cognitively there are essentially parallel ontological schemas for individuals and collectives.</Paragraph> <Paragraph position="1"> Thus we have the parallel ontology fragments in Figure 2. Table 2 illustrates cross-classification at the root of the ontology, where ENTITY cross-classifies as either REAL or ABSTRACT, and as either INDIVIDUAL or COLLECTIVE. Cross-classification is handled as in McCord (1985, 1987).</Paragraph> </Section> </Section> <Section position="7" start_page="0" end_page="0" type="metho"> <SectionTitle> INDIVIDUAL COLLECTIVE ENTITY ENTITY ABSTRACT REAL ABSTRACT REAL / / P~SICAL, N~U RAI_, SE ~M~ING PH~IC~,~TURAL,SE~VING / / ANIMAL FAUNA / / COW HERD </SectionTitle> <Paragraph position="0"/> </Section> <Section position="8" start_page="0" end_page="0" type="metho"> <SectionTitle> INDIVIDUAL COLLECTIVE REAL cow herd ABSTRACT idea book </SectionTitle> <Paragraph position="0"> Multiple attachments of instantiations to leaves is possible. For example, an entity, John, is both a HUMAN with the physical properties of a MAMMAL, and is also a PERSON, a SENTIENT. As a SENTIENT, John can be the subject of mental verbs such as think and say. Institutions are also SENTIENTs, so that the SENTIENT node reflects English usage in pairs like (4).</Paragraph> <Paragraph position="1"> 4. John sued Levine. The government sued Levine.</Paragraph> <Paragraph position="2"> Computational Linguistics, Volume 15, Number 3, September 1989 151 Kathleen Dahlgren, Joyce McDowell, and Edward P. Stabler, Jr. Knowledge Representation for Commonsense Reasoning with Text On the other hand, John, as a HUMAN, is like animals, and has physical properties. Both John and a cow can figure as subjects of verbs like &quot;eat&quot; and &quot;weigh.&quot; Multiple attachment was justified by an examination of the texts. It was found that references to human beings in text, for example, deal with them either as persons (SENTIENTs) or as ANIMALs (physiological beings), but rarely as both at the same time.</Paragraph> </Section> <Section position="9" start_page="0" end_page="0" type="metho"> <SectionTitle> 2.1.1.2 ONTOLOGICAL CATEGORIES </SectionTitle> <Paragraph position="0"> The INDIVIDUAL/COLLECTIVE cut was made at the level of ENTITY (the highest level) because all types of entities are conceived individually or in collections. COLLECTIVE breaks into sets of identical members (herd, mob, crowd, fleet), masses that are conceived as stuff (sand, water), and structures where the members have specified relations, such as in institutions (school, company, village). Leaf node names, such as ANIMAL and FAUNA, are shorthand for collections of categories inherited from dominating nodes and by cross-classification. Thus &quot;cow&quot; and &quot;herd&quot; share all categories except that &quot;cow&quot; is an INDIVIDUAL term and &quot;herd&quot; is a COLLECTIVE term.</Paragraph> <Paragraph position="1"> The REAL node breaks into the categories PHYSI-CAL, TEMPORAL, and SENTIENT, and also NATURAL and SOCIAL. Entities (or events) that come into being (or take place) naturally must be distinguished from those that arise through some sort of social intervention. Table 3 illustrates the assignment of example words under the REAL cross-classification.</Paragraph> </Section> <Section position="10" start_page="0" end_page="0" type="metho"> <SectionTitle> INDIVIDUAL COLLECTIVE NATURAL SOCIAL NATURAL SOCIAL </SectionTitle> <Paragraph position="0"> PHYSICAL rock knife sand fleet SENTIENT man programmer mob clinic TEMPORAL earthquake party winter epoch Table 3. Attachment of Nouns under REAL The SENTIENT/PHYSICAL distinction is placed high because in commonsense reasoning, the properties of people and things are very different. Verbs select for SENTIENT arguments or PHYSICAL arguments, as illustrated in (5). Notice also, that as a physical object, an individual entity like John can be the subject both of verbs that require physical subjects and those that John sued Levine.</Paragraph> <Paragraph position="1"> The cow fell.</Paragraph> <Paragraph position="2"> We make the SENTIENT/NON-SENTIENT distinction high up in the hierarchy for several reasons. Philosophically, the most fundamental distinction in epistemology (human knowledge) is arguably that between thinking and non-thinking beings (Strawson 1953). Psychology has shown that infants are able to distinguish humans from all other objects and they develop a deeper and more complex understanding of humans than of other objects (Gelman and Spelke 1981). In the realm of linguistics, a class of verbs selects for persons, roles, and institutions as subjects or objects. Thus the SENTIENT distinction captures the similarity between persons and institutions or roles. There is a widespread lexical ambiguity between a locational and institutional reading of nouns, which can be accounted for by the SENTIENT distinction, as in (7).</Paragraph> <Paragraph position="3"> 7. The court is in the center of town.</Paragraph> <Paragraph position="4"> The court issued an injunction.</Paragraph> <Paragraph position="5"> The NATURAL/SOCIAL distinction also was placed high in the hierarchy. Entities (including events) that are products of society, and thereby have a social function, are viewed as fundamentally different from natural entities in the commonsense conceptual scheme. The distinction is a basic one psychologically (Miller 1978; Gelman and Spelke 1981). SOCIAL entities are those that come into being only in a social or institutional setting, with &quot;institution&quot; being understood in the broadest sense, for instance family, government, education,, warfare, organized religion, etc.</Paragraph> </Section> <Section position="11" start_page="0" end_page="0" type="metho"> <SectionTitle> 2.1.1.3 CONSTRUCTION OF THE ONTOLOGY </SectionTitle> <Paragraph position="0"> The ontological schema was constructed to handle the selectional restrictions of verbs in 4,000 words of geography text and 6,000 words of newspaper text. These were arranged in a hierarchical schema. The hierarchy was examined and modified to reflect cognitive, philosophical, and linguistic facts, as described above. It was pruned to make it as compact as possible. We minimized empty terminal nodes. A node could not be part of the ontology unless it systematically pervaded some subhierarchy. Distinctions found in various places were relegated to feature status. The full ontology with examples may be found in Dahlgren (1988).</Paragraph> </Section> <Section position="12" start_page="0" end_page="0" type="metho"> <SectionTitle> 2.1.1.4 VERBS </SectionTitle> <Paragraph position="0"> In KT, verbs are attached to the main ontology at the node TEMPORAL because information concerning the temporality of situations described in a sentence is encoded on the verb as tense and because the relations indicated by verbs must be interpreted with respect to their location in time in order to properly understand the discourse structure of a text. Thus the ontology implies that events are real entities, and that linguistic, not conceptual, structure distinguishes verbalized from nominalized versions of events and states.</Paragraph> <Paragraph position="1"> We view the cognitive structure of the concepts associated with nouns and verbs as essentially different.</Paragraph> <Paragraph position="2"> Non-derived nouns in utterances refer to objects. Lexical nouns name classes of entities that share certain features. Verbs name classes of events and states, but 152 Computational Linguistics, Volume 15, Number 3, September 1989 Kath|een Dahlgren, Joyce McDoweli, and Edward P. Stabler, Jr. Knowledge Representation for Commonsense Reasoning with Text these do not share featural descriptions. Psychologically, verbs are organized around goal orientation and argument types (Huttenlocher and Lui 1979; Graesser and Hopkinson 1987).</Paragraph> <Paragraph position="3"> The primary category cut at the node TEMPORAL is between nouns, which name classes of entities, in this case temporal entities like &quot;party,&quot; &quot;hurricane,&quot; and &quot;winter,&quot; and verbs, which indicate relations between members of these nominal classes, like &quot;hit,&quot; &quot;love,&quot; &quot;remember.&quot; We attach temporal nouns to the TEM-PORAL/NON-RELATIONAL node and verbs to the TEMPORAL/RELATIONAL node. Many nouns are, of course, relational, in the sense that &quot;father&quot; and &quot;indictment&quot; are relational. Our node RELATIONAL does not carry this intuitive sense of relational, but instead simply indicates that words attached here require arguments for complete interpretation. So, while &quot;father&quot; is relational, it is possible to use &quot;father&quot; in a text without mentioning the related entity. But verbs require arguments (usually overtly, but sometimes understood, as in the case of commands) for full interpretation. Nominalizations like &quot;indictment&quot; are a special case. In our system, all deverbal nominalizations are so marked with a pointer to the verb from which they are derived. Subsequent processing is then directed to the verb, which, of course, is attached under TEMPORAL,</Paragraph> </Section> <Section position="13" start_page="0" end_page="0" type="metho"> <SectionTitle> RELATIONAL. 2.1.1.5 THE VENDLER CLASSIFICATION </SectionTitle> <Paragraph position="0"> One basis of the relational ontology is the Vendler (1967) classification scheme, which categorizes verbs into aspectual classes (see Dowty 1979). According to this classification, RELATIONAL divides into EVENT or STATIVE, and EVENT divides into ACTIVITY, ACHIEVEMENT, or ACCOMPLISHMENT. Vendler and others (particularly Dowty 1979) have found the following properties, which distinguish these classes.</Paragraph> <Paragraph position="1"> 8. STATIVE and ACHIEVEMENT verbs may not appear in the progressive, but may appear in the simple present.</Paragraph> <Paragraph position="2"> ACTIVITY and ACCOMPLISHMENT verbs may appear in the progressive and if they appear in the simple present, they are interpreted as describing habitual or characteristic states.</Paragraph> <Paragraph position="3"> ACCOMPLISHMENTs and ACHIEVEMENTs entail a change of state associated with a terminus (a clear endpoint). STATIVEs (&quot;know&quot;) and ACTIVITYs (&quot;run&quot;) have no well-defined terminus. ACHIEVEMENTs are punctual (John killed Mary) while ACCOMPLISHMENTs are gradual (John built a house).</Paragraph> <Paragraph position="4"> STATEs and ACTIVITYs have the subinterval property (cf. Bennett and Partee 1978).</Paragraph> <Paragraph position="5"> Table 4 summarizes these distinctions. There are several standard tests for the Vendler (1967) system which can be found in Dowty (1979) and others and which we apply in classification.</Paragraph> <Paragraph position="6"> The Vendler classification scheme is actually more accurately a classification of verb phrases than verbs. KT handles this problem in two ways. First, in sentence processing we take into account the arguments in the verb phrase as well as the verb classification to determine clause aspect, which can be any of the Vendler classes. Second, we classify each sense of a verb separately.</Paragraph> <Paragraph position="7"> The other nodes in the relational ontology are motivated by the psycholinguistic studies noted above. The MEN-TAL/NONMENTAL/EMOTIONAL distinction is made at the highest level for the same reasons that led us to place SENTIENT at a high level in the main ontology. All EVENTs are also cross-classified as GOAL oriented or not. This is supported by virtually every experimental study on the way people view situations, i.e., GOAL orientation is the most salient property of events and actions. For example, Trabasso and Van den Broek (1985) find that events are best recalled which feature the goals of individuals and the consequences of goals and Trabasso and Sperry (1985) find that the salient features of events are goals, antecedents, consequences, implications, enablement, causality, motivation, and temporal succession and coexistence. This view is further supported by Abbott, Black, and Smith (1985) and Graesser and Clark (1985). NONGOAL, ACCOMPLISHMENT is a null category because AC-COMPLISHMENTS are associated with a terminus and thus inherently GOAL oriented. On the other hand NONGOAL, ACHIEVEMENT is not a null category because the activity leading up to an achievement is always totally distinct from the achievement itself.</Paragraph> <Paragraph position="8"> SOCIAL, NONGOAL, ACTIVITY is a sparse category. null Cross-classifications inherited from the TEMPORAL node are SOCIAL/NATURAL and INDIVIDUAL/ COLLECTIVE. The INDIVIDUAL/COLLECTIVE distinction is problematical for verbs, because all events can be viewed as a collection of an infinitude of subevents. null</Paragraph> </Section> <Section position="14" start_page="0" end_page="154" type="metho"> <SectionTitle> 2.1.2 GENERIC KNOWLEDGE </SectionTitle> <Paragraph position="0"> Generic descriptions of the nouns and verbs were drawn from psycholinguistic data to the extent possible. In a typical experiment, subjects are asked to freelist features &quot;characteristic&quot; of and common to objects in Computational Linguistics, Volume 15, Number 3, September 1989 153 Kathleen Dahlgren, Joyce McDoweH, and Edward P. Stabler, Jr. Knowledge Representation for Commonsense Reasoning with Text categories such as DOG, LEMON, and SECRETARY (Rosch et al. 1976; Ashcraft 1976; Dahlgren 1985a). The number of subjects in such an experiment ranges from 20 to 75. Any feature that is produced in a freelisting experiment by several subjects is likely to be shared in the relevant subpopulation (Rosch 1975; Dahlgren 1985a). Features that were freelisted by at least one-fifth of the subjects are chosen for a second experiment, in which different subject s are asked to rate the features for typicality. Those features rated as highly typical by the second group can be considered a good first approximation to the content of the cognitive structures associated with the terms under consideration. The number of features shared in this way for a term averaged 15.</Paragraph> <Paragraph position="1"> The generic knowledge in the KT system is contained in two generic data bases, one for nouns and one for verbs. There is a separate entry for each sense of each lexical item. The content of the entries is a pair of lists of features drawn from the psycholinguistic data (as described above) or constructed using these data as a model. A feature is, informally, any bit of knowledge that is associated with a term. In informal terms, these can be any items like &quot;wears glasses&quot; (programmer), &quot;is red&quot; (brick), or &quot;can't be trusted&quot; (used-car salesman). For each entry, the features are divided into two lists, one for typical features and one for inherent features. For example, a brick is typically red, but blood is inherently red. Together the lists comprise the entry description of the term that heads the entry.</Paragraph> <Paragraph position="2"> The source of descriptions of social roles were data collected by Dahlgren (1985a). For physical objects we used generic descriptions from Ashcraft (1976), including raw data generously supplied by the author. An informal conceptualization for &quot;lawyer&quot; was shown in (1). The corresponding generic description is shown in (9). Features of the same feature type within either the inherent or typical list are AND'ed or OR'ed as required. Some features contain first-order formulas like the conditions in discourse representations. For example, one function feature has (advise(E,noun,Y) & client(Y) & regarding(E,Z) & law(Z)). The first argument of the predicate advise is an event reference marker. This event is modified in the regarding predicate. The second argument of advise is instantiated in the processing as the same entity that is predicated as being in the extension of the noun generically described in the representation, in this case, &quot;lawyer.&quot;</Paragraph> <Paragraph position="4"/> <Paragraph position="6"> The entire set of features for nouns collected in this way so far sort into 38 feature types. These are age, agent, appearance, association, behavior, color, construction, content, direction, duration, education, exemplar, experienced-as, in-extension-of, frequency, function, goal, habitat, haspart, hasrole, hierarchy, internal-trait, legal-requirement, length, level, location, manner, material, name, object, odor, operation, owner, partof, physiology, place, processing, propagation, prototype, relation, requirement, rolein, roles, sex, shape, size, source, speed, state, status, strength, structure, taste, texture, time. A given feature type can be used in either typical or inherent feature lists. Since these are not primitives, we expect the list of feature types to expand as we enlarge the semantic domain of the system.</Paragraph> <Paragraph position="7"> There is a much smaller set of feature types for verbs. We were guided by recent findings in the psycholinguistic literature which show that the types of information that subjects associate with verbs are substantially different from what they associate with nouns. Huttenlocher and Lui (1979) and Abbott, Black, and Smith (1985) in particular have convincingly argued that subjects conceive of verbs in terms of whether or not the activities they describe are goal oriented, the causal and temporal properties of the events described, and the types of entities that can participate as arguments to the verb. For the actual feature types, we adapted the findings of Graesser and Clark (1985), whose research focused on the salient implications of events in narratives. A small number of feature types is sufficient to represent the most salient features of events. These can be thought of as answers to questions about the typical event described by the verb that heads the entry. In addition, selectional restrictions on the verb are also encoded as a feature type. Feature types for verbs are cause, goal, what.enabled, what.happened_next, consequence_oLevent, where, when, implies, how, selectioD..\]-restriction. An example of a generic entry for verbs follows.</Paragraph> <Paragraph position="9"> if someone buys something, typlcaPSly he can afford it, he uses money, he buys it in a store, he needs it, and later he uses it.</Paragraph> <Paragraph position="10"> Inherently, his goal is to own it, and after buying it, he does OwTt it.</Paragraph> <Paragraph position="11"> The buyer is sentient and what is bought is merchandise.</Paragraph> <Paragraph position="12"> Computational Linguistics, Volume 15, Number 3, September 1989 Kathleen Dahlgren, Joyce McDowell, and Edward P. Stabler, Jr. Knowledge Representation for Commonsense Reasoning with Text Naive Semantic representations of generic knowledge contain fifteen or more pieces of information per word, relatively more than required by other theories. The magnitude of the lexicon is counterbalanced by constraints that naturally obtain in the generic knowledge. Study of the protocols of subjects in prototype experiments reveals that people conceive of objects in constrained patterns of feature types. For example, animals are conceived in terms of physical and behavioral properties, while social roles are conceived in terms of functional and relational properties. Thus not all feature types occur in the representations of all nouns (or verbs). The pattern of features relevant to each node in the ontology is called a Kind Type. Each feature is classified by type as a COLOR, SIZE, FUNCTION, INTERNAL TRAIT or other. At each node, only certain feature types are applicable. Features at lower nodes inherit feature type constraints from the nodes above them in the ontology. For instance, any node under SOCIAL may have certain feature types, and any node under ROLE may have those feature types inherited from SOCIAL, as well as further feature types. Examples contrasting &quot;elephant&quot; and &quot;lawyer&quot; are shown in Tables 5 and 6.</Paragraph> <Paragraph position="13"> A lexical augmentation facility is used to create generic entries. This facility exploits the fact that possible feature types for any term are constrained by the ontological attachment of the term, by the Kind Type to which they belong (Dahlgren & McDowell 1986a; Dahlgren 1988a). For example, it is appropriate to encode the feature type &quot;behavior&quot; for &quot;dog&quot; but not for &quot;truck.&quot; Similarly, it is appropriate to encode the feature type &quot;goal&quot; for &quot;dig&quot; but not for &quot;fall.&quot; The lexical augmentation facility presents the user with appropriate choices for each term and then converts the entries to a form suitable for processing in the system.</Paragraph> <Section position="1" start_page="154" end_page="154" type="sub_section"> <SectionTitle> 2.2 TEXT INTERPRETATION ARCHITECTURE </SectionTitle> <Paragraph position="0"> The overall goal of the KT project is text selection based on the extraction of discourse relations guided by Node in Feature types associated Feature values for Ontology with the node lawyer SOCIAL function types function practice law function argue cases function advise clients function represent clients in court naive semantic representations. This goal motivated the choice of DRT as the compositional semantic formalism for the project. The particular implementation of DRT which we use assumes a simple, purely syntactic parse as input to the DRS construction procedures (Wada and Asher 1986). Purely syntactic parsing and formal semantics are unequal to the task of selecting one of the many possible parses and interpretations of a given text, but human readers easily choose just one interpretation. NS representations are very useful in guiding this choice. They can be used to disambiguate parse trees, word senses, and quantifier scope. We use an existing parser (MODL) to get one of the syntactic structures for a sentence and modify it based upon NS information.</Paragraph> <Paragraph position="1"> This allows us to isolate the power of NS representations with respect to the parsing problem. Not only are NS representations necessary for a robust parsing capability, but also in anaphora resolution and discourse reasoning. Furthermore, the parse tree must be available to the discourse coherence rules. Thus our research has shown that not only must the NS representations be accessible at all levels of text processing, but purely syntactic, semantic, and pragmatic information that has been accumulated must also be available to later stages of processing. As a result, the architecture of the system involves separate modules for syntax, semantics, discourse and naive semantics, but each of the modules has access to the output of all others, as shown in Figure I.</Paragraph> <Paragraph position="2"> The parser chosen is the Modular Logic Grammar (McCord 1987), (MODL). Both MODL and KT are written in VM/PROLOG (IBM 1985). The input to MODL is a sentence or group of sentences (a text). In KT we intercept the output of MODL at the stage of a labeled bracketing marked with grammatical features and before any disambiguation or semantic processing is done. In effect, we bypass the semantics of MODL in Computational Linguistics, Volume 15, Number 3, September 1989 155 Kathleen Dahlgren, Joyce McDowell, and Edward P. Stabler, Jr. Knowledge Representation for Commonsense Reasoning with Text order to test our NS representations. In the labeled bracketing each lexical item is associated with an argument structure that can be exploited in semantic interpretation. The labeled bracketing output by MODL is slightly processed before being passed to our module DISAMBIG. Here the commonsense knowledge base is accessed to apply rules for prepositional phrase attachment (Dahlgren and McDowell 1986b) and word sense disambiguation (Dahlgren 1988a), as well as to assign the correct scope properties to operators and quantifiers. The output is a modified parse. All of these modules are in place and functional. The word sense disambiguation rules are in the process of being converted. An example of the input to DISAMBIG and the resulting output is as follows: 11. Input S: John put the money in the bank.</Paragraph> <Paragraph position="3"> Input to DISAMBIG:</Paragraph> <Paragraph position="5"> The differences are that the PP &quot;in the bank&quot; is VP-attached in the output of DISAMBIG rather than NP-attached as in the output of MODL, and that the words &quot;put&quot; and &quot;bank&quot; are assigned index numbers and changed to putl and bank2, selecting the senses indicated by the word sense disambiguation algorithm.</Paragraph> <Paragraph position="6"> The modified parse is then submitted to a semantics module, which outputs a structure motivated in part by current versions of discourse representation theory (DRT) (Kamp 1981; Wada and Asher 1986; Asher and Wada 1988). The actual form of the discourse representation structure (DRS) and its conditions list in the KT system differ from standard formats in DRT in that tense arguments have been added to every predicate and tense predicates link the tense arguments to the tense of the verb of the containing clause. The analysis of questions and negation was carried out entirely with respect to the KT system and to serve its needs. The DRT semantics is in place and functional for most structures covered by MODL. The commonsense knowledge representations are accessed in the DRT module for semantic interpretation of modals, the determination of sentence-internal pronoun anaphora (where simple C-command and agreement tests fail), and to determine some cases of quantifier scoping.</Paragraph> <Paragraph position="7"> As each sentence of a text is processed it is added to the DRS built for the previous sentence or sentences. Thus an augmented DRS is built for the entire text. In the augmentation module the commonsense knowledge representations are accessed to determine definite noun phrase anaphora, sentence-external pronoun anaphora, temporal relations between the tense predicates generated during sentence-level DRS construction, discourse relations (suc\]h as causal relations) between clauses, and the rhetorical structure of the discourse as a whole. The discourse work is being carried out mainly by Dahlgren and is in various stages of completion.</Paragraph> <Paragraph position="8"> Since standard proof techniques are available for use with logical forms, the DRS formulated by the sentence-level and discourse-level semantic components is converted to standard logic. A number of difficulties present themselves here. In the first place, given any of the proposed semantics for DRSs (e.g., Kamp 1981; Asher 1987), DRSs are not truth functional. That is, the truth walue of a DRS structure (in the actual world) is not generally a function of the truth values of its constituents. For example, this happens when verbs produce opaque contexts (Asher 1987). Since general proof methods for modal logics are computationally difficult, we have adopted the policy of mapping DRSs to naiw ~ , first-order translations in two steps, providing a special and incomplete treatment of non-truth functional contexts. The first step produces representations that differ minimally from standard sentences of first-order logic. The availability of this level of representation enhances the modularity and the extensibility of the system. Since first-order reasoning is not feasible in an application like this, a second step converts the logical forms to clausal forms appropriate for the problem solver or the textual knowledge base. We describe each of these steps in turn.</Paragraph> </Section> </Section> <Section position="15" start_page="154" end_page="154" type="metho"> <SectionTitle> 2.2.4.1 THE TRANSLATION TO STANDARD LOGICAL FORMS </SectionTitle> <Paragraph position="0"> The sentence-level and discourse-level semantic components disambiguate names and definite NPs, so that each discourse reference marker is the unique canonical name for an individual or object mentioned in the discourse. The scoping of quantifiers and negations has also been determined by the semantic processing. This allows the transformation of a DRS to FOL to be rather simple. The basic ideas can be briefly introduced here; a more thorough discussion of the special treatment of queries is provided below. The conditions of a DRS are conjoined. Those conditions may include some that are already related by logical operators (if-then, or, not), in which case the logical form includes the same operators. Discourse referents introduced in the consequent of an if-then construction may introduce r quantifiers: these are given narrow scope relative to quantifiers in the consequent (cf. Kamp's notion of a &quot;subordinate&quot; 156 Computational Linguistics, Volume 15, Number 3, September 1989 Kathleen Dahlgren, Joyce McDowell, and Edward P. Stabler, Jr. Knowledge Representation for Commonsense Reasoning with Text DRS in Kamp 1981; Asher and Wada 1988). Any quantifiers needing wide scope will already have been moved out of the consequent by earlier semantic processing. The DRSs of questions contain special operators that, like the logical operators, take DRS arguments that represent the scope of the material being questioned. A similar indication is needed in the logical form. In the case of a yes-no question, we introduce a special vacuous quantifier just to mark the scope of the questioned material for special treatment by the problem solver (see below). In the case of wh-questions, a special wh-quantifier is introduced, again indicating the scope of the questioned material and triggering a special treatment by the problem solver. Verbs of propositional attitude and other structures with opaque contexts are treated as, in effect, introducing new isolated subtheodes or &quot;worlds.&quot; For example, &quot;John believes all men are mortal&quot; is represented as a relation of belief holding between John and a logical form (see below for more detail), though it is recognized that this approach will need to be supplemented to handle anaphora and propositional nominals (cf., e.g., Asher 1988).</Paragraph> </Section> <Section position="16" start_page="154" end_page="154" type="metho"> <SectionTitle> 2.2.4.2 THE CONVERSION TO SPECIALIZED CLAUSAL FORMS </SectionTitle> <Paragraph position="0"> Considerations of efficiency motivate a further transformation in our logical forms. After the first step of processing, we have standard first-order logical forms, except that they may include special quantifiers indicating questioned material. Consider first those logical forms that are not inside the scope of any question quantifier. These are taken as representations of potential new knowledge for the textual data base. Since the inference system must solve problems without user guidance, it would be infeasible to reason directly from the first-order formulations. Clausal forms provide an enormous computational advantage. For these reasons, we transform each sentence of first-order logic into a clausal form with a standard technique (cf., e.g., Chang and Lee 1973), introducing appropriate Skolem functions to replace existentially quantified variables. The textual database can then be accessed by a clausal theorem prover. In the current system, we use efficient Horn clause resolution techniques (see below), so the knowledge representation is further restricted to Horn clauses, since completeness is less important than feasible resource use in the present application. Definite clauses are represented in a standard Prolog format, while negative clauses are transformed into definite clauses by the addition of a special positive literal &quot;false(n),&quot; where n is an integer that occurs in no other literal with this predicate. This allows a specialized incomplete treatment of negation-as-inconsistency (cf.</Paragraph> <Paragraph position="1"> Gabbay and Sergot 1986). The definite clause translations of the text can then be inserted into a textual knowledge base for use by the reasoning component.</Paragraph> <Paragraph position="2"> The presence of question quantifiers triggers a special treatment in the conversion to clausal form. Our problem solver uses standard resolution techniques: to prove a proposition, we show that its negation is incompatible with the theory. Accordingly, the material inside the scope of a question operator is treated as if it were negated, and this implies an appropriately different treatment of any quantifiers inside the scope of the operators.</Paragraph> <Paragraph position="3"> In the architecture of the system, the reasoning module is broken into two parts: the specialized query processing system and a general purpose problem solver. The special processing of queries is described in detail below. The problem solver is based on a straightforward depth-bounded Horn clause proof system, implemented by a Prolog metainterpreter (e.g. Sterling and Shapiro 1986). The depth bound can be kept fixed when it is known that no proofs should exceed a certain small depth. When a small depth bound is not known, the depth can be allowed to increase iteratively (cf. Stickel 1986), yielding a complete SLD resolution system. This proof system is augmented with negation-as-failure for predicates known to be complete (see the discussion of open and closed world assumptions below), and with a specialized incomplete negation-as-inconsistency that allows some negative answers to queries in cases where negation-as-failure cannot be used.</Paragraph> <Paragraph position="4"> The RELEVANCE module will have the responsibility of determining the relevance of a particular text to a particular user. The text and user profiles will be processed through the system in the usual way resulting in two textual data bases, one for the target text and one for the profile. Target and profile will then be compared for relevance and a decision made whether to dispatch the target to the profiled user or not. The relevance rules are a current research topic. The commonsense knowledge representations will form the primary basis for determining relevance.</Paragraph> </Section> <Section position="17" start_page="154" end_page="154" type="metho"> <SectionTitle> 3 NAIVE SEMANTICS IN THE KT SYSTEM </SectionTitle> <Paragraph position="0"> From the foregoing brief overview of the KT system, it should be clear that naive semantics is used throughout the system for a number of different processing tasks. In this section we show why each of these tasks is a problem area and how NS can be used to solve it.</Paragraph> <Section position="1" start_page="154" end_page="154" type="sub_section"> <SectionTitle> 3.1 PREPOSITIONAL PHRASE ATTACHMENT 5 </SectionTitle> <Paragraph position="0"> The proper attachment of post-verbal adjuncts is a notoriously difficult task. The problem for prepositional phrases can be illustrated by comparing the following sentences.</Paragraph> <Paragraph position="1"> 12. \[S The government \[VP had uncovered \[NP an entire file \[PP about the scheme\]\]\]\]. 6 13. \[S Levine's lawyer \[VP announced \[NP the plea bargain\] \[PP in a press conference\] Computational Linguistics, Volume 15, Number 3, September 1989 157 Kathleen Dahigren, Joyce McDowell, and Edward P. Stabler, Jr. Knowledge Representation for Commonsense Reasoning with Text 14. \[S\[S The judge adjourned the hearing\] \[PP in the</Paragraph> <Paragraph position="3"> Each of these sentences takes the canonical form Subject-Verb-Object-PP. The task for the processing system is to determine whether the PP modifies the object (i.e., the PP is a constituent of the NP, as in (12)), the verb (i.e., the PP is a constituent of the VP, as in (13)), or the sentence (i.e., the PP is an adjunct to S, as in (14)). Some deny the need for a distinction between VP and S-modification. The difference is that with S-modification, the predication expressed by the PP has scope over the subject, while in VP-attachment it does not.</Paragraph> <Paragraph position="4"> For example, in (15), &quot;in the park&quot; applies to Levine, while in (16), &quot;with 1,000 dollars&quot; does not apply to Levine.</Paragraph> <Paragraph position="5"> 15. Levine made the announcement in the park.</Paragraph> <Paragraph position="6"> 16. Levine bought the stock with 1,000 dollars.</Paragraph> <Paragraph position="7"> A number of solutions for the problem presented by post-verbal prepositional phrases have been offered.</Paragraph> <Paragraph position="8"> The most common techniques depend on structural (Frazier and Fodor 1978), semantic (Ford, Bresnan, and Kaplan 1982), or pragmatic (Crain and Steedman 1985) tests. MODL (McCord 1987) employs a combination of syntactic and semantic information for PP attachment.</Paragraph> <Paragraph position="9"> Independently, we formulated a preference strategy for PP attachment which uses ontological, generic and syntactic information to cover 99% of the cases in an initial test corpus, and which is 93% reliable across a number of types of text. This is the preference strategy we employ in KT. The PP attachment rules make use of information about the verb, the object of the verb, and the object of the preposition. A set of global rules is applied first, and if these fail to find the correct attachment for the PP, a set of rules specific to the preposition are tried. Each of these latter rules has a default. The global rules are stated informally in (17). with example sentences.</Paragraph> <Paragraph position="10"> 17a. time(POBJ)-, s_attach(PP) If the object of the preposition is an expression of time, then S-attach the PP.</Paragraph> <Paragraph position="11"> The judge adjourned the hearing in the afternoon. null b. lexical(V+Prep)--~ vp_attach(PP) If the verb and preposition form a lexicalized complex verb, then VP-attach the PP.</Paragraph> <Paragraph position="12"> The defense depended on expert witnesses.</Paragraph> <Paragraph position="13"> c. Prep=of--~ np_attach(PP) If the preposition is of then NP-attach the PP.</Paragraph> <Paragraph position="14"> The ambulance carried the victim of the shooting. null</Paragraph> <Paragraph position="16"> If the verb is an intransitive verb of motion and the object of the preposition is a place then VP-attach the PP.</Paragraph> <Paragraph position="17"> The press scurried about the courtroom.</Paragraph> <Paragraph position="19"> If the verb is intransitive and the object of the preposition is a place, temporal, or abstract, then S-attach the PP.</Paragraph> <Paragraph position="20"> Levine worked in a brokerage house.</Paragraph> <Paragraph position="21"> f. epistemic(POBJ)--* s_attach(PP) If the prepositional phrase expresses a propositional attitude, then attach the PP to the S. Levine was guilty in my opinion.</Paragraph> <Paragraph position="22"> g. xp(.. ~_dj-PP...)--~ xp_attach(PP) If PP follows an adjective, then attach the PP to the phrase which dominates and contains the adjective phrase.</Paragraph> <Paragraph position="23"> Levine is young for a millionaire.</Paragraph> <Paragraph position="24"> h. measure(DO)--~ np_attach(PP) If the direct object is an expression of measure, then NP-attach the PP.</Paragraph> <Paragraph position="25"> The defendant had consumed several ounces of whiskey.</Paragraph> <Paragraph position="26"> i. comparative-* np_attach(PP) If there is a comparative construction, then NP-attach the PP.</Paragraph> <Paragraph position="27"> The judge meted out a shorter sentence than usual.</Paragraph> <Paragraph position="28"> j. mental(V) & medium(POBJ)-~ vp_attach(PP) If the verb is a verb of saying, and the object of the preposition is a medium of expression then VP-attach the PP.</Paragraph> <Paragraph position="29"> Levine's lawyer announced the plea bargain on television.</Paragraph> <Paragraph position="30"> Example 14 is handled by global rule (17a). Example 13 is handled by global rule (17j). The global rules are inapplicable with example 12, so the rules specific to &quot;about&quot; are called. These are shown below. 18a. intrsmsitive(V) & mental(V)--~ vp_attach(PP) If the verb is an intransitive mental verb, then VP-attach the PP.</Paragraph> <Paragraph position="31"> Levine spoke about his feelings.</Paragraph> <Paragraph position="32"> b. Elsewhere--* np_attach(PP) Otherwise, NP-attach the PP.</Paragraph> <Paragraph position="33"> The government had uncovered an entire file about the scheme.</Paragraph> <Paragraph position="34"> As a filrther example, the specific rules for &quot;by&quot; uses both generic (19a,b) and ontological (19c) knowledge. 19 a. nom(DO)--~ np_attach(PP) If the direct object is marked as a nominalization, then NP-attach the PP.</Paragraph> <Paragraph position="35"> The soldiers withstood the attack by the enemy. null b. location(DO,POBJ)--~ np_attach(PP) If tJ~e rel~.tion between the d.ireot obJeot and t~he object of t~he preposition Is one of loos,~ion, then lq'P-~t~ch t, he PP.</Paragraph> <Paragraph position="36"> The clerk adjusted the microphone by the witness stand.</Paragraph> <Paragraph position="37"> 158 Computational Linguistics, Volume 15, Number 3, September 1989 Kathleen Dahlgren, Joyce McDowell, and Edward P. Stabler, Jr. Knowledge Representation for Commortsense Reasoning with Text c. proposltional(DO) & sentlent(POBJ)--, np_at-</Paragraph> <Paragraph position="39"> If the direct object is propositional and the object of the preposition is sentient, then NP-attach the PP.</Paragraph> <Paragraph position="40"> The judge read out the statement by Levine. d. Elsewhere---> s_attach(PP) Otherwise, S-attach the PP.</Paragraph> <Paragraph position="41"> The lawyers discussed the case by the parking lot.</Paragraph> <Paragraph position="42"> These PP-attachment preference rules are remarkably successful when applied to real examples from actual text. However, they are not foolproof and it is possible to construct counterexamples. Take the global rule illustrated in (17a). We can construct a counterexample as in (20).</Paragraph> <Paragraph position="43"> 20. John described the meeting on Jan. 20th. Sentence 20 is ambiguous. Jan. 20th can be the time of the describing or the time of the meeting. Perhaps there is a slight intuitive bias toward the latter interpretation, but the rules will assign the former interpretation. This is a counterexample only because &quot;meeting&quot; is a TEMPORAL noun and can plausibly have a time feature. Compare (21), which is identical except for the ontological attachment of the direct object and which is handled correctly by the global rule.</Paragraph> <Paragraph position="44"> 21. John described the proposal on Jan. 20th.</Paragraph> <Paragraph position="45"> The problem of the interpretation of event nominals is a research topic we are working on.</Paragraph> <Paragraph position="46"> The PP attachment rules are applied in the module DISAMBIG, which produces a disambiguated parse from the output of the MODL parser. The first step is to identify the sentence elements that form their inputs for each clause using find..args. The output of find_args is a list of the the direct object, object of the preposition, preposition, and main verb of the clause and the index of the clause (main, subordinate, and so on). The PP-attachment rules are in place and functional for one post-verbal prepositional phrase. Where more than one post-verbal prepositional phrase occurs, the current default is to attach the second PP to the NP of the first PP. However, this will not get the correct attachment in cases like the following.</Paragraph> <Paragraph position="47"> 22. The judge passed sentence on the defendant in a terse announcement.</Paragraph> <Paragraph position="48"> A planned extension of the PP-attachment functionality will attack this problem by also keeping a stack of prepositions. The top of the stack will be the head of the rightmost PP. The attachment rules will be applied to PPs and the other constituents in a pairwise fashion until all are attached.</Paragraph> </Section> <Section position="2" start_page="154" end_page="154" type="sub_section"> <SectionTitle> 3.2 WORD SENSE DISAMBIGUATION </SectionTitle> <Paragraph position="0"> The word sense disambiguation method used in the system is a combined local ambiguity reduction method (Dahlgren 1988b). The method is local because word senses are disambiguated cyclically, from the lowest S-node up to the matrix node. Only when intrasentential sources of information fail are other sentences in the text considered by the disambiguation method. The algorithm is combined because it employs three sources of information. First it tries fixed and frequent phrases, then word-specific syntactic tests, and finally naive semantic relationships in the clause. If the fixed and frequent phrases fail, the syntactic and naive semantic rules progressively reduce the number of senses relevant in the clausal context. The algorithm was developed by considering concordances of seven nouns with a total of 2,193 tokens of the nouns, and concordances of four verbs with a total of 1,789 tokens of the verbs.</Paragraph> <Paragraph position="1"> The algorithm is 96% accurate for the nouns in these concordances, and 99% accurate for the verbs in these concordances.</Paragraph> <Paragraph position="2"> Fixed phrases are lists of phrases that decisively disambiguate the word senses in them. For example, the noun &quot;hand&quot; has 16 senses. Phrases such as &quot;by hand,&quot; &quot;on hand,&quot; &quot;on the one hand&quot; have only one sense.</Paragraph> <Paragraph position="3"> Syntactic tests either reduce the number of relevant senses, or fully disambiguate. For nouns, syntactic tests look for presence or absence of the determiner, the type of determiner, certain prepositional phrase modifiers, quantifiers and number, and noun complements. For example, only five of the 16 senses of &quot;hand&quot; are possible in bare plural noun phrases. For verbs, syntactic tests include the presence of a reflexive object, elements of the specifier, such as particular adverbs the presence of a complement of the verb and particular prepositions. For example, the verb &quot;charge&quot; has only its reading meaning &quot;indict&quot; when there is a VP-attached PP where the preposition is &quot;with&quot; (as determined by the prepositional phrase attachment rules).</Paragraph> <Paragraph position="4"> Syntactic tests are encoded for each sense of each word. The remainder of this section will illustrate disambiguation using naive semantic information and give examples of the naive semantic rules. (The complete algorithm may be found in Dahlgren 1988a).</Paragraph> <Paragraph position="5"> Naive semantic information was required for at least a portion of the disambiguation in 49% of the cases of nouns in the concordance test. Naive semantic inference involves either ontological similarity or generic relationships. Ontological similarity means that two nouns are relatively close to each other in the ontology, both upwards and across the ontology. If there is no ontological similarity, generic information is inspected.</Paragraph> <Paragraph position="6"> Generic information for the ambiguous noun, other nouns in the clause, the main verb, often disambiguate an ambiguous noun.</Paragraph> <Paragraph position="7"> Ontological similarity is tested for in several syntactic constructions: conjunction, nominal compounds, possessives, and prepositional phrase modifiers. Many Computational Linguistics, Volume 15, Number 3, September 1989 159 Kathleen Dahlgren, Joyee McDowell, and Edward P. Stabler, Jr. Knowiedge Representation for Commonsense Reasoning with Text of the 16 senses of &quot;hand&quot; are ontologically distinct, as shown in Table 7.</Paragraph> <Paragraph position="8"> I. HUMAN human body part 2. DIRECTION right or left 3. INSTRUMENT by hand 4. SOCIAL power, authority 5. TEMPORAL applause 6. ROLE laborer 7. ARTIFACT part of a clock In (23), only the HUMAN and ROLE senses (I and 6) are possible, by ontological similarity. Generic knowledge of the verb &quot;clear&quot; is inspected for the final disambiguation to sense 6.</Paragraph> <Paragraph position="9"> 23. The farmer and his hand cleared the field.</Paragraph> <Paragraph position="10"> In contrast, in (24), the relevant senses of &quot;hand&quot; are the HUMAN and ARTIFACT senses (1 and 7).</Paragraph> <Paragraph position="11"> 24. His arm and hand were broken.</Paragraph> <Paragraph position="12"> At the point in the algorithm where naive semantic tests are invoked, syntactic tests have already eliminated the ARTIFACT sense, which does not occur with a personal pronoun. Thus the HUMAN sense is selected. In (25), only the ARTIFACT sense (7) is possible, by ontological similarity of &quot;clock&quot; and &quot;hand.&quot; 25. The clock hand was black.</Paragraph> <Paragraph position="13"> In (26), again the HUMAN and ROLE senses (1 and 6) are the only relevant ones by ontological similarity.</Paragraph> <Paragraph position="14"> Selection restrictions on &quot;shake&quot; complete the disambiguation. null 26. John shook the man's hand.</Paragraph> <Paragraph position="15"> In (27), sense 4 is selected because &quot;affair&quot; and sense 4 are both SOCIAL.</Paragraph> <Paragraph position="16"> 27. John saw his hand in the affair.</Paragraph> <Paragraph position="17"> In (28), sense I is selected because both sense 1 and a sense of &quot;paper&quot; are attached to PHYSICAL.</Paragraph> <Paragraph position="18"> 28. The judge had the paper in his hand.</Paragraph> <Paragraph position="19"> The word sense disambiguation algorithm tests for generic relationships between the ambiguous noun and prepositional phrases modifiers, adjective modifiers, and the main verb of the sentence. Two of the nine senses of &quot;court&quot; are shown in Table 8. In &quot;the court listened to testimony,&quot; generic information for the second sense of &quot;court&quot; can be used to select sense 2. The generic information includes knowledge that one of the functions of courts has to do with testimony. In (29), sense 1 of &quot;court&quot; is selected because the generic representation of &quot;court&quot; contains information that witness stands are typical parts of courtrooms.</Paragraph> <Paragraph position="21"> PLACE Typically, it has a bench, jury box, and witness stand.</Paragraph> <Paragraph position="22"> Inherently its function is for a judge to conduct trials in.</Paragraph> <Paragraph position="23"> It is part of a courthouse.</Paragraph> <Paragraph position="24"> INSTITUTION Typically, its function is justice. Examples are the Supreme Court and the superior court. Its location is a courtroom. Inherently it is headed by a judge, has bailiffs, attorneys, court reporters as officers.</Paragraph> <Paragraph position="25"> Participants are defendants, witnesses and jurors. The function of a court is to hear testimony, examine evidence and reach a verdict. It is part of the 29. The witness stand in Jones's court is made of oak. In (30), the adjective &quot;wise&quot; narrows the relevant senses from nine to the two INSTITUTION senses of &quot;court.&quot; 30. 'The wise court found him guilty. Generic knowledge of one sense of the verb &quot;find&quot; is then used to select between the court-of-law sense (2) and the-royal-court sense (4). Verb selection restrictions are powerful disambiguators of nouns, as many computational linguists have observed. In (31), the verb Typically, if someone is charged, next they are indicted in court, convicted or acquitted. They are charged because they have committed a crime or the person who charges them suspects they have.</Paragraph> <Paragraph position="26"> Inherently, the charger and chargee are sentient, and the thing charged with is a crime.</Paragraph> <Paragraph position="27"> Inherently, if someone charges that something is true, that someone is sentient, his goal is that it be known, and the something is bad.</Paragraph> <Paragraph position="28"> Typically, if someone charges someone else an amount for something, the chargee has to pay the amount to the charger, and the chargee is providing goods or services. Inherently, the charger and chargee are sentients, the amount is a quantity of money, and the goal of the charger is to have the chargee pay the amount.</Paragraph> <Paragraph position="29"> Generic Information for Thi'ee Senses of charge 160 Computational Linguistics, Volume 15, Number 3, September 1989 Kathleen Dahlgren, Joyce McDoweli, and Edward P. Stabler, Jr. Knowledge Representation for Commonsense Reasoning with Text &quot;last&quot; requires a TEMPORAL subject, thus disambiguating &quot;hand.&quot; 31. They gave him a hand which lasted 10 minutes. Just as selectional restrictions of verbs disambiguate nouns, their arguments are powerful disambiguators of verbs. Subject, object, and oblique arguments are all taken into account by the algorithm. (32) and (33) illustrate the way that objects can disambiguate the verb &quot;charge,&quot; generic entries for which are shown in Table 9.</Paragraph> <Paragraph position="30"> In (32) the SENTIENT object selects sense 1. In (33), the MONEY object selects sense 3.</Paragraph> <Paragraph position="31"> 32. The state charged the man with a felony. 33. The state charged ten dollars for the fine. The verb &quot;present&quot; can be disambiguated by its subject. It has at least three senses: 34. present l--&quot;give&quot; present2---&quot;introduce&quot; present3--&quot;arrive in the mind of&quot; Senses 1 and 2 require SENTIENT subjects, so the third sense is selected in (35).</Paragraph> <Paragraph position="32"> 35. The decision presented a problem to banks. (36) illustrates subject and object disambiguation. The SENTIENT subject narrows the possibilities to senses l and 2, and the &quot;give&quot; sense (1) is selected because it requires a PHYSICAL object argument.</Paragraph> <Paragraph position="33"> 36. John presented a bouquet to Mary.</Paragraph> <Paragraph position="34"> The disambiguation method first tries fixed and frequent phrases, then syntactic tests, and finally naive semantic information. Each set of rules reduces the number of relevant senses of a word in the sentential (and extrasentential) context. There is a fixed set of commonsense rules for nouns and another one for verbs. They are tried in an order that inspects the main verb of the clause last, because the main verb often chooses between the last two relevant senses. An example of a noun rule is ppmod, which considers an ambiguous noun in relation to the head of a prepositional phrase modifier attached to the same higher NP as the ambiguous noun. There are two versions of the rule, one which looks for ontological similarity between senses of the ambiguous noun and the head of the PP, and one which looks for generic relationships between them. The output of find_args (in DISAMBIG, see Section 3. l) is used as a simplified syntactic structure inspected by these rules. This provides information as to whether or not a head noun is modified by a prepositional phrase. In the first version of the rule, the ontological attachment of the head of the PP is looke d up and then senses of the ambiguous word with that same ontological attachment are selected. SI is the list of senses of the ambiguous noun still relevant when the rule is invoked. $2 is the list of senses reduced by the rule if it succeeds. If the first version fails, the second is invoked. It looks for a generic relationship between senses of the ambiguous word and the head of the PP.</Paragraph> <Paragraph position="35"> 37. ppmod(Ambig_Word,{.. _A_rnbig_Word,Prep,</Paragraph> <Paragraph position="37"/> </Section> <Section position="3" start_page="154" end_page="154" type="sub_section"> <SectionTitle> 3.3 QUANTIFIER SCOPING </SectionTitle> <Paragraph position="0"> The semantic module of the KT system is capable of generating alternative readings in cases of quantifier scope ambiguities, as in the classic case illustrated in (38).</Paragraph> <Paragraph position="1"> 38. Every man loves a woman.</Paragraph> <Paragraph position="2"> In this example either the universally quantified NP (&quot;every man&quot;) or the existentially quantified NP (&quot;a woman&quot;) can take widest scope. In such sentences, it is generally assumed that the natural scope reading (leftmost quantified expression taking widest scope) is to be preferred and the alternative reading chosen only under explicit instructions of some sort (such as input from a user, for example, or upon failure to find a proper antecedent for an anaphoric expression). Under this assumption, in a sentence like (39), the indefinite NP would preferentially take widest scope.</Paragraph> <Paragraph position="3"> 39. A woman loves every man.</Paragraph> <Paragraph position="4"> But there are a number of cases similar to (39) where an expression quantified by &quot;every&quot; appears to the right of an indefinite NP and still seems to take widest scope.</Paragraph> <Paragraph position="5"> Ioup (1975) has discussed this phenomena, suggesting that expressions such as &quot;every&quot; take widest scope inherently. Another computational approach to this problem is to assign precedence numbers to quantifiers, as described in McCord (1987). However, our investigation has shown that commonsense knowledge plays at least as large a role as any inherent scope properties of universal quantifiers.</Paragraph> <Paragraph position="6"> Consider (40). In the natural scope reading, &quot;every lawyer&quot; takes scope over &quot;a letter&quot; and we have several letters, one from each of the lawyers, i.e., several tokens of a one-to-one relationship. In the alternative reading, &quot;a letter&quot; takes scope over &quot;every lawyer&quot; and we have only one letter, i.e., one token of a many-to-one relationship relationship. Both scope readings are plausible for (40).</Paragraph> <Paragraph position="7"> 40. The judge read a letter from every lawyer.</Paragraph> <Paragraph position="8"> In (41) only the alternative reading (several tokens of one-to-one) is plausible.</Paragraph> <Paragraph position="9"> 41. The politician promised a chicken in every pot.</Paragraph> <Paragraph position="10"> In (42), however, only the natural reading (one token of many-to-one) is plausible.</Paragraph> <Paragraph position="11"> 42. The prince sought a wife with every charm.</Paragraph> <Paragraph position="12"> Even in (40), however, speakers prefer the one-to-one relationship, the alternative reading. That is, speakers prefer the reading that denotes several tokens of a one-to-one relationship (several letters) over one which denotes one token of a many-to-one relationship unless Computational Linguistics, Volume 15, Number 3, September 1989 161 Kathleen Dahlgren, Joyce McDowell, and Edward P. Stabler, Jr. Knowledge Representation for Commonsense Reasoning with Text there is strong commonsense knowledge to override this preference. We know that in our culture princes can h~ive only one wife, so in the case of (42) speakers prefer one token (one wife) of many-to-one to several tokens of one-to-one. Similar arguments apply to the following examples (43)-(45), which correspond to (40)-(42) respectively. null 43. A judge decided on every petition.</Paragraph> <Paragraph position="13"> 44. A lawyer arrived from every firm.</Paragraph> <Paragraph position="14"> 45. A company agent negotiated with every union.</Paragraph> <Paragraph position="15"> Thus, if there is an inherent tendency for universal quantifiers to take widest scope where scope ambiguity is possible, it derives from the human preference for viewing situations involving universally quantified entities as many tokens of one-to-one relationships, s In KT, first we prefer wide scope for the universal quantifier. In cases where this is not the natural scope interpretation, (i.e., the universal quantifier is to the right of a containing existentially quantified NP), we can use facts encoded in the generic data base to override the preference. For example, when processing (42) we would discover that a man may have only one wife. The generic entry for &quot;wife&quot; tells us that &quot;wife&quot; is a role in a &quot;marriage.&quot; The generic entry for &quot;marriage&quot; tells us that &quot;marriage&quot; is a relation between exactly one husband and exactly one wife. This knowledge forces the many-to-one interpretation. The cases where this is necessary turn out to be rare. Curiously, the preposition &quot;with&quot; correlates very highly with many-to-one relationships. Thus our strategy for the present has been to consider overriding the preference only when the universal quantifier is in an NP which is the object of &quot;with.&quot; In these cases we access the generic knowledge, as described above.</Paragraph> </Section> <Section position="4" start_page="154" end_page="154" type="sub_section"> <SectionTitle> 3.4 OPAQUE CONTEXTS </SectionTitle> <Paragraph position="0"> In KT, clauses in opaque contexts (embedded under propositional attitude verbs such as &quot;hope,&quot; &quot;believe,&quot; &quot;deny&quot;) are handled by asserting the predicates generated from the clause into partitioned databases, which correspond to the delineated DRSs of Asher (1987). Each partition is associated with the speaker or the individual responsible for the embedded clause.</Paragraph> <Paragraph position="1"> Reasoning can then proceed taking into account the reliability and bias of the originator of the partitioned statements, as in Section 2.2.4.1 An example follows.</Paragraph> <Paragraph position="2"> 46. Text: Meese believes that Levine is guilty.</Paragraph> </Section> </Section> <Section position="18" start_page="154" end_page="154" type="metho"> <SectionTitle> 3.5 MODALS 9 </SectionTitle> <Paragraph position="0"> The English modals can, may, must, will, and should are high-frequency items in all kinds of texts. They can be easily parsed by a single rule similar to the rules that handle auxiliary &quot;have&quot; and &quot;be&quot; because all the modals occupy the same surface syntactic position (i.e., the first element in the auxiliary sequence). However, the~ modals present some considerable problems for semantic interpretation because they introduce ambiguities and induce intensional contexts in which possibility, necessity, belief, and value systems play a role. In the KT system, we are concerned with what is known by the system as a result of reading in a modal sentence.</Paragraph> <Paragraph position="1"> In particular we are interested in what status the system assigns to the propositional content of such a sentence.</Paragraph> <Paragraph position="2"> To illustrate the problem, if the system reads &quot;Levine engaged in insider trading,&quot; then an assertion can justifiably be added to the knowledge base reflecting the fact that Levine engaged in insider trading. The same is true if the system reads &quot;The Justice Department knows that Levine engaged in insider trading.&quot; But this is not the case if the system reads &quot;The Justice Department believes that Levine engaged in insider trading.&quot; In this case the statement that Levine engaged in insider trading must be assigned some status other than fact. Specifically, since &quot;believe&quot; introduces an opaque context, the propositional content of the embedded clause would be assigned to a partitioned data base linked to the speaker the Justice Department, as described in the previous section. A similar problem exists in modal sentences such as &quot;Levine may have engaged in insider trading.&quot; There are two types of modal sentences. In Type I modal sentences the truth value of the propositional content is evaluated with respect to the actual world or a set of possible other states of the actual world directly inferable from the actual world. Examples are 47. Levine must have engaged in insider trading.</Paragraph> <Paragraph position="3"> 48. The Justice Department will prosecute Levine.</Paragraph> <Paragraph position="4"> 49. Levine can plead innocent.</Paragraph> <Paragraph position="5"> In Type II modal sentences, we say that a second speech act is &quot;semantically embedded&quot; in the modal sentence. The modal sentence is successful as an assertion just in case the secondary speech act is in effect in the actual world. In these cases the truth value of the propositional content is evaluated with respect to some set of deontic or normative worlds. The modal is viewed as a quantifier cum selection function. Thus, for a sentence of the form/~ = NP Modal VP, I~ is true in the actual world just in case NP VP is true in at least one/every world (depending on Modal) in the set of deontic or normative worlds selected by Modal. Examples are 50. Levine must confess his guilt.</Paragraph> <Paragraph position="6"> 51. Levine may make one phone call.</Paragraph> <Paragraph position="7"> 52. Levine should get a good attorney.</Paragraph> <Paragraph position="8"> In (50) a command is semantically embedded in the assertion; in (51) a permission is semantically embedded in the assertion; and in (51) the issuance of a norm is semantically embedded in the assertion.</Paragraph> <Paragraph position="9"> Type I modal sentences are of assertive type according to the speech act classification scheme in Searle and Vanderveken (1985). These include the standard assertions, reports, and predictions, and a proposed new 162 Computational Linguistics, Volume 15, Number 3, September 1989 Kathleen Dahlgren, Joyce McDowell, and Edward P. Stabler, Jr. Knowledge Representation for Commonsense Reasoning with Text type, quasi-assertion. They must be distinguished in a text-understanding system from Type II modal sentences that embed other types of speech acts, because only Type I modal sentences make a contribution to the textual knowledge base. In addition, some modal sentences are ambiguous between Type I and Type II, for example (47), (51), and (50).</Paragraph> <Paragraph position="10"> For the KT system, disambiguating the ambiguous modals &quot;may&quot; and &quot;must&quot; results in changing the syntactic input to the semantic module. The surface syntactic parse that is output by the parser is converted into the equivalent logical form where the epistemic (Type I) uses of ambiguous modals are treated as sentential operators and the nonepistemic uses of ambiguous modals (Type II) are treated as modifiers of the verb. Sentences containing ambiguous modals can be assigned the correct status by a simple disambiguation algorithm that makes appeal to the ontological classification of the main verb, specifically whether or not the verb is STATIVE, following Steedman (1977). Disambiguation takes place in DISAMBIG, the same module that converts the labelled bracketing to a modified parse for input to the DRT semantic translator. At this point, a determination is made whether the modal is a sentential operator or a modifier of the verb. The propositional content of quasi-assertions and predictions can be added directly to the dynamically constructed textual data base if they are appropriately marked with probability ratings. On this view, &quot;will&quot; and one sense of &quot;may&quot; are taken as denoting strong and weak prediction and are not viewed as tenses. That is, when using these modals, the speaker is indicating his confidence that there is a high/moderate probability of the propositional content being true in the future. In the present state of the system, every predicate contains a tense argument and there is a tense predicate relating every tense argument to a tense value (such as &quot;pres.,&quot; &quot;future,&quot; etc.).1deg In a planned extension to the system these tense predicates will also contain probability ratings. For example, given the continuum of speaker commitment to the truth of the statement &quot;Leyine engaged in insider trading&quot; illustrated in (53), we would have the corresponding predicates in the DRS shown in (54).</Paragraph> <Paragraph position="11"> 53. Full Assertion: Levine engaged in insider trading Strong Quasi-Assertion: Levine must have engaged in insider trading Weak Quasi-Assertion: Levine may have engaged in insider trading 54. Full Assertion: engage(el,levine), tense(el ,past, 1) Strong Quasi-Assertion: engage(el,levine), tense (el,past,0.9) Weak Quasi-Assertion: engage(el,levine), tense (el,past,0.5) This hierarchy reflects the &quot;epistemic paradox&quot; of Karttunen (1971), in which he points out that in standard modal logic must(P) or necessarily, P is stronger than plain assertion whereas epistemically-must(P) is weaker than plain assertion. This results from the fact that the standard logic necessity operator quantifies over every logically possible world, plain assertion of P is evaluated with respect to the actual world, but the epistemic modal operator quantifies only over the epistemically accessible worlds, a set which could possibly be null.</Paragraph> <Paragraph position="12"> Assertions of possibility (49) tC/igger the inferring of enabling conditions. For any event there is a set of enabling conditions that must be met before the event is possible. For John to play the piano, the following conditions must be met: 55. 1. John knows how to play the piano.</Paragraph> <Paragraph position="13"> 2. John has the requisite permissions (if any).</Paragraph> <Paragraph position="14"> 3. A piano is physically available to John.</Paragraph> <Paragraph position="15"> 4. John is well enough to play.</Paragraph> <Paragraph position="16"> ....</Paragraph> <Paragraph position="17"> These can be ordered according to saliency as above.</Paragraph> <Paragraph position="18"> This, we claim, is why the sentence &quot;John can play the piano&quot; most often receives the interpretation (I), less often (2), and practically never (3) or (4) unless explicitly stated, as in &quot;John can play the piano now that his mother has bought a new Steinway.&quot; The enabling conditions are encoded in KT as part of the generic representation for verbs. When a modal sentence is interpreted as a full assertion of possibility (poss(p)), this triggers the inference that the most salient of the enabling conditions is in fact true. The difference between poss(p) and p being processed for KT, is that ifp is output, then p is added to the textual database and the most salient enabling condition is also inferred. But if poss(p) is output, then only the most salient enabling condition is inferred, but p is not added to the textual database. Notice that this simply reflects the fact that if I say &quot;John can play the piano,&quot; I am not saying that John is playing the piano at that very moment.</Paragraph> <Paragraph position="19"> Type II modal sentences present a more complex problem for interpretation. The commands, permissions, and norms reported in Type II modal sentences are asserted into partitioned databases in the same way as clauses in opaque contexts. The only difference is that in most cases the issuer of the command, permission, or norm reported in a modal sentence is not known. Semantic translation in DRT proceeds via categorial combination. By the time the modal sentence reaches the DRT module, the semantic type of the modal is unambiguous and the appropriate lexical entry can be retrieved. The creation of appropriate predicates to express the variety of modal statements is the task of the DRT module.</Paragraph> </Section> <Section position="19" start_page="154" end_page="154" type="metho"> <SectionTitle> 4 THE QUERY' SYSTEM 4.1 OPEN AND CLOSED WORLD ASSUMPTIONS </SectionTitle> <Paragraph position="0"> It is well known that negation-as-failure is a sound extension of SLD resolution only when the database is Computational Linguistics, Volume 15, Number 3, September 1989 163 Kathleen Dahlgren, Joyce McDowell, and Edward P. Stabler, Jr. Knowledge Representation for Commonsense Reasoning with Text complete, i.e., when it represents a closed world (Lloyd 1984). Since our databases will always include some predicates about which we have only incomplete information, we cannot assume a completely closed world.</Paragraph> <Paragraph position="1"> The open world assumption, though, makes it unsound to use the very useful negation-as-failure rule. Fortunately, it is well known that we can restrict the use of this rule to just those predicates known to be complete, keeping an open world assumption for other predicates.</Paragraph> <Paragraph position="2"> We accordingly specify that some of our general knowledge comprises a complete, closed world in the appropriate sense, but we do not make this assumption about textual knowledge.</Paragraph> </Section> <Section position="20" start_page="154" end_page="154" type="metho"> <SectionTitle> 4.2 FUNCTIONING OF THE QUERY SYSTEM </SectionTitle> <Paragraph position="0"> Queries are handled just like text up through conversion to FOL. The FOL form of the query is then passed to REASONER, which decides which database is the most likely source of the answer. REASONER can access the textual database, the verb and noun generic databases, and the ontology. The reasoning is complex and dependent on whether the query form is ontological, factual, or generic. A search sequence is then initiated depending on these factors. The form of the answer depends on the search sequence and the place where the answer is found.</Paragraph> <Paragraph position="1"> The form of the answer depends on the reliability of the knowledge encoded in the database where the answer is found. The text is considered authoritative. If an answer is found in the text, search is terminated. The ontology is considered a closed world (see discussion above).</Paragraph> <Paragraph position="2"> This means that yes/no ontological questions are answered either &quot;yes&quot; or &quot;no.&quot; The textual and generic databases are considered an open world. If an answer is not found, and no further search is possible, the system concludes, &quot;I don't know.&quot; Answers found in the generic databases are prefaced by &quot;Typically,&quot; for information found in the first list (of typical features) or &quot;Inherently,&quot; for answers found in the second list (of inherent features).</Paragraph> <Paragraph position="3"> An ontological question is in a copular sentence with a non-terminal ontological node in the predicate.</Paragraph> <Paragraph position="4"> 56. Ontological Questions: Is a man human? Is the man a plant? A factual question is one couched in the past tense, present progressive, and/or where the subject is specific (a name or a definite NP). Specific NPs are assumed to have already been introduced into the system. Our simplified definition of generic question is one which contains an inherently stative verb or a non-stative verb in the simple present combined with a non-specific subject (indefinite NP).li 57. Factual Questions: Did John buy a book? Is the man happy? Who bought the book? Generic Questions: Does a man buy a book? Does the man love horses? Ontological questions are answered by looking in the ontological database only. If an answer is not found, the response will be &quot;no&quot; if the query contained an ontological predicate (such as PLANT or ANIMAL) because the ontology is a closed world.</Paragraph> <Paragraph position="5"> 58. Text: John is a man who bought a book.</Paragraph> <Paragraph position="6"> Is a plant living?--Yes.</Paragraph> <Paragraph position="7"> Is the man an animal?--Yes.</Paragraph> <Paragraph position="8"> Is the man a plant?----No.</Paragraph> <Paragraph position="9"> Factual queries (non-generic questions) go to the textual database first. If an answer is not found, then the generic knowledge is consulted. If an answer is found there, the appropriate response (Typically.., Inherently..) is returned. Otherwise the response is, &quot;I don't know.&quot; 59. Text: John is a man who bought a book.</Paragraph> <Paragraph position="10"> Did John buy a book?---Yes.</Paragraph> <Paragraph position="11"> Who bought a book?--John.</Paragraph> <Paragraph position="12"> Is John the President?mI don't know.</Paragraph> <Paragraph position="13"> Does John wear pants?---Typically so.</Paragraph> <Paragraph position="14"> Where did John buy the book?--Typically, in a store.</Paragraph> <Paragraph position="15"> Generic: queries go only to the generic database. 60. Does a man wear pants?--Typically so.</Paragraph> <Paragraph position="16"> What is the function of a lawyer?--Inherently, represents clients.</Paragraph> <Paragraph position="17"> Is an apple red? Typically so.</Paragraph> <Paragraph position="18"> Does a man love flowers?--I don't know.</Paragraph> <Paragraph position="19"> Where does a man buy a book? Typically, in a store.</Paragraph> <Paragraph position="20"> Who buys a book?---Inherently, a sentient.</Paragraph> <Paragraph position="21"> In addition to these general rules, there is special handling for certain types of queries. Questions of the form &quot;Who is . . .&quot; and &quot;What is . . .&quot; are answered by finding&quot; every predicate in any data base that is true of the questioned entity. For example, in any of the examples above, the question &quot;Who is John?&quot; would trigger a response that includes a list of all the nodes in the ontology which dominate the position where &quot;John&quot; is attached plus the information that John is a man, is tall, and bought a book. The question &quot;What is a vehicle?&quot; would trigger only the list of ontological nodes because there is no specific vehicle in the domain of this example. Questions such as &quot;Who buys a book?&quot; and &quot;What is driven?&quot; are answered by stating selectional restrictions on the verb--Inherently, a sentient buys a book, and Inherently, a vehicle is driven. Finally, it is possible to override the generic information in specific cases while still retaining the capability of accessing the generic information later, as the following example shows.</Paragraph> <Paragraph position="22"> 164 Computational Linguistics, Volume 15, Number 3, September 1989 Kathleen Dahlgren, Joyce McDowell, and Edward P. Stabler, Jr. Knowledge Representation for Commonsense Reasoning with Text 61. What color is an airplane?mTypically, white.</Paragraph> <Paragraph position="23"> John bought an airplane.--OK.</Paragraph> <Paragraph position="24"> What color is the airplane. - The text does not say, but typically white.</Paragraph> <Paragraph position="25"> The airplane is red.roOK.</Paragraph> <Paragraph position="26"> What color is the airplane?~The text says red.</Paragraph> <Paragraph position="27"> What color is an airplane?--Typically, white.</Paragraph> <Paragraph position="28"> Thus the system uses default logic (Reiter 1980). REASONER, therefore, is sensitive to a number of factors which make the system seem to understand the queries in a natural way. The responses generated also reflect the continuum of reliability of information which is available to a human reasoner. A flow chart of the search strategies in REASONER is shown in Figure 3.</Paragraph> </Section> <Section position="21" start_page="154" end_page="154" type="metho"> <SectionTitle> 5 NAIVE SEMANTICS AND DISCOURSE PHENOMENA 12 </SectionTitle> <Paragraph position="0"> Most computational treatments of discourse phenomena acknowledge the role of world knowledge in anaphora resolution, temporal reasoning, and causal reasoning (Reichman 1985; Grosz and Sidner 1986; Wada and Asher 1986). However, in the past the only method for encoding and incorporating world knowledge involved writing a detailed script for every real-life situation, directly encoding the probable sequence of events, participants, and so forth (Schank and Abelson 1977). This section will demonstrate that word level naive semantics offers a principled, transportable alternative to scripts. NS is a powerful source of information in discourse reasoning. Along with syntactic, compositional semantic, and discourse cue information, NS can be used to reason heuristically about discourse and drive many of the inferences drawn by people when they read a discourse. The role of syntax and compositional semantics will be underplayed in what follows, only because these contributions have been thoroughly treated by others (Reinhart 1982; Asher and Wada 1988; Kamp 1981; Grosz and Sidner 1986; Reichman 1985; Webber 1985).</Paragraph> <Section position="1" start_page="154" end_page="154" type="sub_section"> <SectionTitle> 5.1 ANAPHORA </SectionTitle> <Paragraph position="0"> In anaphora resolution, syntactic constraints, accessibility (in the sense of Kamp 1981), and discourse segmentation work in concert to limit the number of antecedents available to an anaphoric pronoun or definite NP. However, it is clear that the resultant saliency stack can end up with more than one member (Asher and Wada 1988). It is at this point in the reasoning that naive inference is required. Consider the following.</Paragraph> <Paragraph position="1"> 62. Levine's friend is a lawyer. He won his case for him.</Paragraph> <Paragraph position="2"> Syntactic rules exclude a reading in which &quot;he&quot; and &quot;him&quot; co-refer. Since both Levine and the lawyer are possible antecedents of all of the pronouns and therefore are both in the saliency stack, the following readings remain: 63. i. Levine won Levine's case for the lawyer.</Paragraph> <Paragraph position="3"> ii. Levine won the lawyer's case for the lawyer.</Paragraph> <Paragraph position="4"> iii. The lawyer won Levine's case for Levine.</Paragraph> <Paragraph position="5"> iv. The lawyer won the lawyer's case for Levine.</Paragraph> <Paragraph position="6"> Reading (iii) is the one people select. Generic knowledge of &quot;lawyer&quot; suffices to make this selection. The generic representation of &quot;lawyer&quot; in (9) includes information that lawyers argue cases. An inspection of generic knowledge for &quot;argue&quot; reveals that one goal of arguing is winning. Using these two facts, the system can infer that the lawyer won, so the lawyer is the subject of the sentence. Thus the benefactee, by disjointness, must be Levine. Now the feature of &quot;lawyer&quot; that says that lawyers represent their clients could be used to figure that the case is Levine's.</Paragraph> <Paragraph position="7"> Turning to definite anaphora, a definite description often has an implicit antecedent in the discourse, rather than an explicit or deictic one. Unless world knowledge is used, it is impossible to recover these. Consider the following.</Paragraph> <Paragraph position="8"> 64. Levine's trial was short. The testimony took only one day.</Paragraph> <Paragraph position="9"> A generic representation of &quot;trial&quot; is shown below. Using it, the antecedent of &quot;testimony&quot; in (65) can be identified as a part of the trial.</Paragraph> <Paragraph position="10"> 65. trial-Typically, in a trial first there are oaths, state165 Computational Linguistics, Volume 15, Number 3, September 1989 Kathleen Dahlgren, Joyce McDowell, and Edward P. Stabler, Jr. Knowledge Representation for Comrnonsense Reasoning with Text ments and testimony, then the judge instructs the jury, the jury deliberates and announces a verdict, and the judge pronounces the sentence.</Paragraph> <Paragraph position="11"> There are roles of witnesses and spectators. A trial lasts about three days.</Paragraph> <Paragraph position="12"> Inherently, a trial has roles of judge, clerk, bailiff, attorneys, plaintiff, and defendant. The goal of a trial is to settle a dispute or determine the guilt or innocence of a defendant.</Paragraph> <Paragraph position="13"> An interesting subset of definite anaphora is definite event anaphora (Asher 1988). In (66), we know that &quot;decision&quot; refers to an event, because it is a deverbal nominal. Generic knowledge for deverbal nominals is the same as for verbs. So we know that there is some definite SOCIAL MENTAL event. Both commenting and sentencing are SOCIAL MENTAL verbs. Generic knowledge of the verb &quot;sentence&quot; includes knowledge that a sentencing is enabled by a decision. This can be used to infer that the antecedent of &quot;the decision&quot; is the sentencing event reported in the first sentence, rather than the commenting event. Thus e 4 = e I.</Paragraph> <Paragraph position="14"> 66. (el) The judge sentenced Levine to a short term.</Paragraph> <Paragraph position="15"> (e2) He commented that Levine had been cooperative. null (e3) The decision (e4) surprised an attorney.</Paragraph> <Paragraph position="16"> If the generic knowledge has an implication which is similar to the event nominal, the correct antecedent can be inferred. The resulting DRS is shown in (67).</Paragraph> <Paragraph position="17"> 67.</Paragraph> <Paragraph position="18"> and speech time (now). In a sequence of simple past tense clauses, the reference time is updated each time. So in text (66), the temporal equations look as follows. 68. rl < now.</Paragraph> <Paragraph position="20"> However, the reference time is not always updated by a simple past tense verb. In addition to the effects of clause aspect, which will be discussed below, common-sense knowledge affects temporal reasoning. Two events reported in a sequence of simple past tense clauses can overlap in time, or the second event (in textual sequence) can occur before the first. Naive semantic representations are sufficient to assign these relations correctly in many cases. The typical implications of events in verb representations can be used to infer overlap. Consider the following discourse.</Paragraph> <Paragraph position="21"> 69. (el) Levine made a statement. (e2) He said he was sorry.</Paragraph> <Paragraph position="22"> We know that a &quot;statement&quot; is a deverbal nominal of &quot;state,&quot; and that the goal of &quot;state&quot; and of also &quot;say&quot; is communicating ideas. This knowledge is used to infer that e 2 C e 1.</Paragraph> <Paragraph position="23"> The regular temporal presuppositions of implicational features on verbs are very powerful in inferring the relative order of events. An event e i must occur before an event ej in order to cause ej.</Paragraph> <Paragraph position="24"> Xl, x2, a 1, e I, x 3, p, e 4, x 3, a2, e2, e3</Paragraph> <Paragraph position="26"/> </Section> <Section position="2" start_page="154" end_page="154" type="sub_section"> <SectionTitle> 5.2 TEMPORAL REASONING </SectionTitle> <Paragraph position="0"> Temporal reasoning involves the assignment of relationships among the times of the events reported in the discourse. Following the Reichenbachian approach to tense found in Partee (1984), there are three elements to temporal reasoning: event time (el), reference time (ri), generic: features of verbs. This works well when one of the implications of a verb mentions another verb (or related verb) in the text, and tense or adverbial modifiers do not give clues as to the temporal relations. In (70), the fact that e2 is a typical cause of el, can be used to infer that e 2 occurred before e l.</Paragraph> <Paragraph position="1"> 70. (ea) Levine was found guilty. (e2) He broke the law.</Paragraph> <Paragraph position="2"> Similarly, in (71), the fact that buying typically takes place in stores, can be used to infer that e 2 overlapped el in time. (The stativity of the verb in the second sentence is also a weak, but insufficient indicator of the temporal relationship between the sentences).</Paragraph> <Paragraph position="3"> Computational Linguistics, Volume 15, Number 3, September 1989 166 Kathleen Dahlgren, Joyce McDowell, and Edward P. Stabler, Jr. Knowledge Representation for Commonsense Reasoning with Text 71. (el) Levine bought a book. (e2) He was in a store on 5th Ave.</Paragraph> <Paragraph position="4"> Temporal properties of nouns also require naive semantics. Role terms have a temporal element (Enc 1987). A term such as &quot;President of the U.S.&quot; has a DURA-TION feature in the naive semantic representation. This enables appropriate inferences concerning the occupants of roles. In (72), the system can infer that the individual who declared war on insider trading is probably not the same as the one whose office indicted Levine.</Paragraph> <Paragraph position="5"> 72. The Attorney General declared war on insider trading in 1964. The Attorney General's office indicted Levine six months ago.</Paragraph> <Paragraph position="6"> Another way in which naive semantic representations can be used in temporal reasoning has to do with the assignment of tense to deverbal nominals. In (73), the nominal &quot;violations&quot; refers to events which can be placed in time before the sentencing. Naive semantic knowledge can be used to infer that the violations took place before the sentencing, because violations are illegal, sentencing is the fnal stage of a trial, a trial determines whether the defendant is guilty, and guilty parties have done something illegal.</Paragraph> <Paragraph position="7"> 73. Levine, who engaged in massive insider trading, was sentenced to two years in prison. His violations of the securities laws were shocking.</Paragraph> <Paragraph position="8"> Another way that naive semantics helps temporal reasoning, comes from the assignment of coherence relations using naive semantics. How this can be done is discussed below (Section 5.4). Given certain rhetorical relations between two clauses, certain temporal relationships are indicated. Table 11 lists the temporal predications of certain coherence relations. The first group are relations which hold between events or states 81 and 62, where 81 occurs before 62. In (74) the system can recognize that e I is before e2, once it assigns the relation Cause between them.</Paragraph> <Paragraph position="9"> 74. (e0 Levine broke the law. (e2) He was indicted.</Paragraph> <Paragraph position="10"> The same thing would work if the sentences were in reverse order in the text. The second group in Table 1 1 lists the coherence relations that indicate that events or states overlap in time. The third group lists those relations that require that the source event or state is in the speech time (now).</Paragraph> </Section> <Section position="3" start_page="154" end_page="154" type="sub_section"> <SectionTitle> 5.3 CLAUSE ASPECT </SectionTitle> <Paragraph position="0"> Clause aspect refers to the aspect of an entire clause.</Paragraph> <Paragraph position="1"> The classical example in (75) through (77) illustrates differing clause aspects with the same verb and tense.</Paragraph> <Paragraph position="2"> (75) is telic, while (76) and (77) are activity clauses.</Paragraph> <Paragraph position="3"> 75. John pushed the cart under a shed.</Paragraph> <Paragraph position="4"> 76. John pushed the cart under adverse conditions.</Paragraph> <Paragraph position="5"> 77. John pushed the cart.</Paragraph> <Paragraph position="6"> Ciause-stative means not only the opposition between an inherently stative verb and an eventive verb (as in &quot;be&quot; vs. &quot;hit&quot;), but to the various ways in which a whole 8 1 and 62---event or state reference markers r~ < rE--reference times I. t~ 1 C rl, 62 C r 2 Elaboration(Si,62), Cause(81,62), clause can end up being stative, as with the presence of the progressive, or a number of other factors. Clauserelic means that the clause reports a change of state with a terminus. &quot;John built the house&quot; is clause-telic, while &quot;John was building the house&quot; is clause-stative. A telic clause has an ACHIEVEMENT or ACCOMPLISHMENT verb not in the progressive, not in the simple present (which would be habitual, and with no modal (e.g., &quot;John will build the house&quot; is not telic). Clauseactivity has to do with a clause which reports an event which has no terminus, and which has the sub-interval property (Bennett and Partee 1978), as &quot;John ran.&quot; Naive semantic knowledge is used to assign clause aspect. In (75), naive semantic knowledge that a shed is a PLACE, and generic information of the relative sizes of carts and sheds, can be used to infer that the shed was a destination for the cart. An ACTIVITY verb such as &quot;push,&quot; with a destination argument in the verb phrase results in a telic clause. This inference for (75) would not hold for (76). Similarly, an ACHIEVEMENT or ACCOMPLISHMENT verb indicates a TELIC clause, but other arguments can change them to ACTIVITY (cf. Moens and Steedman 1987). For example, in (78) the clause is TELIC, while in (79) it is ACTIV-ITY, and in (80) it is ambiguous between TELIC and ACTIVITY.</Paragraph> <Paragraph position="7"> 78. The prosecutor questioned the point, 79. The prosecutor questioned the witness for an hour.</Paragraph> <Paragraph position="8"> 80. The prosecutor questioned the witness.</Paragraph> </Section> </Section> <Section position="22" start_page="154" end_page="154" type="metho"> <SectionTitle> S.4 COHERENCE RELATIONS </SectionTitle> <Paragraph position="0"> Coherence relations are handled in the KT system as added predicates in a cognitive DRS. In the DRS representing two clauses connected by a discourse cue word such as &quot;because,&quot; a predicate cause(el,e2) is represented. Similarly, where the first event typically causes the second, the system guesses the cause relation between the two event reference markers, and a Computational Linguistics, Volume 15, Number 3, September 1989 167 Kathleen Dahlgren, Joyce McDowell, and Edward P. Stabler, Jr. Knowledge Representation for Commonsense Reasoning with Text cause predicate is introduced into the DRS, resulting in a cognitive (or inferred) DRS. Coherence relations are assigned using syntax, temporal relations, clause aspect, discourse cues, and naive semantics. An algorithm for coherence relation assignment has been developed (Dahlgren 1988c, 1989). This section will illustrate only the contribution of naive semantics and will not delve into the complex problem of the interactions among the several sources of information. Grosz and Sidner (1986) argue that coherence relations are not a useful analytical tool because no clear, closed set of them has been discovered. However, there is ample psycholinguistic evidence that in constructing the interpretation of a text, and in recalling what it said, coherence relations are inferred and used by readers (Rickheit and Strohner 1985). In terms of computational linguistics, coherence relations are useful for text summarization and relevance reasoning. In text summarization, only the general actions, not the elaborations, can be included in the summary. Descriptive and other background clauses can be ignored. Similarly, relevance can be inferred from the causal implications of events reported in a text.</Paragraph> <Paragraph position="1"> If a reader says that he or she wants to read about events that affect the construction industry, for example, and the typical consequence of some event in a text affects the construction industry, then that reader is interested in that text.</Paragraph> <Paragraph position="2"> The naive semantic representations of nouns and verbs contain sufficient information to handle a large number of cases in which world knowledge is required to structure the discourse. Generic representations of the typical implications of verbs such as cause, goal, enablement, and consequence are the very same information as coherence relations. Their content means, &quot;If there was an event (or state) of VERBling, then it probably had as goal a later event (or state) of VERB2ing.&quot; For example, &quot;If there was an event of buying, it probably had as goal a state of owning.&quot; Naive semantic representations contain generalizations about objects and actions which are common to a linguistic community, and thus explain the ability to understand a discourse without resort to particular scripts describing familiar real-world situations.</Paragraph> <Paragraph position="3"> Using generic and ontological representations derived from psycholinguistic data, coherence relations can be assigned. To infer goal(el,e2) for text (81), knowledge that &quot;profit&quot; is money can be used to relate (e2) to the the goal feature of &quot;invest.&quot; 81. (e0 John invested heavily. (e2) He made a huge profit.</Paragraph> <Paragraph position="4"> Some of the naive semantic representation of &quot;invest&quot; is shown below: 82. Investing is typically lucrative and is accomplished with money. Inherently, sentients do the investing with the goal of making money.</Paragraph> <Paragraph position="5"> Similarly, in (83), the generic entry for &quot;insider trading&quot; can be used to infer that Levine broke the law. The entry associated with reading 1 of the verb &quot;charge,&quot; as shown in Table 9, includes information that a typical cause of charging someone is that that someone has committed a crime. Putting these two together, cause(el,e2) can be inferred.</Paragraph> <Paragraph position="6"> 83. (e0 Levine engaged in insider trading.</Paragraph> <Paragraph position="7"> (e2) The government charged him with violations of the securities laws.</Paragraph> <Paragraph position="8"> The segmentation of discourse takes into account paragraphing, discourse cues such as Turning to .... In summary, clause aspect, temporal relations, and coherence relations. In this section we will briefly illustrate that nai, ve semantics is one source of information in discourse segmentation. In narrative, a clear distinction can often be made between segments consisting of sequences of actions that are the foreground of the narrative, and segments that provide the background or setting for the action. If the author does not give clear discourse cues of the switch to a setting or situation segment, the shift can be inferred using naive semantics. In our method, discourse segments are related to each other the same way as clauses (as in Mann and Thompson \[1987\] and Hobbs \[1985\]), so the relationship here is one of situation_activity(Seg2,Segl) where Segl is a sequence of actions. Consider (84).</Paragraph> <Paragraph position="9"> 84. Levine engaged in insider trading at his firm. He was charged and found guilty of violations of the, securities laws. He was sentenced by Judge GoetteL Levine was happy at his firm. The audience waited with baited breath to hear what Judge Goettel would say.</Paragraph> <Paragraph position="10"> In the text, there is a change of segment at &quot;Levine was happy at his firm.&quot; The segment is a SITUATION-ACTIVITY segment. It describes what was going on when Levine was engaging in illegal practices. The change from a sequence of actions to a background segment is indicated by several factors, including the use of the stative and the place adverbial. Another factor is naive semantic knowledge of &quot;firm.&quot; Working takes place at a firm, and this knowledge can be used to infer that &quot;Levine was happy at his firm&quot; refers to a long-term situation in which Levine was working. In that situation, he was happy. Thus the sentence is not about some specific action, but is a generalization about Levine's condition as a worker. Such a generalization indicates a change of segment from a sequence of actions to a SITUATION_ACTIVITY segment.</Paragraph> </Section> class="xml-element"></Paper>