File Information

File: 05-lr/acl_arc_1_sum/cleansed_text/xml_by_section/metho/79/j79-1070_metho.xml

Size: 105,015 bytes

Last Modified: 2025-10-06 14:11:16

<?xml version="1.0" standalone="yes"?>
<Paper uid="J79-1070">
  <Title>American Journal of Cornput ational Linguistics TEXT UiiDERSTANDING: A SURVEY</Title>
  <Section position="3" start_page="0" end_page="0" type="metho">
    <SectionTitle>
1.0 THE CONTENT OF CONNECTED DISCOURSE
</SectionTitle>
    <Paragraph position="0"> This scctfon will present a brief survey of the recent work of some experimental psychologist,^ concerned with thc process of understanding corlncctcd discourse.</Paragraph>
    <Paragraph position="1"> Thc details of their cxperi~n~nts will not be presented, but the types of cxperirneals and general conclusions will be discussed, as well as any madcls proposed by the investigators. Th~~se results seem to cotlve rge toward a single model, in spite of apparent cont radictions, Psychological work on text understanding is relevant to computational models because humans are still the only good tcxt undcrstnndcrs. Also, although some aspects of psychological research, such as forgetting, are not currently included in any serious way in computational models, it would appear that such models have a very useful role to play in pointing out where significant details of the process of text understanding are being handled intuitively by the theorist, and have not really been defined in the tlleory,</Paragraph>
    <Section position="1" start_page="0" end_page="0" type="sub_section">
      <SectionTitle>
1.1 The Nature Of Discourse Content
</SectionTitle>
      <Paragraph position="0"> The nature (general characteristics and features) of the meaning of connected text will be examined first. The principle focus of these studies is to compare the understanding of a text to the actual text and attempt to charnctefize the types of differences that are found.</Paragraph>
      <Paragraph position="1">  I\,~rc*l,ty, Bra~~sford and Franks in n large body of work (Bransford, Rarclay irnd l'rrwks [72], l3nrcl~y [73], Bransford and Franks (731, Branrrforrl atld Ptt*('*n 1,-11 1741 nnd Frn~~ke and Rraneford [74]) argue for the cxletence of II t t c rcy~rcscnt nt ions indepcndcnt of the surface input, and propose active 11 1 t'haiCS 011(~11t it18 on thc input. They describe eontencea as Information whl(\r is ttscld by tlrc undcrstnnder to construct a descript-ion of a situation. F n t11oy clnim tl\r~t the itrformnt ion used to construct the selnantic dt~*~-lpt [ot~ is not wholly contained in the sentences, but that an underatander  lr is ~rrtsvious knowlcdgc. to a great extent. They do not make any definite 01 s r r 11 tlrc Form of the setnant ic representat ions. Typical cxl~l1~tt~~wt:: pcrformtbd to test thcsc hypotheses consist of recognition or r I 1 QP I-ompoi~nci rtrntt\ncea 1 ikc: I. Tllr~c turt lcs rastcd hoaidc u floating log, and a fish swam beneath t 111-in.</Paragraph>
      <Paragraph position="2"> 3. 'l'1rrt.c turtlm r~8t~d on a floating log, and a fish swam beneath ttiem.</Paragraph>
      <Paragraph position="4"> r;\~ll jr'ct rz i~r;r\cl thc2 t spntinl kt.rowlcdgc to crclate situatiot~ dcscriprions, and when two different sentences produce the same description, it is difficult to distinguish them.</Paragraph>
      <Paragraph position="5"> Similar investigations used a sequence of sentences, each comparing two objects from a set of five according to some dimension (e.g. positjon, height, speed, weight). Subjects who knew that a single situation was being described had a better memory for the situation, but were less able to distinguish sentences they had heard from those implied by the described situation. Once again, the conclusion is that a non-linguistic representation is created when a meaningful text is understood.</Paragraph>
      <Paragraph position="6"> A final group of studies involved descriptive texts that were ambiguous or difficult to comprehend without an appropriate context. The context might be a picture of the described situation or a meaningful title for the text. In every case the results indicated the understander attempted to build a description of the whole situation, and when this was very difficult or impossible, comp rehension was low. (Related studies by Anderson, Reynolds, Schallert and Coetz [76] and Scallert [76] utilize texts, each with two totally different meanings. Evidence is obtained that one or the other alternative is almost always chosen for the entire text. The effect of subjects backgrounds on interpretations of these ambiguous texts Is also Gtudied, and a correlation between background and chosen interpretation is claimed. ) The principal conclusion of these studies is that the process of understanding text involves constructing a consistent unified meaning, wlli ch is not simply the set of individual sentenee meanings. It diffcrs in thht Page 8 additional knowledge and orga11ir.n t i 011 i .; Int rod~iccd by tire uilderstander.</Paragraph>
    </Section>
  </Section>
  <Section position="4" start_page="0" end_page="1" type="metho">
    <SectionTitle>
1 Frcderikscn -
</SectionTitle>
    <Paragraph position="0"> Frederikscn [75p, 7Sb1 contin~~cs this line of thought, treating the undcrqtanding of n disccl\irsc as the. scmarlt i c knowlcdgs that is acquired in listening to the disco\rrsc. Thi r; knowlc.~l~~,~~ f o r~cqutrcd by processes that utilize prior knowlcdgc, cont cxt , c t c. nnd includcs inf armat ion which is inferred as well AS that explicitly prcs~b~~tctl. Fredcuiksen also argues thaf: the processes involved in undcrstnndinp, a discourse are direqtly related to the process limitntior~s of thc PL'OCCSSO~. The two processes that he discusses are overgcnera1i;cat ion and infert&gt;~~cr. Ovc.rgcncirnlizetion is the discarding of detail informat ion resulting in n more p,th1'icr~1 concept. 'tie suggests that overgenera1,ization reduccs the amcwnt of infornlotion to be understood, thus reducing the processing Zond on the UII~C ~-stni\de Inferred information is infannation which is assunlcd to hc. t even if it- is not explicitly present.</Paragraph>
    <Paragraph position="1"> This process is clhimcd to reduce the pruccssing load by eliminating the necessity of completely andc rst lrndiny, t\vtary sctlteilce. FredarAksen performed neutral recall cxpcrimcnts nnd roct~ll cxl~rrin~~nts in wh.lch the context - the task assigwcd to the subjc.ct prlor to pli-:;c~itntion of the story - was changed,.</Paragraph>
    <Paragraph position="2"> His conclusfons arc that. ovc rgcncr;tliz,~r 1011 rind inferences are incorporated into the semantic reprcsi-ntnticrl~ of tl~cb story as it is ut~derstood, and that the context can inf lu~nccl tl~c nii)ou~~t of infc rcnciny, that is done during Page 9 Frederiksen uses a simple, inPS ormally def incd semantic rep resentation in the above studies. It consists of set Inclusion, identity, and logical implication relationships that hold among the concepts of the text,. He indicates the need for&amp; detailed model of the sciwntic representation, and in other work, Frederiksen [75cJ begin's to clef ine such a rep resentational scheme. He propcses the use of two networks: a semantic network and a logical network. The semantfc network comains the rep resent at ion of individual propositions while the logical network is composed of relations that hold between propositions that exist in the semantic network. ~rederiksen's system is probably the most elaborate yet proposed using the basic ideas of case grammar represented by semantic networks. He begins with the fundamental ideas of object and action hierarchies, although his are quite large end contain numerous distinctions. He proposes a system that distinguishes sixteen verb, cases, and a large numhc r of relations specifying states, quantification locati~n, manner, time, order, proximity, tense and aspect.</Paragraph>
    <Paragraph position="3"> His logical network consists of relations drawn from propositional logic plus causality, and he proposes several modal operators. The complexity of this proposal gives the impression of representational power, but it seems that Frederiksen is still primarily oriented toward the repreqentation of  indiyidual propositians. Although he shows the representation of simple time and causally ordered actions sequences, he has not yet demonstrated the adequacy of his system for text, in gerlerdl. And, he has utilized a number of elements of quantified modal logic without demonstrating their usefulness for modelling human understanding of d'iscoursc. His discussions do raise a number of interesting questions about representation, and certainly deserve Page 10 consideration in the development of any set of semantic types and relat~ons.</Paragraph>
  </Section>
  <Section position="5" start_page="1" end_page="1" type="metho">
    <SectionTitle>
113 Kintsch -
</SectionTitle>
    <Paragraph position="0"> Kintsch [74] C reviewed by van Di jk [75bJ ) attempts to set forth a fairly complete theory of langudge understanding. Kintsch's basic rep resentational unit is the proposition, and he includes discus&amp;ion of the usual difficult problems of definiteness (including generic versus specific distinct ions) of noun phrases, quantification, modality, implicetian and presupposition, location, time and tense. He proposes a text base which underlies discourses, but his revised and elaborated thought on this is discussed in Section 1.2.</Paragraph>
    <Paragraph position="1"> ,Me, describes a model of discourse processing in which he makes a distinction between episodic and semantic memory, argues that the processing operations must be well defined (he suggests pattern matching and completion, abstraction and generation) and argues that recall is essentially a different pmcess from recognition in that recall requires input organization while recognition does not, Kintsch reports a number of experiments undertaken to test aspects of h;i$ theory. One conclusion that he reaches is that the semantic representation is partially independent of the actual input sentences. One kind of experiment performed to test this was the performance of a cammsn task by subjects who had read substantially the same material but in forms ot varying comnlexity.</Paragraph>
    <Paragraph position="2"> In other experiments he notes that sentences containing multiple propositions are much less likely to be recalled (completnly) than are single propositibn sentences. This also supports the idea of an undctlying representaticn. A Page 11 related conclusion is that the agent of a proposition is the most likely case to be recalled. However, Kintsch did find that in passive sentences, the subject, rather than the agent, was the most likely to be kecalled. A final conclusion conce ms the presence of inferred inPS ormation in memory. Kintsch found that immediate recall showed some difference between implicit and explicit propositions, but that after twenty minutes or more the difference had disappeared and the two were indistinguishable to the subjeet. An interesting side result was the distinct difference in reading and recognition times for argumentative versus descriptive discourse. The argumentative discourse required significantly more time in both tasks.. Kintsch also includes reaction time experiments to study proposition retrieval, deterruin-ation of the truth or falsity of general propositions and processing of complex lexical items but these will not be discussed.</Paragraph>
  </Section>
  <Section position="6" start_page="1" end_page="1" type="metho">
    <SectionTitle>
1.1.4 Thorndyke -
</SectionTitle>
    <Paragraph position="0"> Additional investigationb on the role of inference in understanding text have been carried out by Thorndyke [76]. He uses compound sentences asserting a causal connection which is not familiar or obvious. For example: The hamburger chain owner was aPS raid that his love fox f rench fries would ruin his marriage.</Paragraph>
    <Paragraph position="1"> Sentences like these are imbedded in a meaningful text, and are followed later in the teFt by a continuation sentence which references the previously mentioned relationship.</Paragraph>
    <Paragraph position="2"> This continuation sentence might be neutral like Page 12 He decided to see a marriage counselor in order to save his marriage. or it might encourage one particular eqlanatcrry inference like He decided to join weight watchers in order to save his marri ~ge. which strongly suggests the inference He is fat from eating Erench fries.</Paragraph>
    <Paragraph position="3"> Thorndyke uses recognition tests after the presentation of the story to compare inferences that have been reinforced by a continuation sentence with neutral and inapp rop riate inferences. He found the reinforced inferences much more likely to be recognized as part of She text than the neutral inferences, while recognition of inappropriate inf erenc~s was very unlikely. Thorndyke suggests that although this evidence indicates that inferences are made and stored as part of the understanding of a text, a more important implicaOion exrsts. This is that the role of inferencl ng is to aid in the 4ntegretiot.r of new information into tble larger framewi3tk of the understanding o,f a text when no appropriate understanding could be obtained from only the explicitly stated information.</Paragraph>
    <Section position="1" start_page="1" end_page="1" type="sub_section">
      <SectionTitle>
1.2 The Structure Of Discourse Content
</SectionTitle>
      <Paragraph position="0"> Attempts to characterize the content of text necessarily lead to a structuring of that content. Simple one-dimensional representations in which propositions are connected only by co-reference or by time or causal ordering Page 13 are adequate for only a restricted class of texts.</Paragraph>
      <Paragraph position="1"> The following studier all p ropose some type of multi-level representation of content. It seems appropriate to begin a discussion of memory structure with Bartlett [32] since he is often referenced as the originator of several cu rrently popular hypotheses. It should be remembered that Bart1 et t was concerned with memory in general and the phenomena associated with recall, and did, not restrict himself to the study of disc~urse. His pxinciple conclusions were that memories are not stored in isolated, static units, and that exact recall is very rare. \In fact, he often found cases of gross distortions in the recall of his subjects. He suggests that instead, memory is composed of a number of active, organized masses of past reactions and experiences, which he designates schemata, and a small amount of outstanding detail- Remembering is seen as a constructive process strongly affected by memories other than the one being retrieved. His ideas of organization beyond the sentence, of storage of something other than actual input sentences, and of active processes that modify text prior to its reproduction in recall are all currently en joying wjde acceptance.</Paragraph>
      <Paragraph position="3"> Crotbhers 1721 is concerned with the recall of short, expository paragraphs containin!:. material not likely to be familiar to his subjects, He presents results obtained wi tll paragraphs about hebulae. He proposes the existence of a semantic representation underlying any particular discourse which contains the rncal~lng of the discourge, but, does not reflect the details of the surface form of the teyt. Thus, a single semantic represent~tion might underlie numerbus actual discourses. His semantic representation assumes a conceptual taxonomy showing the relationship of each known concept to its superordinate concept (i.e. the familiar semantic hierarchy), but he does not provide details on this. He also proposes an additkonal set of hierarchies showing the relationships of the concepts in a particular discourse. For example, the following: means that a nebula is either seen or it is not seen, where all of the concepts are defined on the conceptual graph. Crothers does not carefully define his notation, so it is ~ot clear exactly what may or may not occur at a node, Primarily, concepts (words) or connectives are used (e.g. NEBULA. IS AND, OR, WHY). Since hc deals only with expository, descriptive material he finds no need to represent actions or time, Crothers performed recall experiments on two different versions of the same material, and arrives at three conclusibns. First, he concludes that the surface paragraph *s not a Page 15 -fieant factor in recall, thus arguing for a surface-independent semantic repnsentation for the paragraphs. His- last two conclusions are negative xejecting hypotheses he had previously suggested.</Paragraph>
      <Paragraph position="4"> The first of these is that superordinate nodes in the meaning of the paragraph will be recalled more frequently than subordinate nodes. The second rejected hypothesis is that iaforaation not directly connected to the most superordinate pode of the paragraph I&amp; lese likely to be recalled than information that is connected. Since his results do not support either of these hypotheses, Crothers suggests tbat other variablea, such as frequency of concepts, are probably also of irportance.</Paragraph>
      <Paragraph position="5">  Grimes [75] has proposed an extended case grammr supplemented with what he calls rhetorical predicates, which are higher-order predicates that take other propositions as their arguments. He subdivides rhetorical predicates into three groups: paratactic, hypotactic and neutral, Parataet ic predicates always take arguments of equal weight (i. e. no argument is subordinate to any other argument). Hypotactic predicates have one dominant argument, to which the others are ,subordinate. Neutral predicates may be used as either paratactic or hypotactic predicates. The following is a list of some of the rhetorical predicates proposed by Grimes, and their basic meaning:</Paragraph>
    </Section>
  </Section>
  <Section position="7" start_page="1" end_page="1" type="metho">
    <SectionTitle>
COLLECTION -
</SectionTitle>
    <Paragraph position="0"> Options, or Question and answer, problem and solution Gives qualities of the dominant proposition Gives restatement of the domiriafit proposition Gives more specific infomation about a general dominant proposition Gives an abstract ekplanation for the specific dominant proposition Gives an analogy to support the dominant proposition List of elements related in some unspecified way Causality, an antecedent and consequent Meyer [75] uses Grimes' rhetorical predicates and most of his casegrammar to create analyses of paragraphs she uses in recall experiments. For example: Page 17</Paragraph>
  </Section>
  <Section position="8" start_page="1" end_page="1" type="metho">
    <SectionTitle>
RE SP ON-S E
PROBLEM
COLLECTION
</SectionTitle>
    <Paragraph position="0"> need to gene rate elect ri c power p t~tect envi ronment rational utilization of natural resources</Paragraph>
  </Section>
  <Section position="9" start_page="1" end_page="1" type="metho">
    <SectionTitle>
SOLUTION
</SectionTitle>
    <Paragraph position="0"> breeder reactors would be the essentials of an analysis of a paragraph which stated that the need to generate electric power while protecting the environment and rationally utilizing natural resources is a problem which breeder reactors solves. Meyer indicates the hierarchical structure through the use of indentation. Since the rhetorical predicates in the above example were 911 paratactic, nb indentation was used. However, the following example: finite reserves of natural resources</Paragraph>
  </Section>
  <Section position="10" start_page="1" end_page="1" type="metho">
    <SectionTitle>
SPECIFIC
COLLECTION
</SectionTitle>
    <Paragraph position="0"> coal oil gas shows the subordination of the specific set of resources to the generaJ idea. eyer's analyses of expository paragraphs using this scheme always follow the author's organizational st rvcture and always result in a purely hierarchical structure like those depicted above Meyer also recognizes class of sentences that can occur in expository paragraphs, but do not contribute any Page 18 content to the passage. She calls these&gt; signalling because their function is. to explicktly indicate the structure of the passage. She notes four types of signallins  1. ~~ecificatibn of the structure (e. g. &amp;quot;Two options exist.'') 2. Prematurely revealed information abstracted from the remainder of the passage (em go &amp;quot;The alternatives are solar energy, nuclear energy, geothermal energy aad laser fusion energy. &amp;quot;) 3. Summary statements t t 4. Pointer words (e.g. unfortunately&amp;quot;, &amp;quot;an impdrtant point is&amp;quot;)  Meyer conducted recall experiments varying the position of certain material in the content hierarchy of passages, and including or omitting certain s ~gnalling inPS omat ion. Her principle conclusion is that material structurally higher in the content hierarchy is remembered substantially better than identical information placed lower in the hierarchy in another passage. She also notes that passages with identical structure but totally different content exhibit very similar patterns of recall at the higher leve$s. However, at the lower levels, the pattern of recall varied, indicating content dependence. In regard to signalling, she concludes that it has very little' effect at the top level, but seems beneficial at the middle levels.</Paragraph>
    <Paragraph position="1"> Page 19 In regard to a theoretical explanation of why the highest ideas are recalled best, Meyer suggests three proposals and points out the weaknesses of each. First, lower propositions might be subsumed by hieher ones with the passage of time. Immediate recall experiments shw that the phenomenon occurs even without a time lapse, however. Secondly, at is possible that all propositions are stored but that retrieval is easier at the higher levels. She criticizes this proposal because even cued recall experiments were unable to retrieve the lower propositions. Finally, perhaps only higher ideas are ever stored. But her results show that the lose of lower level propositions begins immediately, but continues with time at a more rapid rate than for the higher level propositions. Meyer suggests that probably some combination of these processes is occurring, as well as other processes sensitive to the structure of the passage as a whole.</Paragraph>
    <Paragraph position="2"> 1.24 Kintsch And Van Dijk Kintsch and van Dl jk (van Di jk [74,7Sa,76], van Di jk and Kintsch [forthcoming] and Kintsch and van Di jk [76]) present a model for the organization of discourse as whole, and a number of (sometimes informal) experiments attempting to validate the model. Beginning at the lowest level, a discourse representation consists of a set of propositions. A distinction is made between two different types of representation, callCd text bases, as to whether all implied information is made explicit or whether it is left implicit. The notion of coherence is introduced, which is the property that distinguishes a discourse from a random set of sentences, Referential Page 20 identity has been auggqsted as a major test of coherence, but it ie not an adequate definition (see discuesion of coherence and Bellert's proposals in Section 3). Kintsch and van Di jk argue that coherence is more accuratelycaptured by the requirement that each proposition of the text base be connected with one or more preceding propositions. Propositions are connected if one is a condition for the other, with the strength of the connectfon ranging from posaible to necessary. Thus, an explicit text base is one containing all of the propositions necessary for coherence, while an implicit text base is one with some of these propositinns deleted. The deleted propositions are those that can be assumed known or which normally would be inf~rred by the unde rs tande r. Different types of discourse would have different rules governing the deletion of propositions. For example, casual conve rgation would allow more deletion thari careful argumentation. The text base is organized hierarchically under macro-s t ructures, which are higher brde r propositions. Macro-st ructures may be related to their propositional arguments by a number of macro-rules. Four of these rules are suggested. =The first, information reduction, is a rule of generalization which would explain the existence of the macro-structure John is ill, which had as its arguments propositions like John has a feve r.</Paragraph>
    <Paragraph position="3"> John has the flu.</Paragraph>
    <Paragraph position="4"> The second rule, deletion, would explain the relationship between Peter saw a ball.</Paragraph>
    <Paragraph position="5"> and its subordinate propositions Peter saw a ball.</Paragraph>
    <Paragraph position="6"> The ball wSas blue.</Paragraph>
    <Paragraph position="7"> The third rule, integration, combines a central event with its normal pre-conditions, results and component actions. Thus John went to Paris.</Paragraph>
    <Paragraph position="8"> might have as its arguments John took a cab TO the station.</Paragraph>
    <Paragraph position="9"> He bought tickets.</Paragraph>
    <Paragraph position="10"> He went to Paris.</Paragraph>
    <Paragraph position="11"> The fourth rule, construction, explains the relationship between a complex fact and its component parts. For example, the macro-structure Peter built a house, might have as its arguments the event sequence Peter laid a foundation.</Paragraph>
    <Paragraph position="12"> Peter built walls.</Paragraph>
    <Paragraph position="13"> Peter built a roof.</Paragraph>
    <Paragraph position="14"> In general, two conditions hold fur all macro-structures. First, the Page 22 macro-structure must be semant lcally implied by its micro-structure (i.e. its propositional arguments). The exact meaning of the term implication is not clearly stated by Kintsch and van Dljk, They sometimes refer to it as entailment and treat it as a formal logical relation, but at other times say that it is not logical in the strict sense. In any case, it is clear that there are semantic rules or structures that allow creation of macro-~tructures such as those above. Thc second coqdition is that the sequence of macro-st ructures representing a coherent text base must itself be coherent. Recalling the definition of koherence, this implies that no macro-structure may delete information contained in its micro-structure which fs a c~ndition for another macro-structure An important consequence of this is that the macro-structures 06. a ttext, taken by themselves, form a coherent summary of that text.</Paragraph>
    <Paragraph position="15"> A final component of a tept representation is the specification of the PSom or structure of the discourse, but discussion of this proposal will be postponed until Section 2. Their investigations are primartly with narrative text, and they call the narrative structure a schema.</Paragraph>
    <Paragraph position="16"> Kintsch and van4i jk shggcst that the concept of -narrative st-ructure coabined with the idea of macro-structures leads them to the follawing hypotheses regarding co~np rehension and recall of neriative- Ffrs t macro-structures are primarily what is stored when a text is understood. Recall uses macro-structures as a starting point in retrieval, and summaries directly reflect the nlacro-st ructures. Secondly, since macro-st ructures are essential to comprehension, they must be constructed at the time of reading. Page 23 Finally, a narrative schema is necessary for the organization of the text representation. Kintscll and van Dijk have done a number of summary, recall and other experiments to test these hypotheses. Most use a 1600 word text from - The ------..-- Decamcrot~. In recall experiments, the propositions corresponding to the macro-st ructures were found to be recalled most often, and were very unlikely to disappear in delayed recall (nine days later). Xn contrast, many other propositioi~s recalled immediately were omitted in the delayed recall, The propositions most likely to be recalled have the following functions:  1, Introduce main characters 2, Give major goals of characters 3. Describe actions leading to these goals 4. Describe events occurring to these characters leading to or from their goals while propositions having the following functions are likely to be forgotten: 1, Sctting description 2. Prepnratoq act ions 3. Mental actions 4. Comp oncnt act ions Page 24 5. Probable consequences of an act ion or an event 6. Meta-narrative statements by the author  Summarizing experiments showed thbt summaries of the story *wet.e very much like delayed recall. When a summary was written after presentation of each successive part of the story, followed immediately by a complete summary, propositions were included Fn the partial summaries that were omitted in the final summary. A flnal group of recall experiments compared recall of a 70 word paragraph in isolation, to cued recall of the same paragraph imbedded in ran 85O word text. Surprisingly, the recalls were almost identical But when the entire text was recalled, the reproduction of the particular paragraph wa8 much smaller and less accurate than in the first two cases. Their conclusions from these results axe that the recall of small amounts of text is a different process from the recall of a long text, that recall of a long text relies . on the macro-structure of the text as a means of organizing the text, that the micro-st ructure is forgotten much more easily and that summarizing is based on the macro-stcructuxe.</Paragraph>
    <Paragraph position="17"> Finally, some experiments were done in which incorrect summaries were presented prior to the presentation of the story. These were found to have practically no influence on the final understanding o'f the story, and the subjects were unable to accurately recall the incorrect summaries.</Paragraph>
    <Paragraph position="19"> Some Conside rati on's Regarding Memory Experiments Before attemptiqg to describe an informal madel based upon these hypotheses and experimental resulta, some general criticisms of memory experiments exp ressed by Spiro [ 751 deserve consideration. spire's principle argument is that recall consists of active reconstruction processes, rather than passive processes which merely reproduce that which was stored st the time of comprehension. His general position is thus much like that of Bartlett, and he Observes that almo3t all experimentets since Bartlett have failed to replicate his findings of significant errors in recall. As a result, psychologists have tended to concent rate on the process of understanding, or construction, and have treated recall as somethihg of a simple retrieval process. Spiro argues that for reconstructive errors to occur, it is necessary that the input be understood in terms of some preexisting schema. Any text that results in the creation of a new achema would not be subject to the same kind of effects of previous knowledge. The interference caused by the pre-existence of schemata is even more pronounced if other uses of the schema are made between the time of comprehension and the time of recall. Spiro also points out that any experimental subject would be unlikely to integrate material read or heard in an experiment into his general knowledge since its truth and source are unknom. This coupled with the desire to perform well, could easily result id a considerably modified process of understanding text in experimental settings.. Spiro performed a number of recall experiments which support his hypothesis. He used text about interpersonal relationships and instructed t31e subjects that the experiment was concerned with their reactions to the incidents described, thus maximally Page 26 involving their pre-existing struerurea in the understanding process. Spiro found substantial errors in the subjects recal,l~, which he explaine as the effect of interaction wi ti1 prc-cxisting structures for interpersonal relationships. (Kintsch and van Dijk [75] found similar results using stories from the Bible.) Spiro conclr~des that most text recall experiments have not shown these effects because they did not meet the required conditions, but frequently involved the prcscntation of material that was totally new and which was kept isolated by the subject. These criticisms not only are relevant to the question of construction versus reconstruction, but also to inferring discourse organf zation f rofn recall experiments, and indicate that one must be very cautious in generalizing from the results of text recall.</Paragraph>
    <Paragraph position="20"> experiments due to the many intewctions between the subject's knowledge, the experimental setting, the type and content of the text and the extra-experimental effects (or lack of them) in long term experiments.</Paragraph>
    <Section position="1" start_page="1" end_page="1" type="sub_section">
      <SectionTitle>
1.4 Discussion And Conclusions
</SectionTitle>
      <Paragraph position="0"> A great deal of agreement is seen in the preceding studies. The idea that the meaning of a text differs from the meaning of its individual sentences, in that prior knovledge is used to infer implicit information during comprehension, is generally accepted. The idea that this meaning must be well structured, or coherent, is also expressed directly or indirectly by most of the researchers. However, explicit disagreement exists on both the nature of content structure, and its relationship to recall. Crothers and Meyer disagree as to whether a neutral struccure reflecting pre-existing Page 27 knowledge is built, or the author's structilre used. Kintsch and . van Dijk essentially use both approaches since the time ordering in narratjve is author defined (although the concept of time is pre-existent), but the macro-st ructures which hierarchically organize the text are pre-existent-Meyer, Kintsch and van Dijk all agree that propositions higher in the structure are more likely to be recalled, but Crorhers evidence did not support this conclusion. Clearly, this hypothesis, if true, could be confirmed only if it is tested against the appropriate content structure.</Paragraph>
      <Paragraph position="1"> Spiro's comments regarding the use (pr laah of use) of pre-existing structures are relevant, In the case of unfamiliar expository material, such as that used by Crothers and Meyer, it seems reasonable to suppose that the author s organization would be used to organize the semantic representation due to the lack (or n,on-use) of any appropriate pre-existing content structures. Thus Meyer's results suppart the importance of propositional level, and Crothera results do not- Ylowevev, a familiar structure, like an action sequence, is understpod in the. same way regardless of presentation setting. Thus, Kintsch and van Di jk found that the role of a proposition in the pre-existing structure (the macro-structures) was primary in determining its importance, It should be noted that even Meyer found variations in the recall of midclevel propositions that she could not explain by reference to the author's organization. It is plausible to assume that these variations were caused by individual differences in prior knowledge, resulting in differing content structures.</Paragraph>
      <Paragraph position="2"> Page 28 In light of thts, it seems plausible to propose a general model that distinguishes the following component processes bf text understaading (realizing that like almost all distinctions, such as syntax versus semanticq, they are really ~nadequate, but useful, oversimplifications). First there is the process of comprehensioh of the surface structure which builds a representation that is independent of the surface form of the text, This representat$on contains as much inferred information as is necessary to meet some minimal level of understandability, or coherence. This representation is organized hierarchically by higher level content st ructures ghich summarize or generalize over a larger amount of detailed inPS onnation. These higherlevel structures would normally be pre-existing (i. e. known to the understander) although a text could result in the creation of new structures. Different types of texts could certainly differ in the complexity of each of these tasks, explaining the variation ,in di f f iculty of understanding.</Paragraph>
      <Paragraph position="3"> The process(es) explaining forgettfng, and the integration of new information into prior knowledge, operates on this text representat-ion. This process is poorly understood, and there are no well defined hypotheses about its operation. It appears to operate primarily on the content structure of a text. So, for example, Spiro's subjects integrated information from the stories they read into their general expectations about love and courtship, and lost the specifics of the tpxt, even recalling it with radical mods f i,ca t ions This pxocess would explain the almost total loss of detail information with time, as well as the interactions and distortions caused by previous knowledge.</Paragraph>
      <Paragraph position="4"> Page 29 Finally, recall is a retrieval process which operates on the current form of the rep resentation. If that form is relatively unchanged since comprehension, recall may be almost distortion-f ree. But if significant changes have occurred, recall will be distorted. As suggested by Kintsch and van Dijk, summary is seen as recall to a controlled depth. Retrieval accesses the higher-level organization first, and uses these structures to access more detailed information. As retrieved information is generated as output text, a test of coherence would be made. If something has been forgotten or changed so that a necessqry explanation or link is missing, the retrieval process would supply a probable piece of information from the general pre-existing structures. This, along with modifications from integration into prior knowledge, accounts for the reconstructive aspect of recall-This general model does not account for the recall of very specific surface details (e.g. words and constructions used in the input text) but as suggested by Kintsch's work, it seems best to treat this as a process separate from the general understanding of text. One other inadequacy of the model is that it specifies complete comprehension of all input propositions, including assignment to an appropriate role in a higher level structure. Frederiksen suggests that overgeneralization is a loss of detail which reduces the processing load. The idea that partial processing is sometimes done on input information is also suggested by Bobrow and Norman [75]. They describe such processes as resource-limited. This kind of processing seems to be indicated by facts such as loss of detail in immediate recall and the subjective impression of partial comprehension in a number of situations (such as reading verg rapidly, reading while partially distracted or reading very complex Page 30 material). Wowever, while the idea of discarding comp rellended inPS ormation* could be described computationally, selective comprehension ia a process that requites further investigation.</Paragraph>
      <Paragraph position="5"> One of the principle contributions of computational investigations with this sort of model will be specification of the processes of comprehen~ion, which can only be done by specifying the nature of pre-existing memory structures. Meyer's work is an example of the use of undefined basic stmctures. She criticizes Crothera for requiring two graphs to represent a text one of which ia his Eundame~ital hierarchy of kndwn concepts. She, instead, simply doesn't define any of her predicates first or higher order, apparently taking them as primitive. Assuming that one attempted to define even the argument roles f~r the rhetorical predicates, enormous difficulty would be encountered. Advocates of case systems have always had difficulty defining the exact role of cases, or delimiting the class of entities that could fill a case role. Consider how much more difficult it would be to define what pairs of things could exist in the RESPONSE oF COVARIANCE relations, or exactly what the charactexistics of these arguments are.</Paragraph>
      <Paragraph position="6"> Kintsch and van Di jk attempt to set out much more well defined structures, but even they have little to say about the set of structures that must exist, what information they must contain and how the camprehension algorithms use the text and the structures to build higher-level structures. The section on computational models will consider algorithmic attempts at text understanding, and should clearly indicate the complexity of these issues.</Paragraph>
      <Paragraph position="8"/>
    </Section>
  </Section>
  <Section position="11" start_page="1" end_page="1" type="metho">
    <SectionTitle>
THE FORM OF CONNECTED DISCOURSE
</SectionTitle>
    <Paragraph position="0"> The general model of text understanding suggested in the last section describes the structured content of text representation. This secaan contains work investigating a similar, yet distinct, type of structure - the abstract, underlying form of a text. (Not everyone would accept this distiaction, but it seems analogous to the distinction between the syntactic structure of sentences and the semantic structure of their underlying meaning.) Much of this work has been done by cultural anthxopolopists or by those interested in literary analysis. Little of the traditional work ha8 been done with cognitive or computational models in mind, and hence tends to rely on intuitive understanding of the terms. This less computational work will be presented quite briefly, since the presentation is intended to suggest the kinds of results that have been obtained, rather than the details of these conclusions. It is necessary to define these ideas in a computational way that can be integrated into the models of text understanding, and some recent work has begun this task.</Paragraph>
    <Section position="1" start_page="1" end_page="1" type="sub_section">
      <SectionTitle>
2.1 Structural Analysis Of Text
</SectionTitle>
      <Paragraph position="0"> Most modern work in the structural analysis of texts has roots in the work of the Russian structuralist, Propp [68]. Propp was concerned with the form of Russian folktales. and developed a method for describing the similarities he found in a corpus of 100 folktales. The major structural unit that he developed , he designated a function. These functions act like meta-actions in that they describe major events or event complexes that were Page 32 repeatedly found in the tales.</Paragraph>
      <Paragraph position="1"> For example, Absentation is the function that describes the situation in which one ~f the members of the family absents himself from home, and Trickery is the function which describes a villain's attempt to deceive his victim in order to take possession of him or his belongings. Propp fourid thirty-one such functions to be adequate to describe the events in the tales that he studied. Importantly, the ordering of the functions was found to be fixed. Some functicrns are optional, but if they am present their position is fixed with respect to the other ftlnctions. All of the functions are defined in terns of actors that participate in the situations they describe. P.ropp did some analysis of the restrictions on the assignment of various characters to these functional roles. He notes that the aame character is likely to play a fixed set of roles, which he designates a sphere of action. These spheres of action include such familiar story roles as hero, villain and false hero. Finally, Propp notes that a single folktale may be composed of a sequence of basic tales. He designates the basic tale a  move&amp;quot;. Thus, a folktale ,may be a one-move or a multi-move tale, with each move conforming to the definition of a tale. These rules capture the similarity of the folktales which could differ widely in the details of characters, setting, specific actions, etc. It Is clear that Propp's rules could be expressed formally, and that is what Klein et a1 f?4j did by writing a computer program to generate the function sequences of a number of folktales, using Prupp's definitions.</Paragraph>
      <Paragraph position="2"> The ideas of Propp have been refined and generalized by Barthes, Bremond, Greimas and Todorov (their work is not generally available in English but is reviewed in van Di jk [72] and Weinolci (72) ). Their work has been directed Page 33 toward the redefinition of Propp's functions as propositions, and of actor8 as case relations of these propositions.</Paragraph>
      <Paragraph position="3"> They have also attempted to generalize the functions by making them less specific, and to apply them to other forms df texts. However, Hendricks [72] demonst rates the flexibility of Propp's original functions by using them to analyze part of the Bradbury novel Something Wicked This Way Comes.</Paragraph>
      <Paragraph position="4"> Chafe [72, 751 discusses the process of &amp;quot;verbalizatiou&amp;quot;, by which he means the translation of non-verbal knowledge into verbal output. One of the proposed component processes of verbalization is breaking the knowledge into smaller chunks according to patterns, which are called &amp;quot;s~hemata'~. These schemata are grammar like descriptions of possible verbal sequences. Thus, he notes that a certain class of stories will always be of the form:</Paragraph>
    </Section>
  </Section>
  <Section position="12" start_page="1" end_page="1" type="metho">
    <SectionTitle>
PLOT + MORAL
</SectionTitle>
    <Paragraph position="0"> Chafe also mentions such schemata as TRICK and VISIT as occurring in the stories he has analyzed. He explicitly states that, at this time, schematic analysis of a story is done by &amp;quot;imagination abd intuition&amp;quot;. He discusses aspects of the schematic analysis of several simple tales, but these are mostly illustrative rather than complete, in any sense.</Paragraph>
    <Paragraph position="1"> Colby [73] analyzes Eskimo folktales, identifying &amp;quot;eidons&amp;quot;, which are similar to Propp's functions, and producing a grammar capable of generating some of these tales. Colby observes that most native speakers are probably not very familiar with story grammars of this type (or the knowledge they represent), since only a few individuals are able to generate such tales Nonetheless, it seems reasonable to assume that the same type of structure can Page 34 be found in more conventional stories Although most work in this area has involved analysis of some form of traditional narrative, Labov and Waletzky [67] present analyses of oral vprsions of personal experiences. Even in this informal type of narrative, they find five regular components: an orientation, a complication, an evaluation, the resolution and a coda or moral, The last two components are optional. A totally different type of material is examined by Becker [65] who discusses patterns in expository paragraphs.</Paragraph>
    <Paragraph position="2"> All of these analyses rely on the intuitive understanding of the analyzer to grasp the meaning of the structures from infomal descriptions and to find these structures in the text being analyzed. For this reason, the particular structures that have been suggested are of less interest than is the general result that quite regular high level patterns are found in texts of many types, which may be characterized independently of the specific textual content.</Paragraph>
    <Section position="1" start_page="1" end_page="1" type="sub_section">
      <SectionTitle>
2.2 Kintsch And Van Dijk
</SectionTitle>
      <Paragraph position="0"> In addition to the content macro-structures (discussed in section 1.2.4) Kintsch and van Di jk [75] and van Di jk [75a, 761 discuss even higher level &amp;quot;super-structures&amp;quot;, used to organize the content of a text according to the type of the text, which are derived from the work in structural analysis of texts. They limit their work to narratives, which they define as a specific type of action discourse. Narrative categories such as Episode Setting, Page 35 Complication, etc.</Paragraph>
      <Paragraph position="1"> are superstructures under which are organized either more narrative categories or the content macro-structures of the text. Van Dijk [75aJ presents a fairly complete example. It includes a 1600 word text from The Decameron, a propositional analysis of the story, the content macro-st ructures of the story and the narrative categories under which the macro-st NCtures are organized. The following abbreviated fragment should give an indication of the kind of analysis being proposed.</Paragraph>
    </Section>
  </Section>
  <Section position="13" start_page="1" end_page="1" type="metho">
    <SectionTitle>
COMPLICATION RESOLUTION
</SectionTitle>
    <Paragraph position="0"/>
    <Paragraph position="2"> (detail propositions about this loss) Some expetiments were done to test the role of the understander's notion of narrative structure. An Apache myth and three Decameron stories, which were of comparable difficulty at the sentence level, were presented. They differed in that the Apache myth had an organization not familiar to the subjects. There was much greater variety in the propositions different people Page 36 used in tllc recall' of the myth than in the other three stories. Another' experiment comparcd recalls of subjects who had read a normal story to those of subjects who had rend scrambled versions of the same story. The recalls were indistinguishable by judges. Both of the~e groups of expeiiments indicate the active role of one's expectations about the fom of a discourse in arganizir~g the representation of that discourse, When this is impossible, as in the Apache myth, recall is less organized and more random.</Paragraph>
    <Section position="1" start_page="1" end_page="1" type="sub_section">
      <SectionTitle>
2.3 Rumelhart
</SectionTitle>
      <Paragraph position="0"> Rumelhart (74, 751 proposes a story grammar that is capable of producing structures similar to those described by Kintsch and van Dijk (section 2.2) His grammar contains rules like the following:  1. STORY -&gt; SETTING + EPISODE 2. EPISODE -&gt; EVENT + REACTION 3. REACTlON -&gt; INTERNAL-RESPONSE + OVERT-RESPONSE  Using these rules, he describes the syntactic structure of simple stories. The following fragment illustrstes the types of structures derived (ignoring the parenthc?sizcd predicates for now):  Associated with most syntacrtc story rules are semantic rules which are used to generate a corresponding semantic structure for the story, The semantic rules for the syntactic rules given above are:</Paragraph>
    </Section>
  </Section>
  <Section position="14" start_page="1" end_page="1" type="metho">
    <SectionTitle>
1, ALLOW (SETTING, EPISODE)
2, INITIATE (EVENT, REACTION)
</SectionTitle>
    <Paragraph position="0"> where the semantic predicates are intended to mean what is suggested by their names (e.g. the SETTING ALLOWS the EPISODE to occur). The figure given above contains the appropriate semantic predicate, in parentheses, at each node in the syntactic structure. (Rumelhart uses a completely separate semantic structure,) Page 38 Rumelhart digcusses some elementary summarization rules that operate on the semantic structure of a story. Some illustrative rules are:  1. MOTIVATE (thought, response) =&gt; response 2. INITIATE (X, Y) =&gt; Y 3. INITIATE (X, Y) =&gt; Y (because 1 when 1 after) X These rules could produce the either of the following summaries (for the EPISODE in the story fragment given above): 1. Margie cried.</Paragraph>
    <Paragraph position="1"> 2, Margie cried because her balloon had popped,  Rume.lhart8s ideas are intended to be primarily illustrative (&amp;quot;a tentative beginning of a theoryu), and deal with very simplified stories. They do argue clearly for the notion of both a syntactic and a semantic structure for the representation of a story's meaning. The syntactic structure allows production of the semantic stucture, which is justified by its usefulness in p roducing summaries.</Paragraph>
    <Section position="1" start_page="1" end_page="1" type="sub_section">
      <SectionTitle>
2.4 Mandler And Johnson
</SectionTitle>
      <Paragraph position="0"> Mandler and Johnson [77] discuss an extended form of ~umelhart's story grammar, which they propose as a basis for text recall studies. The extended Page 39 grammar allows multiple episodes and does not utilize the additional structure containing the semantic interpretation which Rumelhart suggests. Mandler and Johnson feel that this second structure is unwieldy and frequently redundaat., Conjunction, time-ordering and causality are included as categories in the grammar. They observe that restrictions imposed on a story by this type of grammar include allowing only a single protagonist per episode, and prohibiting the realization of higher level nodes in the actual text. They demonstrate the utility of their grammar on several stories, including &amp;quot;The War of the Ghosts&amp;quot;. They also note that a transformational component remains to be developed which would provide for deletions and reorderings of the ideal story structure- They discuss a set of predictions for recall studies, which are suggested by this definition of story. structure. These include:  1. Accuracy will be better as the story is complete and ordered as defined by the ideal structure2, Elaborations will be poorly recalled.</Paragraph>
      <Paragraph position="1"> 3. Optional nodes' realizations will be less well recalledm 4.</Paragraph>
      <Paragraph position="2"> Causality will be better recalled than simple time-ordering.</Paragraph>
      <Paragraph position="3"> 5. Inversions should be fewer as the story is closer to the ideal  structure6. Omissions and violations of the ideal structure will be reflected by additions and distort ions.</Paragraph>
      <Paragraph position="5"> Unfortunately, the only experimental work reported is a cowrison of recalls of students from first grade, fourth grade and university levels, on a set of stories which are very near the ideal structure. Hopefully, this is only the beginning stage of validation of the proposed model* An interesting result from this experinlent is the probability of recall of a proposition as a function of its role For university students, Settings (introduction of  characters) and Beginnings (initiation of event sequences) were best remembered* Attempts and Outcomes (which together formed Goal Paths) were next best recalled* Reactions (mental events) and Endings (which were usually either very predictable or omitted in the stories used) were least well recalled. 'These results are in harmony with the results reported by Kintsch and van Dijk (which are discussed in Section 1.2.4).</Paragraph>
    </Section>
    <Section position="2" start_page="1" end_page="1" type="sub_section">
      <SectionTitle>
2.5 Thorndyke
</SectionTitle>
      <Paragraph position="0"> Thorndyke [771 attempts to validate the basic notion of underlyihg text form as a necessary part of a text understanding. Re applies a story structure grammar, which is nearly identical to that proposed by Rumelhart, to two stories, &amp;quot;Ci rcle Island&amp;quot; (analyzed by Frederiksen [75b] ) and &amp;quot;The Old Farmer&amp;quot; (given by Rumelhart [74J), each pf which results in a straightforward hierarchical structure. The initial rule of the grammar is</Paragraph>
    </Section>
  </Section>
  <Section position="15" start_page="1" end_page="1" type="metho">
    <SectionTitle>
STORY -&gt; SETTING + THEME t PLOT + RESOLUTION
</SectionTitle>
    <Paragraph position="0"> Thorndyke wishes to test the effects of violating this rule. He does so by performing recall, summary and recognition experiments on four vari-ations of Page 41 the stories. Thesc are a normal Form of the story, a story with the Theme moved to the end, a story with the Theme omitted and a descriptive version which omits causal and temporal continuity. using only stative or single action sentences. In comp rehensibili ty judgements and recall tests, the normal form story was best, the Theme-after form next, with the last two cases more dependent on the particular story. Recalls of the Theme-after passages alao showed a strong tendency to relocate the Theme to its normal position, near the beginning of the story. There was also a tendency in the more structured passages for higherlevel propositions to have a higher probability of recall. Summarizations showed that, of the recalled propositions, the higherlevel ones were much more likely to be included as part of the summary. Summaries of the descriptive presentations yielded a wider selection of proposi,tions. In recognition tests, the more structured passages produced more recognition errors when the test proposition was consistent with the story, while the less structured passages produced more accurate recognition.</Paragraph>
    <Paragraph position="1"> In attempts to separate the effects of content and structure, Thorndyke tested the four stories obtained by using each of the original plots with two character / object sets. He found that the Farmer plot was always more comprehensible. He also found that presentation of a second story using the same plot structure improved later recall of the first story.</Paragraph>
    <Paragraph position="2"> Thorndyke concludes that text structures are basic to comprehension, although some are inherently more complex than others. The distinct ion between structure and content is supported by the positive effect of repeated structure prior to recall. And finally, the structural position of a Page 42 p roposition has significant inPS hence on both recall end summarization.</Paragraph>
    <Section position="1" start_page="1" end_page="1" type="sub_section">
      <SectionTitle>
2.6 Text Grammar
</SectionTitle>
      <Paragraph position="0"> &amp;quot;Text grammar&amp;quot; is the designation used for another currently active line of investigation into the structure of text. The fundamental ideas of text grammar are discussed by van Dijk [72], Ihwe [72], Kummer [72], Petofi [72] and Petofi and Rieser [73]. (Unfortunately, a great portion of the work is not available in English.) The principle concern is the formulation of generative descriptions of texts, rather than individual sentences, and is usually approached using a model similar to that of the generative semantics  school of linguistics. Van Dijk suggests the following five components of a text grammar: 1. Semantic formation rules for the meanings of texts, as a whole.</Paragraph>
      <Paragraph position="1"> 2. Transformations which operate on this text meaning.</Paragraph>
      <Paragraph position="2"> 3. Transformations which produce a sequence of sentential semantic rep resentations from the text meaning, 4. Transformations which produce q sequence of sentential syntactic representations (including lexical i terns) from tho sequence of sentential semantic rep resent at ions.</Paragraph>
      <Paragraph position="3"> 5. Rules pairing syntactic representations with morphonological representations.</Paragraph>
      <Paragraph position="4"> Page 43 nc- cqonenta would operate in the order given to generate a text* He  cqhasires this is only the outline of a theory, and that most of the hard dttdls rerain to be specified. A significant amount of work has been done on ccqoaents 1 and 2, formall,y defining represenmtions of text. This has in&amp;ded numems examinations of the use of some extended predicate calculus u a basic representation (primarily addressing the meaning of individual satemces and objects) but has also led to psychologically motivated imestiwtioas (as discussed in Section 1.2.4) and to studies in the general stacture of textual types (as discussed in Section 2.2). Discussions of capmmnts 3 and 4 have been mostly descriptive, indicating phenomena that rst be accounted fo~ without actually defining them. Many of these phenomena seem best characterized as operatdons operating on two or more underlying propositions- These phenomena include anaphora (including pronominalization md article selection), sentence stress, contrast, use of clausal caajtanctions, and verb tense determination. Of course, certain literary styles ma alter the tt normal&amp;quot; rules, choosing repetition over prooariaanzation omission of causal connect ion, repet ition of certain #.tcnce forms, etc. Operations which occur on a single proposition include: .amtic transfonations, such as personification and the use of metaphor; wmctic traasformations, such as inversions and other stylistic operations; .rd pbaaological transformations, such as those producing rhyme, alliteration .od meter. The set of realization rules used to produce a particular text fm its generated underlying structure is assumed to be limited by the type of tart that being generated (e.g. mystery, narrative, etc.), which is Lpdicated by some underlying abstract type marker. The selected rules will Page 44 also determine the textual characteristics usually referred to as style or literary merit.</Paragraph>
      <Paragraph position="5"> It seems that any complete text understanding model must have a set of rules relating the underlying structure of the text - individual and groups of propositions together with global aspects of the text - to its surface realization. Although using a generative model, text grammarians are concerned with identifying and characterizing these relationships, and the results of their investigations should be of definite value in describing text understanding, once their research has reached the point of clearly defining the nature of these surface phenomena in terms of the meaning of text.</Paragraph>
    </Section>
    <Section position="2" start_page="1" end_page="1" type="sub_section">
      <SectionTitle>
2.7 Discussion And Conclusions
</SectionTitle>
      <Paragraph position="0"> The principal hypotheses of these investigators is that there is are very high level structures which organize content structures in the representation \of a text. This section was begun with the warning that the distinction between these two different types of structures might be difficult to define. It seems clear that although each proposal that has been considered claims to be concerned with this high level structure, some confusion exists. The interactions between the abstract structures and the content structures are not characterized at all. Thus, Kintsch and van Dijk do not describe the restrictions on what type of events may realize a CON.PLICATION category.</Paragraph>
      <Paragraph position="1"> Similarly, Rumelhart does not explain how the correct semantic rule would be selected when a choice is possible (e.g. Two sequential EVENTS may be related by either CAUSE or ALLOW). Mandler and Johnson, and to some extent Thorndyke, create additional confusion by mixing syntactic categories with categories that would usually be considered semantic. Thus, conjunct ion time ordcdng and causality may appear in the &amp;quot;syntactic&amp;quot; structure of a story. Although this difficulty in distinguishing between the two types of structure.</Paragraph>
      <Paragraph position="2"> suggest that there is no distinction. the evidence provided by the exirrtenca of describable claesea of texts seems to indioate that some type of high level structures doexist. It remains to clearly define the nature of thasa struckures, and to explain their relationship with content structures, without repeating the fallacy of creating a multitude of subcategories (e.g. wny different EVENT subtypes) which are alleged to be purely syntactic categories.</Paragraph>
      <Paragraph position="3"> These structures fit neatly into the model discussed in the conclusion of the content section (section 1.4). They are the highest level of structure, under which content structures are organized. This is consistent with the hypothesis of Kintsch and van Dijk (discussed in section 1-2*4) that the structural description of a text is one of the important aspects of its representation. They specifically suggest that these st mctures organize macro-structures, which provide the content organization of the text. This ia quite ~0nSiSteht with the results of structural analysis which, as Hendrickc [731 points out, ordinarily uses a summary or synopsis, not the actual text.</Paragraph>
      <Paragraph position="4"> as a basis for analysis. (Recall that macro-structures are proposed ae the basis for summary generatfon). Of course very simple or very navel texta might not have any structure of this type, scnce no appropriate structure would be pre-exis'tent. Of course, the problem of learning these structure is as difficult to explain (if not more digficult) as the learning of content struct-uree. Such acquisition is simply not well understood. The work of the Page 46 text grammarians suggests the complexity and Alversity of linguistic phenomena that nsay be related to underlying textual representnt:ion. This work also ruggsats that many aspects of a text, other than its literal meaning and general type, may require explicit representation. Although a eound computational representatian of atyli~tic aspects of text0 may seem a distant goal, it is still important to retain it as part of the goal of understanding text, and to realize that the richness of human use of language will be only partially accounted for without this component</Paragraph>
      <Paragraph position="6"/>
    </Section>
  </Section>
  <Section position="16" start_page="1" end_page="1" type="metho">
    <SectionTitle>
COMPUTATIONAL MODELS OF TEXT UNDERSTANDING
</SectionTitle>
    <Paragraph position="0"> This section will survey computational work on the process of text understanding. Computational models will be seen to be of value in at least two ways. FLrst, many of the features postulated in other models will be found in these computational models, frequently motivated primarily by computational concerns. This tends to support these hypotheses. Secondly, many points passed over very quickly by other workers are seen to offer formidable problems when one attempts to set out a full computational description. The discovery of inadequate description is essential in develeping sound models. &amp;quot;Computational&amp;quot; will be taken to mean any model that is actually programmable or that is formally defined. The work selected for discussion in this section is generally treated in more detail than that in previous sectibns. The reason for this is that if the complexity of rcomputationally specifying certain representations and processes is to be made clear, it is impossible to treat these matters cursorily. This, in turn, has necessitated greater selectivity on the part of the author in choosing work that seems to best convey certain types of problems.</Paragraph>
    <Paragraph position="1"> Bellert 1701 attempts to define the notion of the coherence of a text, and in doing so suggests some of the same conclusions reached by computational linguists in their text understanding work. Coherence refers to the property of a set of utterances which make it a connected discourse rather than a random collect ion of utterances. Referential identity has of ten been suggested as an indication of coherence, but it is clearly insufficient.</Paragraph>
    <Paragraph position="2"> For example: Page 48 John drinks a lot of coffee. John married a blonde.</Paragraph>
    <Paragraph position="3"> Both sentences refer to John, but do not form a coherent discourse. Furthermore, referential identity is unnecessary, Consider the two sentences: There has been a drought. People are starving, These are coherent without any explicit referential identity. Bellert defihes a coherent text as one in which the semantic interpretation of each sentence is dependent on the semantic interpretation of the preceding sentences. The semantic interpretation of a sentence ia defined as the set of conclusions that can be inferred from that sentence. She suggests that there are two types of conclusions that may be drawn:  1. those drawn only from knowledge of the language 2. those drawn from knowledge of the world  Both types of conclusions are absolutely necessary in understanding text and may be appropriately drawn when the coherence of the text requires. She concludes that &amp;quot;an utterance has meaning only in the entire context and through our knowledge of the world&amp;quot;. The rest of the work discussed in this section wd 11 strongly reinforce these conclusions.</Paragraph>
    <Paragraph position="4"> The discussion is divided into two sections. The first contains earlier work that supports the claim that world knowledge is essential to the understanding of text. The closely related problem of making implicitly conveyed information explicit is also a principle concern. The second section Page 49 describes later work in whlch the or~anization of world knowledge is recognized as a critical question for text understanding.</Paragraph>
    <Section position="1" start_page="1" end_page="1" type="sub_section">
      <SectionTitle>
3.1 The Necessity Of World Knowledge
</SectionTitle>
      <Paragraph position="0"> Charniak's work [72, 74, 761 is probably the first attempt to set out in a well defined fashion the dimension of information processing that must be carried on in the understanding of a story. The tarn &amp;quot;understanding&amp;quot; is necessarily vague, but Charniak suggests an intuitive definition. Consider the following story fragment: Fred was goipg to the store. Today was Jack's birthday and Fred was going to get a present.</Paragraph>
      <Paragraph position="1"> It should be clear that if a human had read and understood this fragment that he would be able to answer such questions as  Charniak claims that a semantic representation of this fragment (i.e. an understanding of it) should explicitly contain the answers to ordinary questions such as these. Note the important point that this type of Page 50 understanding could be attained only through the use of general knowledge of the world, along with the explicit statements of the text. Knowledge required to answer the above questions would include such facts as:  1. A person having a birthday is likely ta receive presents. 2. Presents are of ten bought at stores.</Paragraph>
      <Paragraph position="2"> So Charniak's goal becomes outlibing an answer to the question of hw common  sense knowledge may be incorporated into the process of understanding natural language. A closely related goal is the determination of how much knowledge of this type is required by the basic problems of natural language, such as the resolution of pronominal reference, Charniak breaks the problem of processing natural language into two parts. The fir~lt part is the translation of natural l'anguage into a farm that is convenient for use in making deductions. This internal representation is like the understanding that a person would be capable of obtaining without cqntext. The internal representation is a canonical propositional form, Thus, either of the sentences Jack caught a cold.</Paragraph>
      <Paragraph position="3"> Jack came down with a cold, might be represented by the proposition</Paragraph>
    </Section>
  </Section>
  <Section position="17" start_page="1" end_page="1" type="metho">
    <SectionTitle>
(BECOME-SICK-WITH JACK COLD)
</SectionTitle>
    <Paragraph position="0"> which represents the explicit meaning of eithcr of the sentences. The second Page 51 part of the problem is what Charniak calls Deep Semantic Processing (DSP). This is the processing that makes explicit the implicit information conveyed by the story.</Paragraph>
    <Paragraph position="1"> Charniak elects to examine only DSP.</Paragraph>
    <Paragraph position="2"> The function of the system that Charniak proposes is to read a story which has already been translated into internal representation and output a data base of propositions that explicitly represent all of the information conveyed by the story. In order to do this, the system uses commdn sense knowledge to make implicit information explicit-This knowledge is coded into the system initially, and is not learned or modified, Charniak's goal of understanding stories leads directly to the questions of what kind of implicit information is conveyed, what kinds of common sense knowledge are required to explicate that information, how should this common sense knowledge be represented and when and how is it used. In answering the question about the types of implicit inPS ormation Charni ak discovered that resolution of pronominal reference frequently requi red common sense knowledge. Consider Charniak's most discussed example [Charniak, 741 : Today was Jack's birthday. Penny and Janet went to the store. They were going to get presents. Janet decided to get a top. &amp;quot;Don't do that&amp;quot; Penny said. &amp;quot;Jack has a top. He will make you take it bock.&amp;quot; The &amp;quot;it&amp;quot; in the last line is normally understood to refer to the top that Penny would buy, but any purely syntactic procedures would probably select the top that Jack already has, on the basis of recency. The correct choice of referents seems to be based on the common sense knowledge that if a present is purchased and given, it may be returned or exchanged. The fact that if a person receives a duplicate present he may not wish to keep it is also Page 52 relevant. This realization led Charniak to concentrate on the problem of pronominal reference. Instances of reference are clearly recognizable in the input, and the need to utilize common sense knowledge seems fully present. Charniak decided to represent common sense knowledge as inferences. That is, implicit knowledge is made e~plicit by having rules which infer the implicit fact upon presentation of the necessary explicit information. Por example, if one knows that it is raining, the inference that anyone outside will get wet is valid, The presentation of the sentence John went outside.</Paragraph>
    <Paragraph position="3"> is sufficient to trigger the common sense inference John got wet, in a situation in which it is known to be raining. Charniak's chaice of inPS erences as a represent ation was probably strongly influenced by the availability of MICRO-PLANNER. (For a discugsion of MICRO-PLANNER, see Winograd [74] and Charniak 1761) Before examining his inferences in more detail, one other question must be considered. Since an inference is drawn at some particular time, the question of when inferences should be made arises. Three possible answers to the question are suggested:  1. Make no inferences until the story is accessed (e.g. to answer a question about it, summarize a part of it, etc,) Page 53 2. Make inferences as the story is read, but only those that are necessary to solve some particular problem (e. g. to solve p roblems of ambiguity or reference) 3. Make non-problem inferences as the story is read, making explicit aa  much information as possible Charniak rejects the first possibility because if it is necessary to make dedu.cti0ns from all previous propositions when accessing some proposition, there is no theoretical difference from making them as the story is read, And he argues thqt there is no way to be sure that you have made the correct and necessary inferences without examining all previous propositions. An argument against both the first and second possibilities is that the meaning of any proposition may be context-sensitive Consider a story in which Janet wants to trade with Jack for his paints. The sentence &amp;quot;Those paints make your airplane look funny&amp;quot; Janet said. should probably be understood as part of a bargaining strategy and not an expression' of Janet's true feelings. Since no explicit problem is presented by this sentence, no inferencing would be done, and hence the correct understanding would be missed. So Charniak selects the third possib'ili ty , making inferences whenever pos~ible as the story is read.</Paragraph>
    <Paragraph position="4"> Charniak implements inferences essentially as MICRO-PLANNER antecedent theorems. He distjnguishes two types of common sense inferences which he designates base rout ines and demons. Base routines rep resent knowledge that Page 54 should always be avatlable and does not need to be triggered by a specific context* For example, knowledge about t rsding should always be available ro that any statement about a trade that has occurred will cause the ibfarencer that the ownership relations have been reversed. Base routincar are implemented as antecedent theorems matched against each input proporition. Some inferences are not always appropriate, such as the previous example about getting wet in the rain. The inference would be incorrect if made in r non-raining situation. This type of inference, called a demon, is implementad as antecedent theorems that are not always active. Demons are activated by base routines. Thua, there would be a base routine about raining which could make immediate inferences and activate demons appropriate to the context, rain. The previous inference rule uould be a demon of this type.</Paragraph>
    <Paragraph position="5"> Charniak refers to the set of propositions which match a base routine's pattern as its topic concept. He notes that a topic concept may occur either before or after a proposition which would match one of the topic concept'^  demons. Consider the two sequences: 1, It was raining. Jack was outside.</Paragraph>
    <Paragraph position="6"> 2. Jack was outside. It was raining.</Paragraph>
    <Paragraph position="7">  In the first sequence, the rain base routine is matched which activates the outside-implies-wet demon. The demon matches the next proposition causing the inference that Jack is wet. Charniak calls this looking fomard, and it is handled correctly by antecedent theorems. In the second sequence, however, the demon is not activated until after the proposition so the inference is Page 55 mirsed. This is called looking backward. Charniak proposes an extenaion to antecedent theorems so that when a demon is activated, the data base is marched for matches. Then, in the second sequence, the inference would be made when the the demon was activated.</Paragraph>
    <Paragraph position="8"> A final observation about demons concerns deactivation. Obviously, if demons are context dependent, they must be deactivated when the context is no longer preeent. Thi~ is illustrated by It was raining. When it quit, Jack went outside.</Paragraph>
    <Paragraph position="9"> It should not be inferred that Jack got wet. Although thia simple example could probably be handled by the rain base routine, in general the problem is quite difficult and Charniak offers no real solutions. He assumes that deactivating demons after some fixed number of intermediate propositions would be a satisfactory first approximation.</Paragraph>
    <Paragraph position="10"> Charniak's model includes two other components. Be suggests a bookkeeping component to keep the data base updated and tonsistent. When inferences are made, they will frequently replace other assert ions previously true. For example if Jack ia inside, then goes outside, bookkeeping would mark the original fact as no longer true, but would keep it for historical purposes (such as answering the question, &amp;quot;Was Jack inside?&amp;quot;). The fourth component is made up of fact finders. These are necessary as a result of Charniak's decision to make all possible inPS erences which are exp resaed as demons as the story is read.</Paragraph>
    <Paragraph position="11"> Clearly, many possible inferences should not be made to avoid clogging the system with a huge number of assertions.</Paragraph>
    <Paragraph position="12"> To avoid Page 56 this, these unnecessary inferences will not be realized as dcmons. Inferences such as the facts that John is in his house, his neighbor!~ood, ctc, if he is in his kitchen fall into this category. Some situations may require the availability of these fact so The sequence Jack was in the house. Later he was in the kitchen.</Paragraph>
    <Paragraph position="13"> should not be interpreted as a change in location, from the house to the kitchen. Fact finders are implemented as MICRO-PLANNER consctluent theorems, with patterns which are matched against desired goals. They thus make certain information available through deduction that is not sufficiently important to infer as soon as possible. Fact finders are always available, not activated and deactivated like demons.</Paragraph>
    <Paragraph position="14"> Charniak's complete model is depicted below. The model shows that the inferences are treated ;Like additional propositions and are thus subject to  In some ways Rieger's work [74. 751 seems closely related to Charniak's, but it is not clear that this is completely true. Ricger's system accepts sentences as input which are already analyzed into their semantic representations, and makes explicit additional information that he claims is Page 57 iqlicitly in the input.</Paragraph>
    <Paragraph position="15"> Re designates these additional pieces of information inferences, and agrees with Charniak that they should be made whenever possible Rieger discusses the use of inPS erences in handling problems such as reference, but the bulk of his work is specification of the inferences, themselves, and it is here that the main value of his work is found. Rieger Sdentlfies sixteen classes of inferences which will be discussed in three brod categories (not Rieger's categories).</Paragraph>
    <Paragraph position="16">  Pirst, there are inferences which are concerned with the causal camectiopg between states and acts (Rieger's treatment resembles the use in robotics wo* (e.8. Fikes and Nilsson [71] ) of preconditions - states that be true for an act to occur - and postconditions - states that result fror m act occurring). Given that a state or act is true or has occurred, rbrt inferences may be made?  1. Causative inferences suggest the likely cause input: Mary has the diamond ring.</Paragraph>
    <Paragraph position="17"> inference: Someone must have given or sold the ring to Mary.</Paragraph>
    <Paragraph position="18"> 2. Enablement inferences suggest states that were necessarily t me input: Mary gave John a book, inference: Mary had the book just before she gave it to John.</Paragraph>
    <Paragraph position="19"> 3. Resultative inferences suggest results that f ollswed input: Mae gave John the book.</Paragraph>
    <Paragraph position="20"> inference: John has the book, 4. Missing enablement inferences explain why something cannot occur input: Mary couldn't see the horses finish.</Paragraph>
    <Paragraph position="21"> inference: Something must be blocking her view.</Paragraph>
    <Paragraph position="22"> Page 58 5. Intervention inferences explain how something may be stopped or  p revent ed input: Mary was hicting John ljith a bat.</Paragraph>
    <Paragraph position="23"> inference: Taking the bat away from Mary would stop her.</Paragraph>
    <Paragraph position="24">  The second category of inPS e rences concerns supplying common knowladgo about familiar objects or actions that is not in the input.</Paragraph>
    <Paragraph position="25"> 1, Specification inferences fill in missing parts input: John hit Mary.</Paragraph>
    <Paragraph position="26"> inference: John used his hand to hit Mary, 2. Function inferences supply the normal role of objects input: John got a book.</Paragraph>
    <Paragraph position="27"> inference: John will read tbe book.</Paragraph>
    <Paragraph position="28">  3. Normatiwe inferences supply information about what f s normally true input: Pete is a, human.</Paragraph>
    <Paragraph position="29"> inference: Pete probably has a gall bladder.</Paragraph>
    <Paragraph position="30"> 4, State duration inferences suggest how long some state will persist  input: John started eating at 6:00.</Paragraph>
    <Paragraph position="31"> inference: fie is probably still eating at 6:15.</Paragraph>
    <Paragraph position="32"> 5. Feature inferences connect features of objects with the objects input: Fred wagged his tail.</Paragraph>
    <Paragraph position="33"> Xnference: Fred is a non-human animal6. situation inferences supply other likely aspects of a sftuation input: Mary is going to a masquerade.</Paragraph>
    <Paragraph position="34"> inference: Mary is probabky wearing a costume* 3.1.2.3 Motivation Atid Rnwledse The third category of lnfarsoces concerns human mrivationr and knowledge, and their relation to one's actions.</Paragraph>
    <Paragraph position="35">  1. Motivation inferences suggest reasons for an actor to do eo~vlthing input: John hit Mary.</Paragraph>
    <Paragraph position="36"> infetenee John wanted Mary to be hurt.</Paragraph>
    <Paragraph position="37"> 2. Actioa-pwdiction infbmncao suggert a poseible courae of actian from a pamon's wants input: John wants some naila.</Paragraph>
    <Paragraph position="38"> inaereacer John is likely to go to a hardware store.</Paragraph>
    <Paragraph position="39"> 3. Enablement-p rediction inferences suggest reasons for s pamon  bringing about a certain etate input: Mary put on hot glasses.</Paragraph>
    <Paragraph position="40"> inferencd: Mary probably wants to look at emthing.</Paragraph>
    <Paragraph position="41"> 4, Utterance-intent inferences supply information intended, but not actually ~tated input: Mary couldn't jump the fence.</Paragraph>
    <Paragraph position="42"> inference: Mary wanted to jump the fence.</Paragraph>
    <Paragraph position="43"> 5. Knwledge propagation inferences predict whet else a patson would know if given that he knows certain things input: Bill knew that Mary hit. John with a bat.</Paragraph>
    <Paragraph position="44"> inference: Rill knew that John had been hurt.</Paragraph>
    <Paragraph position="45"> One point that ehould be clarified is that the use of the tern inference to designate the implicit information a sentence conveys ie partidlly misleedcng. What is implicit, and thus misht need to be made explicit, is directly a function of the seaant ic representation. Consider the previous examp le Mary gave John the book.</Paragraph>
    <Paragraph position="46"> Page 60 If the fact that John possesses the book after being given it is an inference, It 1s not clear what meaning the representation of the sentence captures. Surely, the transfer df possessian ia the meaning of the sentence, not just an implicit addition. At the other extreme, the example John hit Mary.</Paragraph>
    <Paragraph position="47"> surely does qot necessarily mean that John wanted to hurt Mary. This is clearly conveyed implicitly, if at all.</Paragraph>
    <Paragraph position="48"> It is very difficult to evaluate the completeness or adequacy of such a large set of inference types, but Rieger certainly suggests how large the class of information is that can be implicitly conveyed by text and, therefore which might need to be made explicit during the processing of that text. Any proposed scheme for representing and understanding text should be examined to see how each of these types of information is stored and in what way the information is asserted to be part of the meaning of the input.  Charniak's model of story comprehension primarily discgsses conne~ttot~a between propositions of a story that alter the sequence of events involved in understanding the ston. Far example, the understanding of some sentence of a story. may be correct because a demon had been previously activated, thus allowing the correct interpretation of that sentence. Charniak dges not say much about explicitly rep reseating the connection between the proposition responsible for activating the demon, and the correctly understood Page 61 proposition. However, since he mainly investigates reference, obtaining the correct referent is an explicit representation of the understanding. Clearly, many kinds of connections might exist between propositions of a connected discourse. Schank ( [73b], Schank and Abelson [77] ) investigates causality as one of the primary discourse connectives. Schank's proposals are discussed in terms of his Conceptual Dependegcy theory [Schank, 73a], but are largely independent of it. The term conceptualization is used to refer to either a proposition (using one of the small set of primitive predicates or acts) or to a state (i.e. an object having some value for some attribute). Schank notes that the sentence John cried because Mary said he was ugly.</Paragraph>
    <Paragraph position="49"> asserts a connection between Mary said John was ugly.</Paragraph>
    <Paragraph position="50"> and John cried.</Paragraph>
    <Paragraph position="51"> But closer consideration reveals that some of the links in this connection are unspecified, John must have found out that Mary had said it and this knowledge must have made him unhappy, which was manifested in his crying. In order to be able to recognize instances of missing links and to be able to supply them, Schank develops a classification of types of causality and characterizes their form and meaning. (It should be clear that here, as in all Conceptual Dependency, there is no simple mapping from English to the Page, 62 underlying representation, ) The first type of causation is a result in which an act may bring about a change of state, For example John went to New York, -RESULT-&gt; John is in New York, The second type of causation is enablement This is the situation in which a change of state brings about the conditions necessary for some act to occur, John has a ring, -ENABLE-&gt; John gave it to Mary.</Paragraph>
    <Paragraph position="52"> The third causal type, initiation, is the relationship between any conceptualization and the act bf someone thinking about that thing. John -INITIATE-&gt; I think about Bill.</Paragraph>
    <Paragraph position="53"> (John reminds me of Bill. ) The fourth and final type is reason causation, This is the relationship that holds between the act of deciding to do something and actually doing that thing.</Paragraph>
    <Paragraph position="54"> John decided to leave town. -RFASON-&gt; John left town, Schank also suggests the utility of s non-specific caugal connection, which he designates &amp;quot;lead-t 0&amp;quot;. This would be used to rep resent causally connected events in situations where the specific type of causality or the specific chain of causes is unknown, Page 63 Although Schnnk explores the purely syntactic expansion of English causals into his formal causals, he concludes that the syntactic approach is too limited. Consider the sentence The hurricane caused my depression.</Paragraph>
    <Paragraph position="55"> Treating this as an instance of initiation causation (the hurricane initiated depressed thinking) makes it causal-syntact ically correct, but there ie clearly something missing that explains what led to the depressian. Schank suggests that other knowledge is used to find the connection, and this suggestion seems to indicate a far more general approach to understanding English expressions of causality than any syntactic approach, Schank suggests that a person would, if possible, find a reasonable connection such as that the hurricane probably blew down the speaker's house which caused him to lose money which caused him to become depressed, In order to accomplish this elaboration a person would have to use a great deal of world knowledge. Schank proposes that a number of different kinds of knowledge are involved here. First, there are axioms about the way people feel, such as Al. Bad results can cause depression.</Paragraph>
    <Paragraph position="56"> Associated with an axiom would be a number of more idiosyncratic beliefs, such as IB1. Less money is a bad result.</Paragraph>
    <Paragraph position="57"> Finally, general knowledge of the world would also be necessary. This would necessarily include facts auch as WK1. Objects can have monetary value.</Paragraph>
    <Paragraph position="58"> WK2. Changes in an object's physical state can cause a change in its worth* A person making the above connections would essentially be engaged in problem solving. Given the final state, depressioo, he would find en axiom which explains it then find an idiosyncratic belief which would meet the exim's preconditions (bad results), then use world knowledge to establish that a negative change of state could be the real culprit, then notice that a hurricane is able to destroy objects. The real point of this discussion seems to be that much additional knowledge is required to establish conrlections that people are able to (and do) make when understanding connected sentences. This is an affirmation of Bellert's hypothesis, and conPS irms Charniak's conclusions, but is reached after work on a different text understanding p roblem, Schank [75b] attempts to use these types of causality as a basis for defining the semantic rep resentation of a paragraph. lie argues that the collected rep resentat ions of the individual sentences of the paragraph do not form a representation of the paragraph as a whole. Much addttinnal implicit information could be made explicit, but Schank suggests that undirected explication is not plausiblee He offers a solution to the problem of what information should be made explicit by defining a paragraph representation as a causally connected sequence. Inferences are limited to those items of information required t~ find the causal connections.</Paragraph>
    <Paragraph position="59"> Page 65 The etatea that are usually true when an act occurs are called the neceasaty condition8 of that act.</Paragraph>
    <Paragraph position="60"> These states would be ~onnectcd to the act by an enabling causation.</Paragraph>
    <Paragraph position="61"> Schank divides necessary conditions into two types: absolutely necessary conditions and reasonable necessary conditions (ANCs and RNCs). The former must always be true when an act occurs. RNCs, however, tare normally true, but may be violated without creating an anomalous situation. For example John was working in his yard.</Paragraph>
    <Paragraph position="62"> has a8 an ANC (among others) that John has a yard. An RNC might be that it was nice weather, but this could be violated* It would only indicate something unusual, and perhaps significant. Schank illustrates the use of neceeeary conditions to establish causal chains by analyzing several stories. A brief description of one of these should indicate his methods. He considers the following paragraph: John began to mow his lawn. Suddenly his toe started bleeding, ... When he cleaned off his toe, he discovered that he had stepped in tomato sauce. Mowing the lawn has as ANCs such things as the existence of John, the lawn, a mower, etc. and as RNCs that it is good weather and the grass needs mowing. These conditions would be inferred when the first sentence was processed since they were not explicitly stated. If the first sentence had been preceded by an explicit statement of either or both of these RNC6, they would have been connected as enabling the mowing. (There is great difficulty in finding a general scheme which allows inferring of normal necessary conditions, but prohibits inferring unusual ones. In this case, it secms reasanable to infer Page 66 all of them.) The second sentence presents a more difficult problem. Same sequence of inferences would have to be made which produces the chain: mowing involves turning blades, which could hit a toe, which could cause it to bleed. The final sentence of the story illustrates the point about normality considerations. To explain the tomato sauce, one must infer that John got it an his toe, but this seems like an abnoml inference to make with no other explanation. However, if the paragraph was preceded by the sentences: John was eating pizza outside on his lawn, He noticed that the grass was very long and and he got out his mower.</Paragraph>
    <Paragraph position="63"> Then the final sentence seems much more reasonable. One would infer that pizza has tomato sauce, and John dropped some on the lawn and later stepped in it, In summary, Schank states the following conditions for the representation of a paragraph:  1. Each input sentence is represented.</Paragraph>
    <Paragraph position="64"> 2. These rep resentat ions should be conceptually connected, primarily by causal chains, 3. The necessary condl tfons for every conceptualization must be explicitly represented, and may originate as input sentences or may be inferred either from previous conceptualizations or because they are normal.</Paragraph>
    <Paragraph position="65"> 4. A story is the joining of causal chains that crilminote in the &amp;quot;point&amp;quot;  of the story.</Paragraph>
    <Paragraph position="66"> Other paths are of less interest.</Paragraph>
    <Section position="1" start_page="1" end_page="1" type="sub_section">
      <SectionTitle>
3.2 The Organization Of World Knowledge
</SectionTitle>
      <Paragraph position="0"> A characteristic common to all of the preceding work is that a great deal of world knowledge ia required, but there is no clear organization of this knowledge. It is clear that any system with large amounts of knowledge represented simply as demons or inference rulea would become bogged dawn searching for relevant knowledge and would quite probably draw incorrect inPS erences because knowledge would be applied in inappropriate contexts (recall Charniak's concern about the deactivation of demons). Charniak [ 75, 76, 771 himself suggests that more organized knowledge would be superior to the demon approach. These realizations have led to a number of proposals for the organization of knowledge. Minsky [75] iatroduced the term &amp;quot;frame&amp;quot; for knowledge structures, but Winograd I751 Bobrow and Norman [75] (using tb term &amp;quot;schema&amp;quot;) and Rumelhart and Ortbny (771 discuss closely related ideas. Recently, Bobrow and Norman 1771 have described a language for knowledge representation (Ec2RL) in which the frame concept plays a central role. These knowledge structures are intended to organize conventional or encyclopedic kncrwledge, rather than def initionsl features or characteristics, This knowledge is described in terms of roles or slots which participate in the situation. These are the variables of the frame. Frames must also have the ability to reference other frames. A detailed discussion ot the nature of Page 58 frames is inappropriate here, primarily because it is outside the scope of this survey, but in part because many of the computational questions regarding frames are not completely answered. However, recent computational work haa begun the attempt to incorporate such organization of knowledge into text understanding systems.</Paragraph>
      <Paragraph position="1">  Schank ([75bl, Schank et a1 1751 and Cullingford [751) extends his ideas wfth the addition of large, pre-existing knowledge structures that he calls scripts. He says that it is necessary to have large amounts of specific knowledge about known situations, since otherwise it is difficult or impossible to recognize the causal relationships between events. A script is a sequence of actions that provide knowledge about the typical occurrence of some situation. The following is a partial description of the script for going to a restaurant:  The scenes are the main episodes of the event'and each is defined in terms of a sequence of specific actions. Schank suggests that scripts would need to be divided at the top level into different tracks whicih distinguish the sequences for different types of restaurants.</Paragraph>
      <Paragraph position="2"> Th&amp; usefulness of a script is seen by considering the sequence Johnwent to a restaurant. He ordered a hamburger from the waitress. He paid and left.</Paragraph>
      <Paragraph position="3"> Many details of this sequence have been omitted and connections between the actions would be impossible to establish wf thout knowledge of what constitutes going to a restaurant. Furthermore, the definite reference to the waitress is meaningful only because all participants in a script are automatically introduced by any reference to that script. The script also contains the conditions that must be true for successful execution of the script, as well as the results of successful execution. Each scene has one act that is considered the essential act of that part of the script. These acts are Page 70 called the MAIN CONceptualizations of a scene.</Paragraph>
      <Paragraph position="4"> Scripts represent the knowledge about conventional situatiow that any reader would be assumed to know. But the separate question of how to represent a text which contains script usages must be answered. If the rcy resentat ion of the restaurant story above wee merely an elaborated causal chain containing many events inPS erred from the restaurant script, the unifying description that these events together constitute going to a restaurant worrld be lost. Schank. proposes that that the elaborated causal chain is only one level of the story's representation - the Conceptual Dependency (CD) level. There is also a second level of representation - the Knowledge Structure (KS) level. At the KS level the above story would have a representation like:</Paragraph>
    </Section>
    <Section position="2" start_page="1" end_page="1" type="sub_section">
      <SectionTitle>
ScriptsRestaurant
Custome rcJohn
Food=Hambu rge r
</SectionTitle>
      <Paragraph position="0"> Additionally, the script representation would be linked to each of the events at the CD level which was part of that script instance. Schank suggests that the causal chain at the CD level should contain all the events explicitly mentioned plus the MAINCONS of any scene that is mentioned. If any event is encountered which is not anderstandable in terms of the current script, the event will be represented at the CD level, but will also be placed on a Wierd list. For example Schank discusses the processing of a story in which John has his pocket picked while riding a subway. Later, he has no money with which to pay the bill after eating at a restaurant- The first sequence of events is understood in terms of the subway script, but the pocket picking event is placed on the Wierd list. When applying the restaurant Gcript, the &amp;aability to pay is encountered in the text and is inexplicable from the rotm~rmt script. The script applier asks the monitoring program to look for cprrs~al occurrences that could result in John having no money. The dtor finds the pocket picking event, and returns the informati6n to the scrlpt qplier which uses it as the required explanation.</Paragraph>
      <Paragraph position="1"> Of course, the problems in general are much more difficult than this riqle example indicates. For example, an unusual event within a script could imterrupt the script's normal continuation. Obstacles, such as not being able to get wbat you want at a restaurant, may cause altering or abandoning the restaurant script. Distractions, sueh as a robbery in a testaurant may lead to suspension of the restaurant script for a sub-story, or to its abandonment, Furthermore, simultaneous,, independent scripts are possible. The situation of eating in a dining car involves both the train script and the restaurant script. Finally, the activation and termination of scripts is a very complex p roblem. For example, consider the sequence: John went to a restaurant, After eating lobster, he bought a watch.</Paragraph>
      <Paragraph position="2"> Does this describe a restaurant event followed by a purchasing event, or the more unusual case of buying a watch in a restaurant? Either is possible, and the story in which this s,equence is found would determine the most probable interpretation.</Paragraph>
      <Paragraph position="3"> Page 72  Schank and Abelson (75, 771 (also see Meehan 1761 and Wilensky [76]) have recently suggested that a11 connected event sequences are not apprcrp riatoly represented by scrlpts. Consider the sequence: Johh wanted to become king- He went to get some arsenic.</Paragraph>
      <Paragraph position="4"> It seems that treating the poisoning-the-king situation as so conventioarl that a script for it exists is unrealistic, The solution lies in realixlag that while scripts handle well known situations, mechanisms rmst exist which are able to handle novel situations using general knowledge. &amp;quot;Plans&amp;quot; are suggested as the appropriate mechanism. A plan is a general course of actaon intended to realize some goal(s), A possible hierarchy of high-level goals is shown in Figure 3-l(a). Motivated actions are always associated with the satisfaction of some high-level goal- Frequently, more specific sub-goals exist which are motivated by a high-level goal. For example, someone may want to go to the train statton (instrumental goal) so he can go to New York (a</Paragraph>
    </Section>
  </Section>
class="xml-element"></Paper>
Download Original XML