File Information
File: 05-lr/acl_arc_1_sum/cleansed_text/xml_by_section/metho/92/p92-1030_metho.xml
Size: 36,510 bytes
Last Modified: 2025-10-06 14:13:17
<?xml version="1.0" standalone="yes"?> <Paper uid="P92-1030"> <Title>TENSE TREES AS THE &quot;FINE STRUCTURE&quot; OF DISCOURSE</Title> <Section position="4" start_page="0" end_page="232" type="metho"> <SectionTitle> 2 Farewell to Reichenbach </SectionTitle> <Paragraph position="0"> Researchers concerned with higher-level discourse structure, e.g., Webber \[1987; 1988\], Passonneau \[1988\] and Song and Cohen \[1991\], have almost invariably relied on some Reichenbach \[1947\]-1ike conception of tense. The syntactic part of this conception is that there are nine tenses in English, namely simple past, present and future tense, past, present and future perfect tense, and posterior past, present and future tense 1 (plus progressive variants). The semantic part of the conception is that each tense specifies temporal relations among exactly three times particular to a tensed clause, namely the event time (E), the reference time (R) and the speech time (S). On this conception, information in discourse is a matter of &quot;extracting&quot; one of the nine Reichenbachian tenses from each sentence, asserting the 1Exarnples of expressions in posterior tense are would, was going to (posterior past), is going to (posterior present), and will be going to (posterior future).</Paragraph> <Paragraph position="1"> appropriate relations among E, R and S, and appropriately relating these times to previously introduced times, taking account of discourse structure cues implicit in tense shifts.</Paragraph> <Paragraph position="2"> It is easy to understand the appeal of this approach when one's concern is with higher-level structure. By viewing sentences as essentially flat, carrying tense as a top-level feature with nine possible values and evoking a triplet of related times, one can get on with the higher-level processing with minimum fuss. But while there is much that is right and insightful about Reichenbach's conception, it seems to us unsatisfactory from a modern perspective. One basic problem concerns embedded clauses. Consider, for instance, the following passage.</Paragraph> <Paragraph position="3"> (1) John will find this note when he gets home.</Paragraph> <Paragraph position="4"> (2) He will think(a) Mary has left(b).</Paragraph> <Paragraph position="5"> Reichenbach's analysis of (2) gives us Eb < S, Rb < Ra, Ea, where tl < t~ means tl is before t2, as below.</Paragraph> <Paragraph position="7"> That is, John will think that Mary's leaving took place some time before the speaker uttered sentence (2). This is incorrect; it is not even likely that John would know about the utterance of (2). In actuality, (2) only implies that John will think Mary's leaving took place some time before the time of his thinking, i.e., S < Ra, Ea and Eb < Rb, Ra , as shown below.</Paragraph> <Paragraph position="9"> Thus, Reichenbach's system fails to take into account the local context created by syntactic embedding.</Paragraph> <Paragraph position="10"> Attempts have been made to refine Reichenbach's theory (e.g., \[Hornstein, 1977; Smith, 1978; Nerbonne, 1986\]), but we think the lumping together of tense and aspect, and the assignment of E, R, S triples to all clauses, are out of step with modern syntax and semantics, providing a poor basis for a systematic, compositional account of temporal relations within clauses and between clauses. In particular, we contend that English past, present, future and perfect are separate morphemes making separate contributions to syntactic structure and meaning. Note that perfect have, like most verbs, can occur untensed (&quot;She is likely to have left by now&quot;). Therefore, if the meaning of other tensed verbs such as walks or became is regarded as composite, with the tense morpheme supplying a &quot;present&quot; or &quot;past&quot; component of the meaning, the same ought to be said about tensed forms of have. The modals will and would do not have untensed forms. Nevertheless, considerations of syntactic and semantic uniformity suggest that they too have composite meanings, present or past tense being one part and &quot;future modality&quot; the other. This unifies the analyses of the modals in sentences like &quot;He knows he will see her again&quot; and &quot;He knew he would see her again,&quot; and makes them entirely parallel to paraphrases in terms of going to, viz., &quot;He knows he is going to see her again&quot; and &quot;He knew he was going to see her again.&quot; We take these latter &quot;posterior tense&quot; forms to be patently hierarchical (e.g., is going to see her has 4 levels of VP structure, counting to as an auxiliary verb) and hence semantically composite on any compositional account. Moreover, going to can both subordinate, and be subordinated by, perfect have, as in &quot;He is going to have left by then.&quot; This leads to additional &quot;complex tenses&quot; missing from Reichenbach's list.</Paragraph> <Paragraph position="11"> We therefore offer a compositional account in which operators corresponding to past (past), present (pres), future (futr) and perfect (perf) contribute separately and uniformly to the meanings of their operands, i.e., formulas at the level of LF. Thus, for instance, the temporal relations implicit in &quot;John will have left&quot; are obtMned not by extracting a &quot;future perfect&quot; and asserting relations among E, R and S, but rather by successively taking account of the meanings of the nested pres, futr and perf operators in the LF of the sentence. As it happens, each of those operators implicitly introduces exactly one episode, yielding a Reichenbach-like result in this case. (But note: a simple present sentence like &quot;John is tired&quot; would introduce only one episode concurrent with the speech time, not two, as in Reichenbach's analysis.) Even more importantly for present purposes, each ofpres, past, futr and perf is treated uniformly in deindexing and context change. More specifically, they drive the generation and traversal of tense trees in deindexing.</Paragraph> </Section> <Section position="5" start_page="232" end_page="236" type="metho"> <SectionTitle> 3 Tense Trees </SectionTitle> <Paragraph position="0"> Tense trees provide that part of a discourse context structure 2 which is needed to interpret (and deindex) temporal operators and modifiers within the logical form of English sentences. They differ from simple lists of Reichenbachian indices in that they organize episode tokens (for described episodes and the utterances themselves) in a way that echoes the hierarchy of temporal and modal operators of the sentences and clauses from which the tokens arose. In this respect, they are anal2In general, the context structure would also contain speaker and hearer parameters, temporal and spatial frames, and tokens for salient referents other than episodes, among other components--see \[Allen, 1987\].</Paragraph> <Paragraph position="1"> ogous to larger-scale representations of discourse structure which encode the hierarchic segment structure of discourse. (As will be seen, the analogy goes further.) Tense trees for successive sentences are &quot;overlaid&quot; in such a way that related episode tokens typically end up as adjacent elements of lists at tree nodes. The traversal of trees and the addition of new tokens is simply and fully determined by the logical forms of the sentences being interpreted.</Paragraph> <Paragraph position="2"> The major advantage of tense trees is that they allow simple, systematic interpretation (by deindexing) of tense, aspect, and time adverbials in texts consisting of arbitrarily complex sentences, and involving implicit temporal reference across clause and sentence boundaries. This includes certain relations implicit in the ordering of clauses and sentences. As has been frequently observed, for a sequence of sentences within the same discourse segment, the temporal reference of a sentence is almost invariably connected to that of the previous sentence in some fashion. Typically, the relation is one of temporal precedence or concurrency, depending on the aspectual class or aktionsart involved (eft, &quot;John closed his suitcase; He walked to the door&quot; versus &quot;John opened the door; Mary was sleeping&quot;). However, in &quot;Mary got in her Ferrari. She bought it with her own money,&quot; the usual temporal precedence is reversed (based on world knowledge). Also, other discourse relations could be implied, such as cause-of, explains, elaborates, etc. (more on this later). Whatever the relation may be, finding the right pair of episodes involved in such relations is of crucial importance for discourse understanding. Echoing Leech \[1987, p41\], we use the predicate constant orients, which subsumes all such relations. Note that the orients predications can later be used to make probabilistic or default inferences about the temporal or causal relations between the two episodes, based on their aspectual class and other information. In this way they supplement the information provided by larger-scale discourse segment structures.</Paragraph> <Paragraph position="3"> We now describe tense trees more precisely.</Paragraph> <Paragraph position="4"> Tense Tree Structure The form of a tense tree is illustrated in Figure 1. As an aid to intuition, the nodes in Figure 1 are annotated with simple sentences whose indexical LFs would lead to those nodes in the course of deindexing. A tense tree node may have up to three branches--a leftward past branch, a downward perfect branch, and a rightward future branch. Each node contains a stack-like list of recently introduced episode tokens (which we will often refer to simply as episodes).</Paragraph> <Paragraph position="5"> In addition to the three branches, the tree may have (horizontal) embedding links to the roots of embedded tense trees. There are two kinds of these embedding links, both illustrated in Figure 1. One kind, utterance pres node. .......... ~home 't &quot;f He left i(&quot; ~ ~ .... ~KP res Ho ,,,,o.vo \ He had left He wbuld He will She will think olo.</Paragraph> <Paragraph position="6"> He would have left indicated by dashed lines, is created by subordinating constructions such as VPs with that-complement clauses. The other kind, indicated by dotted lines, is derived from the surface speech act (e.g., telling, asking or requesting) implicit in the mood of a sentence. On our view, the utterances of a speaker (or sentences of a text, etc.) are ultimately to be represented in terms of modal predications expressing these surface speech acts, such as \[Speaker tell Hearer (That ~)\] or \[Speaker ask Hearer (Whether ~)\]. Although these speech acts are not explicitly part of what the speaker uttered, they are part of what the hearer gathers from an utterance. Speaker and Hearer are indexical constants to be replaced by the speaker(s) and the hearer(s) of the utterance context. The two kinds of embedding links require slightly different tree traversal techniques as will be seen later.</Paragraph> <Paragraph position="7"> A set of trees connected by embedding links is called a tense tree structure (though we often refer loosely to tense tree structures as tense trees). This is in effect a tree of tense trees, since a tense tree can be embedded by only one other tree. At any time, exactly one node of the tense tree structure for a discourse is in focus, and the focal node is indicated by ~). Note that the &quot;tense tree&quot; in Figure 1 is in fact a tense tree structure, with the lowest node in focus.</Paragraph> <Paragraph position="8"> By default, an episode added to the right end of a list at a node is &quot;oriented&quot; by the episode which was previously rightmost. For episodes stored at different nodes, we can read off their temporal relations from the tree roughly as follows. At any given moment, for a pair of episodes e and e' that are rightmost at nodes n and n', respectively, where n' is a daughter of n, if the branch connecting the two nodes is a past branch, Is' before e\]3; if it is a perfect branch, \[e' impinges-on e\] (as we explain later, this yields entailments \[e' before e\] if e' is nonstative and \[e' until e\] if e' is stative, respectively illustrated by &quot;John has left&quot; and &quot;John has been working&quot;); if it is a future branch, \[d after e\]; and if it is an embedding link, \[d at-about e\]. These orienting relations and temporal relations are not extracted post hoc, but rather are automatically asserted in the course of deindexing using the rules shown later.</Paragraph> <Paragraph position="9"> As a preliminary example, consider the following passage and a tense tree annotated with episodes derived from it by our deindexing rules: (3) John picked up the phone.</Paragraph> <Paragraph position="10"> (4) He had told Mary that</Paragraph> <Paragraph position="12"> he would call her.</Paragraph> <Paragraph position="13"> ecall u3 and u4 are utterance episodes for sentences (3) and (4) respectively.</Paragraph> <Paragraph position="14"> Intuitively, the temporal content of sentence (4) is that the event of John's telling, etdz, took place before some time el, which is at the same time as the event of John's picking up the phone, epiek; and the event of John's calling, eean, is located after some time e2, which is the at the same time as the event of John's telling, eteu. For the most part, this information can be read off directly from the tree: \[eple~ orients el\], \[etett before el\] and \[eeatt after e2\]. In addition, the deindexing rules yield \[e2 same-time etell\]. From this, one may infer \[etell before epic~\] and \[ecau after eteu\], assuming that the orients relation defaults to same-time here.</Paragraph> <Paragraph position="15"> How does \[epiek orients el\] default to \[epiek same-time eli? In the tense tree, el is an episode evoked by the past tense operator which is part of the meaning of had in (4). It is a stative episode, since this past operator logically operates on a sentence of form (perf &), and such a sentence describes a state in which & has occurred--in this instance, a state in which John has told Mary that he will call her. It is this stativity of el which (by default) leads to a same-time interpretation of orients. 4 Thus, on our account, the tendency of past perfect &quot;reference time&quot; to align itself with a 3Or, sometimes, same-time (cf., &quot;John noticed that Mary looked pale&quot; vs. &quot;Mary realized that someone broke her vase&quot;). This is not decided in an ad hoc manner, but as a result of systematically interpreting the context-charged relation belT. More on this later.</Paragraph> <Paragraph position="16"> 4 More accurately, the default interpretation is \[(end-of epick ) same-time ell, in view of examples involving a longer preceding event, such as &quot;John painted a picture. He was pleased with the result.&quot; previously introduced past event is just an instance of a general tendency of stative episodes to align themselves with their orienting episode. This is the same tendency noted previously for &quot;John opened the door. Mary was sleeping.&quot; We leave further comments about particularizing the orients relation to a later subsection. We remarked that the relation \[e2 same-time etett\] is obtained directly from the deindexing rules. We leave it to the reader to verify this in detail (see Past and Futr rules stated below). We note only that e2 is evoked by the past tense component of would in (4), and denotes a (possible) state in which John will call Mary. Its stativity, and the fact that the subordinate clause in (4) is &quot;past-dominated, ''5 causes \[e2 bef T eteu\] to be deindexed to \[e2 same-time etch\].</Paragraph> <Paragraph position="17"> We now show how tense trees are modified as discourse is processed, in particular, how episode tokens are stored at appropriate nodes of the tense tree, and how deindexed LFs, with orients and temporal ordering relations incorporated into them, are obtained.</Paragraph> <Paragraph position="18"> The processing of the (indexical) LF of a new utterance always begins with the root node of the current tense tree (structure) in focus. The processing of the top-level operator immediately pushes a token for the surface speech act onto the episode list of the root node. Here is a typical indexical LF: ( decl (past \[John know (That (past (', (perf \[Mary leave\]))))\])) &quot;John knew that Mary had not left.&quot; (decl stands for declarative; its deindexing rule introduces the surface speech act of type &quot;tell&quot;). As mentioned earlier, our deindexing mechanism is a compositional one in which operators past, futr, perf, -,, That, decl, etc., contribute separately to the meaning of their operands. As the LF is recursively transformed, the tense and aspect operators encountered, past, perf and futr, in particular, cause the focus to shift &quot;downward&quot; along existing branches (or new ones if necessary). That is, processing a past operator shifts the current focus down to the left, creating a new branch if necessary.</Paragraph> <Paragraph position="19"> The resulting tense tree is symbolized as /T. Similarly perf shifts straight down, and futr shifts down to the right, with respective results t T and \ T. pres maintains the current focus. Certain operators embed new trees at the current node, written ~--~T (e.g., That), or shift focus to an existing embedded tree, written C/--*T (e.g., decl). Focus shifts to a parent or embedding node are symbolized as T T and .--T respectively. As a final tree operation, OT denotes storage of episode token e T (a focus, as rightmost element of its episode list. As each node comes into focus, its episode list and the lists at certain nodes on the same tree path provide explicit reference episodes in terms of which past, pres, futr, pert, time adverbials, and implicit &quot;orienting&quot; relations are rewritten nonindexically. Eventually the focus returns to the root, and at this point, we have a nonindexical LF, as well as a modified tense tree.</Paragraph> <Paragraph position="20"> Deindexin~ Rules Before we proceed with an example, we show some of the basic deindexing rules here. 6 In the following,&quot;**&quot; is an episodic operator that connects a formula with the situation it characterizes. Predicates are infixed and quantifiers have restrictions (following a colon), r As mentioned earlier, Speaker and Hea~er in the Declrule are to he replaced by the speaker(s) and the hearer(s) of the utterance. Note that each equivalence pushes the dependence on context one level deeper into the LF, thus deindexing the top-level operator. The 6See \[Hwang, 1992\] for the rest of our deindexing rules. Some of the omitted ones are: Fpres ( &quot;futural present,&quot; as in &quot;John has a meeting tomorrow&quot;), Prog (progressive aspect), Pred (predication), K, Ka and Ke (&quot;kinds&quot;), those for deindexing various operators (especially, negation and adverbials), etc.</Paragraph> <Paragraph position="21"> r For details of Episodic Logic, our semantic representation, see \[Schubert and Hwang, 1989; Hwang and Schubert, 1991\].</Paragraph> <Paragraph position="22"> symbols NOWT, Last T and Emb T refer respectively to the speech time for the most recent utterance in T, the last-stored episode at the current focal node, and the last-stored episode at the current embedding node. bet T in the Past-rule will he replaced by either before or same-time, depending on the aspectual class of its first argument and on whether the focal node of T is pastdominated. In the Pert-rule, Last T is analogous to the Reichenbachian reference time for the perfect. The impinges-on relation confines its first argument e T (the situation or event described by the sentential operand of pert) to the temporal region preceding the second argument. As in the case of orients, its more specific import depends on the aspectual types of its arguments. If e T is a stative episode, impinges-on entails that the state or process involved persists to the reference time (episode), i.e., \[e T until LastT\]. If e T is an event (e.g., an accomplishment), impinges-on entails that it occurred sometime before the reference time, i.e., \[e T before LaStT\], and (by default) its main effects persist to the reference time. s An Example To see the deindexing mechanism at work, consider now sentences (ha) and (Ca).</Paragraph> <Paragraph position="23"> (5) a. John went to the hospital.</Paragraph> <Paragraph position="24"> b. (decl Ta (past Tb \[John goto Hospital\] ) ) Tc c. (3 el:tel same-time Now1\]</Paragraph> <Paragraph position="26"> 8We have formulated tentative meaning postulates to this effect hut cannot dwell on the issue here. Also, we are setting aside certain well-known problems involving temporal adverbials in perfect sentences, such as the inadmissibility of * &quot;John has left yesterday.&quot; For a possible approach, see \[Schubert and Hwang, 1990\].</Paragraph> <Paragraph position="27"> The LFs before deindexing are shown in (5,6b) (where the labelled arrows mark points we will refer to); the final, context-independent LFs are in (5,6c). The transformation from (b)'s to (c)'s and the corresponding tense tree transformations are done with the deindexing rules shown earlier. Anaphoric processing is presupposed here.</Paragraph> <Paragraph position="28"> The snapshots of the tense tree while processing (5b) and (6b), at points Ta-Ti, are as follows (with a null initial context).</Paragraph> <Paragraph position="29"> ata atb at C/ atd at e el el el el, e3 el, PS3 * ...... ~ &quot;'&quot;-'. ~ ...... . ...... . .-....--.-. at f at g at h at i el, e3 el, e3 el, e3 el, e3 e2, e4 e e4/(~6 e2, e4 ~: The resultant tree happens to be unary, but additional branches would be added by further text, e.g., a future branch by &quot;It will take several weeks to heal.&quot; What is important here is, first, that Reichenbach-like relations are introduced compositionally; e.g., \[e6 before e5\], i.e., the breaking of the ankle, e6, is before the state John is in at the time of the doctor's talking to him, e4. In addition, the recursive rules take correct account of embedding. For instance, the embedded present perfect in a sentence such as &quot;John will think that Mary has left&quot; will be correctly interpreted as relativized to John's (future) thinking time, rather than the speech time, as in a Reichenbachian analysis.</Paragraph> <Paragraph position="30"> But beyond that, episodes evoked by successive sentences, or by embedded clauses within the same sentence, are correctly connected to each other. In particular, note that the orienting relation between John's going to the hospital, e2, and the doctor's diagnosis, e4, is automatically incorporated into the deindexed formula (6c). We can plausibly particularize this orienting relation to \[e4 after e2\], based on the aspectual class of &quot;goto&quot; and &quot;tell&quot; (see below). Thus we have established inter-clausal connections automatically, which in other approaches require heuristic discourse processing. This was a primary motivation for tense trees. Our scheme is easy to implement, and has been successfully used in the TRAINS interactive planning advisor at Rochester \[Allen and Schubert, 1991\].</Paragraph> <Paragraph position="31"> More on ParCicularizin~ the ORIENTS RelaC/ion The orients relation is essentially an indicator that there could be a more specific discourse relation between the argument episodes. As mentioned, it can usually be particularized to one or more temporal, causal, or other &quot;standard&quot; discourse relation. Existing proposals for getting these discourse relations right appear to be of two kinds. The first uses the aspectual classes of the predicates involved to decide on discourse relations, especially temporal ones, e.g., \[Partee, 1984\], \[Dowty, 1986\] and \[Hinrichs, 1986\]. The second approach emphasizes inference based on world knowledge, e.g., \[Hobbs, 1985\] and \[Lascarides and Asher, 1991; Lascarides and Oberlander, 1992\]. The work by Lascarides et hi. is particularly interesting in that it makes use of a default logic and is capable of retracting previously inferred discourse relations.</Paragraph> <Paragraph position="32"> Our approach fully combines the use of aspectual class information and world knowledge. For example, in &quot;Mary got in her Ferrari. She bought it with her own money,&quot; the successively reported &quot;achievements&quot; are by default in chronological order. Here, however, this default interpretation of orients is reversed by world knowledge: one owns things after buying them, rather than before. But sometimes world knowledge is mute on the connection. For instance, in &quot;John raised his arm. A great gust of wind shook the trees,&quot; there seems to be no world knowledge supporting temporal adjacency or a causal connection. Yet we tend to infer both, perhaps attributing magical powers to John (precisely because of the lack of support for a causal connection by world knowledge). So in this case default conclusions based on orients seem decisive. In particular, we would assume that if e and e' are nonstative episodes, 9 where e is the performance of a volitional action and e' is not, then \[e orients e'\] suggests \[e right-before d\] and (less firmly) \[e cause-of d\]. 1deg</Paragraph> </Section> <Section position="6" start_page="236" end_page="238" type="metho"> <SectionTitle> 4 Beyond Sentence Pairs </SectionTitle> <Paragraph position="0"> The tense tree mechanism, and particularly the way in which it automatically supplies orienting relations, is well suited for longer narratives, including ones with tense shifts. Consider, for example, the following (slightly simplified) text from \[Allen, 1987, p400\]: (7) a. Jack and Sue went{e~} to a hardware store b. as someone had{e~} stolen{~5} their lawnmower.</Paragraph> <Paragraph position="1"> c. Sue had{e4} seen{eh} a man take it 9Non-statives could be achievements, accomplishments, culminations, etc. Our aspectual class system is not entirely settled yet, but we expect to have one similar to that of \[Moens and Steedman, 1988\].</Paragraph> <Paragraph position="2"> 1degOur approach to plausible inference in episodic logic in general, and to such default inferences in particular, is probabilistic (see \[Schubert and Hwang, 1989; Hwang, 1992\]). The hope is that we will be able to &quot;weigh the evidence&quot; for or against alternative discourse relations (as particularizations of orients).</Paragraph> <Paragraph position="3"> d. and had{,,} chased{e,} him down the street, e. but he had{e,} driven{,g} away in a truck.</Paragraph> <Paragraph position="4"> f. After looking{,,o} in the store, they realized{,in} that they couldn't afford{,~} a new one.</Paragraph> <Paragraph position="5"> Even though {b-e} would normally be considered a subsegment of the main discourse {a, f}, both the temporal relations within each segment and the relations between segments (i.e., that the substory temporally precedes the main one) are automatically captured by our rules.</Paragraph> <Paragraph position="6"> For instance, el and ell are recognized as successive episodes, both preceded at some time in the past by e3, es, eT, and eg, in that order.</Paragraph> <Paragraph position="7"> This is not to say that our tense tree mechanism obviates the need for larger-scale discourse structures. As has been pointed out by Webber \[1987; 1988\] and others, many subnarratives introduced by a past perfect sentence may continue in simple past. The following is one of Webber's examples: (8) a. I was{,l} at Mary's house yesterday.</Paragraph> <Paragraph position="8"> b. We talked{,2} about her sister Jane.</Paragraph> <Paragraph position="9"> c. She had{e3} spent{e,} five weeks in Alaska with two friends.</Paragraph> <Paragraph position="10"> d. Together, they climbed{,,} Mt. McKinley.</Paragraph> <Paragraph position="11"> e. Mary askedoe } whether I would want to go to Alaska some time.</Paragraph> <Paragraph position="12"> Note the shift to simple past in d, though as Webbet points out, past perfect could have been used. The abandonment of the past perfect in favor of simple past signals the temporary abandonment of a perspective anchored in the main narrative - thus bringing readers &quot;closer&quot; to the scene (a zoom-in effect). In such eases, the tense tree mechanism, unaided by a notion of higher-level discourse segment structure, would derive incorrect temporal relations such as \[e5 orients e6\] or \[e6 right-after es\].</Paragraph> <Paragraph position="13"> We now show possible deindexing rules for perspective shifts, assuming for now that such shifts are independently identifiable, so that they can be incorporated into the indexical LFs. new-pets is a sentence operator initiating a perspective shift for its operand, and prey-pets is a sentence (with otherwise no content) which gets back to the previous perspective. Recent T is the episode most recently stored in the subtree immediately embedded by the focal node of T.</Paragraph> <Paragraph position="14"> New-pets: (new-pers C/)T *'* \[$,-. T A \[Itecent T orients RecentT,\]\] where T' = $&quot; (~-~ T) When new-pers is encountered, a new tree is created and embedded at the focal node, the focus is moved to the root node of the new tree, and the next sentence is processed in that context. In contrast with other operators, new-pets causes an overall focus shift to the new tree, rather than returning the focus to the original root. Note that the predication \[Recen*c T orients Recen'~T, \] connects an episode of the new sentence with an episode of the previous sentence, prey-pets produces a trivial True, but it returns the focus to the embedding tree, simultaneously blocking the link between the embedding and the embedded tree (as emphasized by use of ~ instead of ~---).</Paragraph> <Paragraph position="15"> We now illustrate how tense trees get modified over perspective changes, using (8) as example. We repeat (Sd,e) below, augmenting them with perspective changes, and show snapshots of the tense trees at the points marked. In the trees, ul,...,u5 are utterance episodes for sentences a,..., e, respectively.</Paragraph> <Paragraph position="16"> T4. Also, note that RecentT1 = e4 and Recenl;T2 = e5. So, by Hew-pets, we get \[e4 orients e5\], which can be later particularized to \[e5 during e4\]. It is fairly obvious that the placement of new-pers and prev-pers operators is fully determined by discourse segment boundaries (though not in general coinciding with them). So, as long as the higher-level discourse segment structure is known, our perspective rules are easily applied. In that sense, the higher-level structure supplements the &quot;fine structure&quot; in a crucial way.</Paragraph> <Paragraph position="17"> However, this leaves us with a serious problem: deindexing and the context change it induces is supposed to be independent of &quot;plausible inferencing&quot;; in fact, it is intended to set the stage for the latter. Yet the determination of higher-level discourse structure--and hence of perspective shifts--is unquestionably a matter of plausible inference. For example, if past perfect is followed by past, this could signal either a new perspective within the current segment (see 8c,d), or the closing of the current subsegment with no perspective shift (see 7e,f). If past is followed by past, we may have either a continuation of the current perspective and segment (see 9a,b below), or a perspective shift with opening of a new segment (see 9b,c), or closing of the current segment, with resumption of the previous perspective (see 9c,d).</Paragraph> <Paragraph position="18"> (9) a. Mary found that her favorite vase was broken.</Paragraph> <Paragraph position="19"> b. She was upset.</Paragraph> <Paragraph position="20"> c. She bought it at a special antique auction, d. and she was afraid she wouldn't be able to find anything that beautiful again.</Paragraph> <Paragraph position="21"> Only plausible inference can resolve these ambiguities. This inference process will interact with resolution of anaphora and introduction of new individuals, identification of spatial and temporal frames, the presence of modal/cognition/perception verbs, and most of all will depend on world knowledge. In (9), for instance, one may have to rely on the knowledge that one normally would not buy broken things, or that one does not buy things one already owns.</Paragraph> <Paragraph position="22"> As approaches to this general difficulty, we are thinking of the following two strategies: (A) Make a best initial guess about presence or absence of new-pers/prevpres, based on surface (syntactic) cues and then use failure-driven backtracking if the resulting interpretation is incoherent. A serious disadvantage would be lack of integration with other forms of disambiguation. (B) Change the interpretation of LaStT, in effect providing multiple alternative referents for the first argument of orients. In particular, we might use Last T = {ei \[ ei is the last-stored episode at the focus of T, or was stored in the subtree rooted at the focus of T after the last-stored episode at the focus of T}.</Paragraph> <Paragraph position="23"> Subsequent processing would resemble anaphora disambiguation. In the course of further interpreting the deindexed LF, plausible inference would particularize the schematic orienting relation to a temporal (or causal, etc.) relation involving just two episodes. The result would then be used to make certain structural changes to the tense tree (after LF deindexing).</Paragraph> <Paragraph position="24"> For instance, suppose such a schematic orienting relation is computed for a simple past sentence following a past perfect sentence (like 8c,d). Suppose further that the most coherent interpretation of the second sentence (i.e., 8d) is one that disambiguates the orienting relation as a simple temporal inclusion relation between the successively reported events. One might then move the event token for the second event (reported in simple past) from its position at the past node to the right-most position at the past perfect node, just as if the second event had been reported in the past perfect. (One might in addition record a perspective shift, if this is still considered useful.) In other words, we would &quot;repair&quot; the distortion of the tense tree brought about by the speaker's &quot;lazy&quot; use of simple past in place of past perfect. Then we would continue as before.</Paragraph> <Paragraph position="25"> In both strategies we have assumed a general coherence-seeking plausible inference process. While it is clear that the attainment of coherence entails delineation of discourse segment structure and of all relevant temporal relations, it remains unclear in which direction the information flows. Are there independent principles of discourse and temporal structure operating above the level of syntax and LF, guiding the achievement of full understanding, or are higher-level discourse and temporal relations a mere byproduct of full understanding? Webber \[1987\] has proposed independent temporal focusing principles similar to those in \[Grosz and Sidnet, 1986\] for discourse. These are not deterministic, and Song and Cohen \[1991\] sought to add heuristic constraints as a step toward determinism. For instance, one constraint is based on the presumed incoherence of simple present followed by past perfect or posterior past. But there are counterexamples; e.g., &quot;Mary is angry about the accident. The other driver had been drinking.&quot; Thus, we take the question about independent structural principles above the level of syntax and LF to be still open.</Paragraph> </Section> class="xml-element"></Paper>