File Information

File: 05-lr/acl_arc_1_sum/cleansed_text/xml_by_section/metho/04/w04-2325_metho.xml

Size: 30,595 bytes

Last Modified: 2025-10-06 14:09:22

<?xml version="1.0" standalone="yes"?>
<Paper uid="W04-2325">
  <Title>Causes and Strategies for Requesting Clarification in Dialogue</Title>
  <Section position="3" start_page="0" end_page="0" type="metho">
    <SectionTitle>
2 Clarification Requests
</SectionTitle>
    <Paragraph position="0"/>
    <Section position="1" start_page="0" end_page="0" type="sub_section">
      <SectionTitle>
2.1 A First Classification
</SectionTitle>
      <Paragraph position="0"> The examples in (1) above together with (5) below illustrate the wide range of CRs that can occur in dialogues, varying with respect to their form (from conventional forms (5-a) to full sentences ((1-d),(5-c-i) and (5-c-ii)), to sentential fragments ((1-a),(1-b),(1-c),(5-b-i) and (5-b-ii))) and with respect to their function (clarifying acoustic understanding ((5-a) and possibly those in  (1)); reference ((1-c),(1-d),(5-b-ii) and possibly (1-a) and (5-b-i)); or pragmatic impact (the examples in (5-c))).</Paragraph>
      <Paragraph position="1"> (5) a. A: Did you talk to Peter? B: Sorry? / Pardon? / You what? b. (i) A: Did you bring a 3-5 torx? B: A what? (ii) A: George Bush is in hospital.</Paragraph>
      <Paragraph position="2"> B: Junior or senior? [from (Gabsdil, 2003)] c. (i) A: What time is it? B: Do you want to leave? (ii) A: Can you pass me the salt? B: Is that a question or a request?  What the questions in these examples have in common is that, unlike &amp;quot;normal&amp;quot; questions, they are not about the state of the world in general, but rather about aspects of previous utterances: they indicate a problem with understanding that utterance, and they request repair of that problem. We take this to be the defining features of CRs. Note that this definition includes correctional uses of CRs as illustrated in (6); in this case the problem is taken to originate on the side of the speaker of the original utter- null ance rather than on the side of the CR initiator. (6) a. A: Dear police men....</Paragraph>
      <Paragraph position="3"> B: Police men? A: Alright then, police people.</Paragraph>
      <Paragraph position="4"> b. Student: 3 + 4 = 8  Teacher: 3 + 4 = 8? We will focus on the possible functions of CRs in this paper, leaving the question of how to map CR form to that function to further work.2 We simply observe at this point that some CRs indicate the kind of understanding problem that occurred; e.g., in (1-c) and (5-b-ii) this seems to be a problem with identifying the intended referent; in (1-d) and (5-b-i) a lexical problem; in (5-c) a problem with recognising the intention behind the utterance. This observation will form the basis of the classification developed below in Section 3, where we further develop extant models to make the pre-theoretic notion of understanding precise.</Paragraph>
      <Paragraph position="5"> Contrasting (1-c) and (5-b-ii) with (7) below illustrates another dimension for classification. Where the former two CRs ask for a confirmation of a hypothesis, the latter asks for a repetition (or reformulation) of the problematic  element.</Paragraph>
      <Paragraph position="6"> (7) A: I talked to Shanti.</Paragraph>
      <Paragraph position="7"> B: WHO?  We call this dimension severity; this represents the intuition that a problem that leads to a request for repetition is more severe than one that leads to a request for confirmation; we will make this notion of severity precise in Section 4. Note that a confirmation request can be realised as an alternative question, as in (5-b-ii) and (4) b, or as a y/n-question, as in (1-c).</Paragraph>
      <Paragraph position="8"> Lastly, we also distinguish between CRs that point out a problematic element in the original utterance, and those that don't. The former category is illustrated by the CRs in (1) and (5-b), the latter by those in (5-a) and (5-c). We call this dimension extent.3 2(Purver et al., 2001) investigates such a mapping, based on the classification discussed below in Section 2.2.  lated to, but different in some aspects from those used in (Larsson, 2003). Our term 'CR' covers what Larsson calls negative feedback as well as what he calls checking feedback, whereas Before we finally come to the description of the dimension level of understanding in the next Section, we will briefly look at an earlier analysis of CR that does not make these distinctions.</Paragraph>
    </Section>
    <Section position="2" start_page="0" end_page="0" type="sub_section">
      <SectionTitle>
2.2 Previous Analyses
</SectionTitle>
      <Paragraph position="0"> In a number of papers (Ginzburg and Cooper, 2001; Purver et al., 2001), Jonathan Ginzburg and colleagues have developed an influential analysis of CR. The authors define two readings that can be ascribed to CRs, which they name the constituent reading and the clausal reading.4 (8-b) shows paraphrases of these readings for the  CR in (8-a).5 (8) a. A: Did Bo leave? -- B: Bo? b. clausal: Are you asking whether Bo left? constituent: Who's Bo?  These readings are defined informally by (Ginzburg and Cooper, 2001) (henceforth G&amp;C) as follows: the clausal reading &amp;quot;takes as the basis for its content the content of the conversational move [[?] speech act, D.S.] made by the utterance being clarified. [It] corresponds roughly to 'Are you asking / asserting that X?', or 'For which X are you asking / asserting that X?'.&amp;quot;; the constituent reading is a reading &amp;quot;whereby the content of a constituent of the previous utterance is being clarified.&amp;quot; Let's look at the conditions under which a dialogue participant (be that a dialogue system or a human) might intend one or the other reading. We begin with the situation shown in (9-a), taken from G&amp;C.</Paragraph>
      <Paragraph position="1">  (9) a. A: Did Bo leave? -- B: Who? b. clausal: For which x are you asking  whether x left? constituent: Who's Bo? severity gives a finer classification of what he calls 'eliciting feedback'. Larsson gives a classification comparable to our extent dimension only indirectly, via a classification of the forms used to express CR as syntactically fragmental or complete; however, since syntactically complete utterances can nevertheless target individual elements (&amp;quot;Which Peter are you talking about?&amp;quot;), these categorisations are not congruent.  namely a lexical reading in which &amp;quot;the surface form of the utterance is being clarified, rather than the content of the conversational move. This reading therefore takes the form 'Did you utter X?'.&amp;quot; (Purver et al., 2001). However, the authors do not offer a formalisation of this reading, so we will concentrate on the two readings for which they do.</Paragraph>
      <Paragraph position="2"> 5These readings are realised technically by a straightforward formalisation of these paraphrases in an HPSG framework, using an illocutionary-act relation for the clausal reading and a relation content for the clausal readings, where both relations take signs as arguments. Since the formalisation is so close to the paraphrases (and is in any case not backed up by a formal semantics of the predicates used), we can use in the following arguments just the paraphrases without missing crucial details. It seems that the clausal reading is appropriate both in situations where A failed to recognise the name acoustically as well as when she failed to resolve the reference. An answer to this question will always resolve both kinds of problems; i.e., this reading does not make a difference between these kinds of understanding problems. The constituent reading, on the other hand, does, and is only appropriate for repetition requests targeting the semantic/pragmatic problem 'reference resolution.' The next example, also from G&amp;C, shows a CR that has the form of a reformulation of the original content. In this case a constituent reading is not available in G&amp;C's analysis, since they postulate a phonologicalidentity condition for constituent readings which is violated here.</Paragraph>
      <Paragraph position="3">  (10) a. A: Did Bo leave? -- B: My cousin? b. clausal: Are you asking whether my cousin left? constituent (not allowed by G&amp;C): Is the  denotation of &amp;quot;my cousin&amp;quot; the denotation of &amp;quot;Bo&amp;quot;? However, in this example the intended distinction between the readings seems to break down, since the function of the clausal reading here is to clarify a hypothesis about the denotation of a referring expression, which in other cases is the function of constituent readings.</Paragraph>
      <Paragraph position="4"> To summarise, G&amp;C's analysis does not explicitly record the problem that leads to the need for clarification. This leads to a loss of information; information which however will be present on the side of the CR producer (who knows where the understanding problem occurred) and presumably should be present on the side of the recipient, who might want to react differently depending on the assumed problem. It seems that the construct &amp;quot;Are you X-ing whether p?&amp;quot; is too general to make these fine-grained distinctions.6  can be illustrated with the following example.</Paragraph>
      <Paragraph position="5"> (i) A: Can you pass me the salt? -- B: The salt?  A's utterance is of course an example of an indirect speech act. Since G&amp;C assume that the illocutionary force of the previous utterance is represented in the CR-reading (&amp;quot;Are you asking/asserting/etc.-ing whether...&amp;quot;) generated by the grammar, they have to find a way to capture this indirectness. There are only two, equally unattractive, options to do this: either the authors have to assume that the grammar directly assigns A's utterance the force request (rather than question), so that the clausal reading of B's utterance can be paraphrased as &amp;quot;Did you request that I pass the salt?&amp;quot;. But interpreting indirect speech acts is a highly context-dependent task and not something that can be decided on syntactical grounds alone. The other option is to stick with the speech act type that is normally associated with interrogatives, and arrive at a reading that can be paraphrased as &amp;quot;Are you asking whether I can pass you the SALT?&amp;quot;. This, however, is presumably in most cases not the right interpretation of  We now turn to exploring a model that does make these distinctions.</Paragraph>
    </Section>
  </Section>
  <Section position="4" start_page="0" end_page="0" type="metho">
    <SectionTitle>
3 A Model of Causes for Clarification
</SectionTitle>
    <Paragraph position="0"/>
    <Section position="1" start_page="0" end_page="0" type="sub_section">
      <SectionTitle>
3.1 The Fundamental Distinctions
</SectionTitle>
      <Paragraph position="0"> Herb Clark (Clark, 1996) and Jens Allwood (Allwood, 1995) independently developed a model of the (hierarchically ordered) tasks involved in communication, as shown schematically in Figure 1.7 (We have also assigned the readings defined by G&amp;C to the appropriate levels in the last column.) This model can serve as a basis for classifying the function of CRs. For example, the CRs shown in (11) can be classified as each targeting a different level according to the model.</Paragraph>
      <Paragraph position="1">  (11) a. [You are sitting in a subway train when A sits down on the seat next to you, talking.</Paragraph>
      <Paragraph position="2"> You might say:] Are you talking to me? b. A: I saw Peter.</Paragraph>
      <Paragraph position="3"> B: What did you say? (I didn't hear you.) c. A: I saw Peter.</Paragraph>
      <Paragraph position="4"> B: Which Peter? d. A: My mother is a lawyer.</Paragraph>
      <Paragraph position="5">  B: Are you trying to threaten me? The distinctions made by this model, however, are still fairly coarse-grained. It seems desirable to further analyse the levels--and especially the third one, that of  'meaning and understanding&amp;quot;--so as to capture for example the difference between the CRs in (12) below. We will do this in the next section.</Paragraph>
      <Paragraph position="6"> (12) a. A: I ate a Pizza with chopsticks the other day.</Paragraph>
      <Paragraph position="7"> B: A Pizza with chopsticks on it? the CR.</Paragraph>
      <Paragraph position="8"> 7There are significant differences between these two models; however, for our purposes the similarities dominate and so we treat the models as terminological variants; keeping in mind that we simplify both models considerably.</Paragraph>
      <Paragraph position="9"> b. A: Please give me the double torx.</Paragraph>
      <Paragraph position="10"> B: What's a torx? c. A: Please give me the double torx.</Paragraph>
      <Paragraph position="11"> B: Which one? d. A: Every wire has to be connected to  a power source.</Paragraph>
      <Paragraph position="12"> B: Each to a different one, or can it be the same for every wire?</Paragraph>
    </Section>
    <Section position="2" start_page="0" end_page="0" type="sub_section">
      <SectionTitle>
3.2 A More Fine-Grained Model
</SectionTitle>
      <Paragraph position="0"> How shall we further carve up the level &amp;quot;meaning &amp; understanding&amp;quot;? One well-known additional distinction that seems useful is that between literal meaning and speaker meaning.8 For instance, this distinction is evoked in the following categorisation given by (Larsson, 2003):  * Semantic Meaning: discourse-independent meaning. E.g. word meanings.</Paragraph>
      <Paragraph position="1"> * Pragmatic Meaning: domain-dependent and discourse-dependent meaning, further split into: - referential meaning, e.g. referents of pronouns, temporal expressions; - pragmatic meaning proper: the relevance of u  in the current context.</Paragraph>
      <Paragraph position="2"> This points in the right direction, but still needs to be made more precise: for instance, what is &amp;quot;relevance&amp;quot; here? Where do the examples in (12) fit in that schema? To make these terms precise, and to add further distinctions, we have devised a model that is closely inspired by how the discourse semantics theory SDRT (Asher and Lascarides, 2003) sets up the syntax/semantics/pragmatics interface. In particular, we use the idea of using semantic underspecification to allow for &amp;quot;pragmatic intrusion&amp;quot; into the determination of the truth value of an utterance; where roughly the underspecified logical form generated by the grammar corresponds to the (set of) literal meaning(s) of an utterance, and the pragmatically resolved LF to the speaker meaning. We also use SDRT's idea of spelling out contextual relevance as the need for determining a rhetorical relation with which to connect a new utterance to the context, and the concept of speech act related goals (SARGs), i.e. goals that are conventionally connected to certain types of speech acts.9 (These ideas will be illustrated with examples below in Section 4 when we discuss the implementation.) This move allows us to say precisely what constitutes an understanding problem in this model--namely, a failure to tackle one (or more) of the precisely defined tasks.  The resulting fine-grained model is shown in Figure 2; the additional levels are further motivated by the examples of CRs shown in (13), which indicate problems on each of these levels (the labels refer to the labels of the (sub-)levels in the figure).10,11 (13) a. 3aa: see e.g. (1-d) or (5-b-i) above, or: A: Peter kowtowed again.</Paragraph>
      <Paragraph position="3"> B: What does &amp;quot;kowtow&amp;quot; mean? b. 3ab: A: The cat that the mouse that the flee bit saw slept.</Paragraph>
      <Paragraph position="4"> B: Who did what? Again, please.</Paragraph>
      <Paragraph position="5"> c. 3ac: A: I saw a man with a telescope.</Paragraph>
      <Paragraph position="6"> B: What do you mean? Did you see a man who was holding a telescope, or did you use a telescope to watch him? d. 3ba: see e.g. (1-c), (12-c) above.</Paragraph>
      <Paragraph position="7"> e. 3bb: see (12-d) above, or: A: I went to the bank yesterday.</Paragraph>
      <Paragraph position="8"> B: As in &amp;quot;monetary institute&amp;quot;, or in &amp;quot;going fishing&amp;quot;? f. 3c: 10Note that despite the rather technical names for the levels this is still a theoretical model of the process of understanding-one, however, that could be implemented in a dialogue system (see (Schlangen et al., 2001) or Section 4 below). Note also that we do not want to make any claims about the psychological status of these postulated levels. All we claim here is that these levels (which are independently motivated in SDRT) are useful for distinguishing types of CRs.</Paragraph>
      <Paragraph position="9"> 11These CRs are not necessarily very natural sounding; the point is just that one can construct CRs that target exactly those postulated levels. The examples are presented here to theoretically motivate those levels. We are currently conducting a detailed corpus analysis to determine the coverage that can be achieved with this model.</Paragraph>
      <Paragraph position="10"> [At court, A being a witness and B the judge.] A: Max fell. John pushed him.</Paragraph>
      <Paragraph position="11"> B: Witness, do you mean that he fell because he was pushed by the de- null fendant? g. 4: (i) A: My mother is a lawyer.</Paragraph>
      <Paragraph position="12"> B: Are you trying to threaten me? Bprime: Why are you telling me that? Bprimeprime: What do you want to say with that? (ii) A: Let's meet next week.</Paragraph>
      <Paragraph position="13"> B: My parents in law are visiting on Tuesday.</Paragraph>
      <Paragraph position="14"> A: So are you saying that Tuesday is good or bad?  To summarise what we have done so far, we have shown a model that can distinguish in a fine-grained manner possible problems during the processing (&amp;quot;understanding&amp;quot;) of an utterance which can lead to the need for clarification.12 This models one of the dimensions for classification which we described in Section 2. What remains to be done is to explain how the problems at each level can be of different severity, leading to either repetition or confirmation requests. This we will do in the next section, where also a general strategy for dealing with processing hypotheses will be discussed.</Paragraph>
    </Section>
  </Section>
  <Section position="5" start_page="0" end_page="0" type="metho">
    <SectionTitle>
4 Clarification Strategies
</SectionTitle>
    <Paragraph position="0"/>
    <Section position="1" start_page="0" end_page="0" type="sub_section">
      <SectionTitle>
4.1 Extending the Concept of Confidence Scores
</SectionTitle>
      <Paragraph position="0"> In current spoken dialogue systems there is a very dominant source for understanding problems: speech recognition (SR). Many existing dialogue systems (cf. e.g. (San-Segundo et al., 2001)) make use of the confidence scores returned by SR systems together with each recognition hypothesis. Based on this value the system can decide whether to reject the hypothesis (which will lead to a repetition request, e.g. &amp;quot;Can you repeat?&amp;quot;, &amp;quot;Pardon?&amp;quot;), whether to confirm it (explicitly with a confirmation request or implicitly), or whether to accept it without generating explicit feedback. This strategy is represented schematically in Figure 3, where the different CSs are distributed over the space of possible confidence values.</Paragraph>
      <Paragraph position="1"> The idea we want to explore here is very straightforward: this concept of confidence score should be extended to all levels of processing. At all processing stages where the system has to rely on &amp;quot;guesses&amp;quot; (nonmonotonic inferences, heuristics, etc.), it should assign a confidence value to its hypothesis. These confidence 12This model is backed up by an independently-motivated formal semantic theory, which however for reasons of space we cannot present here in any detail; cf. (Asher and Lascarides, 2003) for this.</Paragraph>
    </Section>
    <Section position="2" start_page="0" end_page="0" type="sub_section">
      <SectionTitle>
Reject Confirm Accept
</SectionTitle>
      <Paragraph position="0"> values should then be combined in some principled way (for example by taking a weighted average), to determine the Confirmation Strategy (repetition or confirmation request, implicit confirmation, acceptance) and the level of processing that is to be indicated as the primary source of the problem.13 In the simplest case, a system should be able to delay its decision on a CS until processing of (a certain number of) the hypotheses is completed. For example, imagine the case of a travel information system based in the UK. This system might be able to offer flights to Boston (and so should be able to recognise the input &amp;quot;Boston&amp;quot;),  but it also knows that there are no trains from the UK to Boston. So in this case it should be able to make a decision for one of the hypotheses in the following example. (14) Sys: Hello. This is TravelUK. How may I help?  User: I want to go to {Brighton  |Boston} by train.</Paragraph>
      <Paragraph position="1"> The strategy to make use of combined confidence scores could be implemented in many different systems; for example even the rather simple technique of reference resolution via salience lists has an inherent quantitative element that could be used. However, to make the idea more precise, we will in the next section describe an experimental implementation of it in a dialogue system that follows the approach to discourse interpretation described in the previous section. Such a system should ultimately be able to produce the whole range of CRs according to the dimension level of processing as discussed in Section 3.2 as well as along the dimension severity.</Paragraph>
    </Section>
    <Section position="3" start_page="0" end_page="0" type="sub_section">
      <SectionTitle>
4.2 An Experimental Implementation
</SectionTitle>
      <Paragraph position="0"> So far we have said very little about how the theory of discourse semantics alluded to in Section 3.2 tackles its various tasks. In a nutshell, the theory SDRT can be seen as a combination of dynamic semantics (e.g., DRT, (Kamp and Reyle, 1993)) plus (AI-based) pragmatics. In contrast to traditional dynamic semantics, SDRT attempts to represent the pragmatically preferred interpretation of a discourse. The central notion of Discourse Update is for13This is a generalisation of the approach taken for example by (Walker et al., 2000), who use the output of the semantic and pragmatic modules of their dialogue system to dramatically improve the classifier that judges whether a SR hypothesis is correct or not, compared to a classifier that just uses SR-features. mulated in SDRT within a precise nonmonotonic logic, in which one computes the rhetorical relation (or equivalently, the speech act type) which connects the new information to some antecedent utterance. This speech act places constraints on content and the speech act related goals or SARGs; these in turn serve to resolve semantic underspecification. Note that those SARGs are goals that are either conventionally associated with a particular type of utterance or are recoverable by the interpreter from the discourse context; this distinguishes the goals that interact with linguistic knowledge from goals in general.</Paragraph>
      <Paragraph position="1"> The implementation of the theory which we extended for this paper, RUDI (Schlangen et al., 2001; Schlangen, 2003), works in the domain of appointment scheduling (we will refer to the extended version as RUDIclar). It focuses on resolving one particular kind of underspecification, namely that arising from the need to &amp;quot;bridge&amp;quot; definites to their context. To give an example, for (15) the system computes that the &amp;quot;Wednesday afternoon&amp;quot; is 'bridged' via the relation 'next' to the time of utterance: (15) A: What is a good time for you in the next couple of weeks? B: Wednesday afternoon would be good.</Paragraph>
      <Paragraph position="2"> It does this by non-monotonically inferring the rhetorical relation connecting the second to the first utterance (Question-Answer Pair), and using constraints on this relation (roughly, times mentioned in the answer must overlap with that from the question) to resolve underspecification. Before we further describe how the algorithm works, however, we give a couple of examples illustrating the clarification behaviour of the system.14 Example (16) shows an input where the SR component offers two hypotheses that both have to be considered.15 Let's assume that there is no salient Monday the 13th given the dialogue context. In such a situation we want the system to dramatically lower its confidence in the Monday hypothesis, leading to a situation where only the Sunday hypothesis will have to be confirmed, and only implicitly, rather than having to clarify both hypotheses. null 14We should stress that RUDIclar is a proof-of-concept implementation of the model presented here and not a proper dialogue system. Neither is the system actually connected to a speech recogniser--we simulate this by annotating the input with confidence scores--nor does it generate the clarification forms shown in the examples--rather it produces abstract instructions of the form &amp;quot;confirm element x&amp;quot;, &amp;quot;request repetition of element y&amp;quot; etc. which could be used in such a generation. Moreover, RUDI models an overhearer of a conversation, not an actual participant. In RUDIclar this is an overhearer that barges in if it feels the need to clarify something.</Paragraph>
      <Paragraph position="3"> 15At present the system can only handle two alternative hypotheses, where only one constituent may differ.</Paragraph>
      <Paragraph position="4"> - mrs2di - availattach - choose - availantec - speechacts - sa cnstr - sarg - resolve - clarify - update - null B: How about {Sunday &lt;57&gt;  |Monday &lt;43&gt; }? Here it is not the case that one hypothesis is ruled out completely; the difference between the hypotheses here is that one is 'costlier' to maintain. In our approach the dialogue &amp;quot;Let's meet next weekend. - How about Monday?&amp;quot; is just about coherent, under a reading where the second utterance indirectly corrects the plan (a more explicit version of this would be &amp;quot;How about Monday instead?&amp;quot;). This speech act (Plan-Correction), however, is inferred in RUDIclar with a much lower confidence than the more direct speech act Plan-Elaboration that is computed for the Sunday-variant of B's utterance. Hence, the system prefers that latter variant and proceeds accordingly.</Paragraph>
      <Paragraph position="5"> The previous examples combined an ambiguity introduced by the SR module with confidence scores from further levels. However, new ambiguity can also arise during these latter processing stages. In (18) the temporal expression in B's utterance cannot be uniquely resolved (and in this sense the utterance is actually slightly incoherent, since it violates the uniqueness presupposition of definites), and so the system has to clarify the intended reference.</Paragraph>
      <Paragraph position="6"> (18) A: Let's meet this weekend.</Paragraph>
      <Paragraph position="7"> B: How about at 3pm? RUDI: 3pm on Saturday or 3pm on Sunday? The last example, (19), shows another source for quantified hypotheses: resolving any temporal expression other than &amp;quot;tomorrow&amp;quot; to the next day is dispreferred in the system, and so its confidence in this resolution is lowered and it has to be clarified.</Paragraph>
      <Paragraph position="8"> (19) A [on a Friday]: Let's meet this weekend.</Paragraph>
      <Paragraph position="9"> B: How about Saturday? RUDI: You mean tomorrow? We now sketch how the system works. Reflecting the modularity of the underlying theory, RUDI(clar) divides the update process into several stages (shown schematically in Figure 4). The initial module mrs2di postprocesses the semantic representation provided by the grammar, for example by including underspecified bridging relations for definites. RUDIclar allows logical forms to be annotated with confidence values (following an approach similar to that of (Gabsdil and Bos, 2003), associating the confidence values with labels in an underspecified LF), and it allows alternative hypotheses as input (only two at present). In this way we can represent in the system a situation where the speech recogniser cannot make a decision, as in (16) or (17) above.</Paragraph>
      <Paragraph position="10"> At the next stage, an utterance in the context is chosen to which the current one can be attached via a rhetorical relation, and this in turn determines which antecedents for bridging are available. (Should this choice turn out to lead to failure in successive modules, the system can backtrack and choose another attachment site.) The speech act(s) of the current utterance is (are) then inferred non-monotonically (if there is more than one hypothesis coming from the previous step, this is done for each of them) from information about the antecedent and the current utterance and axioms for each relation. The next module, sa cnstr, tests whether certain constraints on the meaning of the speech acts are satisfied by the utterances that are being connected. After this, the SARGs are computed and any remaining underspecification is resolved. null Finally, a new module clarify compares (if there is more than one) and scores the hypotheses, assigning scores for the bridging decisions and for the speech-act inferences. Some of the rules used here are shown in  on the resulting score, the module decides on whether to launch into a clarification sub-dialogue, and if so, which clarification strategy to follow. (We set the thresholds for this and the weights for the average manually to achieve the behaviour described here; see discussion below.) The level at which RUDIclar targets the clarification is always the lowest one where there was a problem; i.e., where alternative hypotheses were introduced, or where no result could be computed. For instance, in (17) this would be the SR level rather than the speech act level.</Paragraph>
      <Paragraph position="11"> This system is capable of producing flexible CRs that adapt to the dialogue context, and this shows the value of the idea of modelling in a fine-grained way sources of CRs and of extending the concept of confidence scores.</Paragraph>
      <Paragraph position="12"> However, the system is only a first proof-of-concept, and we discuss possible improvements in the next section.</Paragraph>
    </Section>
  </Section>
class="xml-element"></Paper>
Download Original XML