File Information

File: 05-lr/acl_arc_1_sum/cleansed_text/xml_by_section/metho/89/j89-1001_metho.xml

Size: 67,253 bytes

Last Modified: 2025-10-06 14:12:17

<?xml version="1.0" standalone="yes"?>
<Paper uid="J89-1001">
  <Title>PARSING WITH FLEXIBILITY~ DYNAMIC STRATEGIES~ AND IDIOMS IN MIND</Title>
  <Section position="4" start_page="0" end_page="0" type="metho">
    <SectionTitle>
2 ENCODING LINGUISTIC KNOWLEDGE
</SectionTitle>
    <Paragraph position="0"> We shall now begin describing our parser, called WEDNESDAY 2 and implemented in Interlisp-D on a Xerox AI-Processor. In the first place we shall give a brief description of how linguistic knowledge is encoded. null WEDNESDAY 2 is based on the assumption that most linguistic knowledge can be conveniently distributed through the lexicon. The lexicon includes in a compact form a lot of information that is normally stored in a grammar. This information includes semantics, syntactic &amp;quot;static&amp;quot; specifications such as category, mood and tense, case marking, the specification of the relation between semantic objects and syntactic objects, such as grammatical functions, but also some more detailed specifications of constraints in binding these objects to other objects that can occur in the sentence. These specifications describe in disjunctive form sets of features that include different aspects: beside the usual ones (syntactic category, generic features, such as gender, number, person, case marking), relative positions of the entities in the string combined with a measure of likelihood of these positions, cospecification of co-occurring phenomena (such as the equinp in a subordinate clause), and measures of likelihood of the obligatoriness of the actual finding of one such entity to be unified with. A characteristic aspect of lexical representation in WEDNESDAY 2 is that semantic objects are part of the value of linguistic functions. null The only centralized component of the system, in the form of a simple (recursive) transition network, deals with aspects of distribution of constituents. The treatment of rather free word order languages becomes quite natural within this approach, and, in particular, the centralized network then becomes very simple. The important thing to note is that the network does not support operations of structure building (like in an ATN: Woods 1970) nor free annotations, including occurrence of equations (such as in LFG rules). It is the Computational Linguistics, Volume 15, Number 1, March 1989 3 Oliviero Stock Parsing with Flexibility, Dynamic Strategies, and Idioms in Mind lexicon that bears feature sets: the network merely has the role of restricting the space where unification can be carried on.</Paragraph>
    <Paragraph position="1"> Each entry in the lexicon may include:  1. A semantic representation of the meaning of the  word, called sere-units. The format of the latter is independent of that of other pieces of information and can be of several different kinds. In our actual implementation we use semantic networks. They can be seen as a set of propositions that use semantic predicates and arguments. Of course, every format has its own operational modality. For instance, unification in a semantic network is carried on as a node merging operation: two or more nodes in one or more net shreds collapse into a single node, producing a new net configuration, that nonetheless subsumes the previous ones. So the basic principle of unification is obeyed.</Paragraph>
    <Paragraph position="2">  2. Syntactic data such as: a. Specification of linguistic (grammatical) functions  with indication of objects in the semantic representation as semantic values. In our implementations the objects are nodes of the semantic network. Arbitrary features and a case marking may be specified, in disjunctive form, in relation to the syntactic aspect connected to that object. A relevant aspect in this representation is therefore that grammatical functions are the point of contact between semantics and syntax.</Paragraph>
    <Paragraph position="3"> b. A particular linguistic function that indicates a node (in the semantic net) that would be referred to as the Main node of the first syntactic constituent that includes the word. (Eventually every constituent, even of higher level, is provided with a Main, in the course of parsing.) It is worth noting that the Main constitutes a concept that bears some similarity with the concept of &amp;quot;head&amp;quot; in some linguistic theories. The most relevant difference, and the one that justifies the use of a different term, is that the value of the Main is an object in the semantic representation, integrated with syntactic specifications. c. category d. mood and tense.</Paragraph>
    <Paragraph position="4"> Syntactic data here are supplemented by other dynamic syntactic information, kept separate from them for the sake of clarity. The latter consists of sets of constraints on variables but a good metaphor would be to view them as impulses to connect fragments of semantic information, guided by syntactic constraints. Yet, impulses are in a declarative format and are actually interpreted by the parsing algorithm. Impulses specify characteristics of an entity to be identified in a given search space. They have alternatives (for instance, the word &amp;quot;tell&amp;quot; has an impulse to merge its object node with the Main of either an NP or a subordinate clause). An alternative includes: a contextual condition of applicability, a category, features, case marking, side effects (through which, for example, coreference between sub-</Paragraph>
    <Paragraph position="6"> ject of a subordinate clause and a function of the main clause can be indicated). Impulses may also be directed to a different search space than the normal one. Furthermore, there can be also impulses that do not involve operations on the semantic representation: they set markings or features for linguistic functions or for the Main.</Paragraph>
    <Paragraph position="7"> Measures of likelihood can also be specified a. for one alternative to be considered; b. for which relative position the entity sought should be in with respect to the present word; c. for the overall necessity of finding the entity. These measures processed together with other ones will give a quantitative account of the likelihood of an analysis and will dynamically play a heuristic role.</Paragraph>
    <Paragraph position="8"> For the sake of example, a lexical entry, corresponding to the word prescrisse, a verb in the past tense (encountered in Section 1.1, in the form of a participle), is shown in Figure 1.</Paragraph>
    <Paragraph position="9"> Sem-units specify the semantics; here they consist of a single proposition. As we use a semantic network format, the n's are called nodes. Besides the three arguments of the predicate, there is a node, nl, that stands for the instantiation of the involved predicate. Next we have the likelihood of the word reading (in our case drawn from frequency figures). Category and verb tense are self-explanatory. The Main specification refers to the instantiation node of the only proposition. Three grammatical functions are indicated, each specifying a particular argument of the predicate. Next, we find unification specifications (impulses). Each indicates a set of constraints: the subject function requires, with an obligatoriness of 0.5, the Main of a noun phrase, which, with likelihood 0.9, has to occur to the left of the present word. The number is singular and the case marking nominative. The initial t indicates that no further condition must be obeyed. The object function</Paragraph>
  </Section>
  <Section position="5" start_page="0" end_page="0" type="metho">
    <SectionTitle>
4 Computational Linguistics, Volume 15, Number 1, March 1989
</SectionTitle>
    <Paragraph position="0"> Oiiviero Stock Parsing with Flexibility, Dynamic Strategies, and Idioms in Mind necessarily requires one of the alternatives: a noun phrase, a simple subordinate clause, or an infinitive clause. This last alternative specifies that the infinitive occurs with likelihood 0.1 before the present word (and with likelihood 0.9 after it) and must be marked with a di case. Moreover the subject of the infinitive is to be made coreferent with the oblique object of the present clause, at the same time as object unification.</Paragraph>
    <Paragraph position="1"> We have briefly referred to the only centralized data structures in WEDNESDAY 2 that handle the distribution of constituents and, in operating terms, the restriction of the space where unification can take place. We say that this part is concerned with opening, closing and maintaining search spaces, where entities sought by impulses are possibly found. It is realized with very simple space management transition networks (SMTNs), in which $EXP, a distinguished symbol on an arc, indicates that only the occurrence of something expected by the preceding context (i.e., for which an impulse was set up) may allow the transition. SMTNs can impose generalized linear precedence on labeled substrings. They encode information that could be expressed by rewriting rules, factorizing fragments that are common to different rules more perspicaciously. In particular for Italian, which has a substantially freer word order language than English, such networks are very simple. Only locally do they impose an ordering (for instance, prepositions come before what they mark). More often they only exclude intraposition of a constituent not belonging to the current space. It should be noted also that a set of acceptable verbal moods and tenses may be specified on entry points in the SMTNs.</Paragraph>
    <Paragraph position="2"> For instance, a (main) sentence needs a verb in a finite tense.</Paragraph>
    <Paragraph position="3"> It is worth noting that SMTNs are converted into an internal format that includes also a table of first transition cross references, i.e., for each space type T, F(T), the set of initial states that allow for transition on T, or, recursively on a space type subsumed by a state in F(T).</Paragraph>
    <Paragraph position="4"> For example, F(Det) = {NP,S}, at least. This is in the same spirit with the concept of reachability discussed in Kay (1980).</Paragraph>
  </Section>
  <Section position="6" start_page="0" end_page="0" type="metho">
    <SectionTitle>
3 THE BASIC PARSER
</SectionTitle>
    <Paragraph position="0"> WEDNESDAY 2 is a parser based on an integration of chart and unification. The parser uses linguistic knowledge of the type outlined in Section 2.</Paragraph>
    <Paragraph position="1"> A morphological analyzer (Stock, Castelfranchi, and Cecconi 1986) is actually attached to it and unifies data included in stems and affixes. Furthermore, to be more precise, stems are put in hierarchies of prototypes: the result is that linguistic data are much less redundant than appeared in our introductory description. The process leans heavily on the idea of chart (Kay 1980, Kaplan 1973). Chart parsing is a very general concept for non-deterministic parsing (see, for instance, Thompson 1981 and Ritchie and Thompson 1984), historically  inspired by Earley's (1970) work on CFL parsing. We shall briefly review the main concepts here.</Paragraph>
    <Paragraph position="2"> A chart is a directed graph that represents the state of the analysis. Given the input string, the junctures between words are called vertices and are represented as nodes in the chart. Each vertex has an arbitrary number of arcs, called edges, entering and leaving it. An edge is therefore a link between two vertices. In the classic chart definition there are two possible types of edge: inactive or active. An inactive edge stands for a recognized constituent (the edge spans the words that are included). An active edge represents a partially recognized constituent. For an inactive edge there is a specification of the category of the constituent. In the case of an active edge, a rewriting rule in the grammar and a position on the right hand side of that rule are provided, thus indicating what is still in order to complete the recognition of the constituent. If the rule is R: Co ---&gt; C~... C, there are specifications R and i, with 0 &lt; i &lt; n, where i is the position on the right hand side of rule R. A word in the string is itself represented as an inactive edge connecting two adjoining vertices. An empty active edge is an active edge that spans no words and is therefore represented in the chart as a link cycling over one vertex. It means, at least in top-down parsing, a prediction of the application of a rule.</Paragraph>
    <Paragraph position="3"> It is important to note that edges are only added to the chart, never removed. A new edge may be added in the following ways: 1. Given an active edge A spanning from Va to V b and an inactive edge I spanning from Vb to V e, where A refers to rule R and to position i, and the category of I is just Ci + ~, i + 1-th symbol of the right hand side of R, then a new edge E can be added to the chart, which will span from Va to Vc, and, if Ci + ~ was the last symbol in R, E will be an inactive edge with category equal to the symbol on the left hand side of R; if not, it will be an active edge with rule R and position i + 1.</Paragraph>
    <Paragraph position="4"> 2. Empty active edges are placed at particular points in the chart, according to the general strategy used. If Computational Linguistics, Volume 15, Number 1, March 1989 5 Oliviero Stock Parsing with Flexibility, Dynamic SttZategies, and Idioms in Mind the parser is a top-down one, when, given an active edge with rule R and position i, that has reached the vertex V, there is a rule R' with left hand side equal to Ci + 1, i + 1-th symbol of the right hand side of R, an empty active edge is introduced on the vertex V, with rule R', provided that one such edge is not already present on that vertex. If the parser is a bottom-up one, when, given an inactive edge with category C that has reached the vertex V, there is a rule R' that has C as the first symbol of its right hand side, an empty active edge is introduced on the vertex V, with rule R', provided that one such edge is not already present on that vertex.</Paragraph>
    <Paragraph position="5"> The whole process of parsing aims at getting one or more (if the sentence is ambiguous) inactive edges to span the whole string, with category S the distinguished initial symbol in the grammar.</Paragraph>
    <Paragraph position="6"> One of the great advantages of chart parsing is that, whatever the strategy adopted, work is never duplicated. For example, if, after backtracking, the analysis is continued by extending a different active edge and a vertex is reached from where an inactive edge starts, and if the edge addition rule 1 can be applied, then the analysis takes advantage of previously done partial analysis. Another advantage is that the mechanism is perfectly suited for both bottom-up and top-down parsing, depending only on the form of the addition rule 2. Another point worth mentioning is that the input relation with other levels of analysis is coherent: lexical ambiguity results in the very simple fact that more than one inactive edge is introduced for one ambiguous word. Furthermore, the basic top-down and bottom-up control schemata can be sophisticated in a number of ways, resulting in better performance (smaller number of unnecessary edges introduced in the chart). Wiren (1987) compares a number of proposals.</Paragraph>
    <Paragraph position="7"> In WEDNESDAY 2 the chart is the basic structure in which search spaces are defined* An active edge defines * an operational environment for unification activity* Some notable overall aspects in the WEDNESDAY 2 chart are: Instead of referring to a set of rewriting rules, edges refer to positions that are states in SMTNs. The case of the $EXP arc will be discussed in further detail in Section 4.</Paragraph>
    <Paragraph position="8"> Parsing goes basically bottom-up, with top-down confirmation, with an improvement of the so-called left corner technique, a version of which is referred to in Wiren (1987). When a word edge with category C is added to the chart, its first left cross references F(C) are fetched and, for each of them, an edge is introduced in the chart. These particular edges are called sleeping edges. A sleeping edge S at a vertex V s is &amp;quot;awakened,&amp;quot; i.e., causes the introduction of a normal active edge iff there is an active edge arriving at V s that may be extended with an edge with the category of S. If they are not awakened, sleeping edges play no role at all in the process.</Paragraph>
    <Paragraph position="9"> An agenda is provided. Tasks can be added to the agenda and at every moment a scheduling function can decide the order in which tasks should be performed, in a multiprogramming fashion. The scheduling function can very easily implement depth-first control and breadth-first control, but any kind of control can in principle be included. Tasks are of several kinds, including lexical tasks, extension tasks, insertion tasks, and virtual tasks. A lexical task specifies a possible reading of a word to be introduced in the chart as an inactive edge. An extension task specifies an active edge and an inactive edge that can extend it (together with some more information).</Paragraph>
    <Paragraph position="10"> Insertion tasks will be explained in Section 4.1. A virtual task consists in extending an active edge with an edge displaced to another point of the sentence, according to the mechanism treating long-distance dependencies, which is explained in Section 5.</Paragraph>
  </Section>
  <Section position="7" start_page="0" end_page="0" type="metho">
    <SectionTitle>
4 UNIFICATION
</SectionTitle>
    <Paragraph position="0"> Let us now focus on the actual problem of unification.</Paragraph>
    <Paragraph position="1"> The sentence is La luna splende (&amp;quot;The moon shines&amp;quot;).</Paragraph>
    <Paragraph position="2"> We consider an active edge with category S, that spans the fragment la luna, and an inactive edge with category V, that spans the fragment splende. Unification is performed simultaneously with the application of the basic edge extension rule. The first frame in Figure 3a shows part of the situation in the preexisting active S edge, before the V edge contribution (shown in the dotted frame). In the first frame it is emphasized that there is a candidate for unification. PI01.1 denotes the first argument of proposition P101. An unsatisfied impulse is emphasized in the dotted frame. Two alternative sets of constraints are actually indicated. The frame in Figure 3b shows the situation within the new S edge after execution.</Paragraph>
    <Paragraph position="3"> Let us informally describe the aspect of an edge in WEDNESDAY 2. An edge is composed of a. a structural aspect, that includes a provenience vertex, a destination vertex, a category (corresponding to a space type) and a state in an SMTN; b. a set of syntactic specifications of various kinds; c. a measure of likelihood for this analysis, resulting from the likelihood of partial interpretations and the likelihood of choices performed at the present level; and d. the semantic representation of the fragment, in the form of a semantic network.</Paragraph>
    <Paragraph position="4"> The syntactic specifications include, among others, the Main, other functions, sets of constraints to be satisfied (impulses), and candidates for satisfying these sets of constraints. An explicit list of candidates is useful because the unification activity is not continuous, and therefore a memory of pending situations is needed.</Paragraph>
    <Paragraph position="5"> Candidates emerge from included edges: a typical case is that a Main of an included edge may be taken as a  candidate in the present edge, if the SMTN arc transited for including the edge specifies it. A candidate is composed of a complex data structure, including syntactic features and a reference to a node in the semantic network.</Paragraph>
    <Paragraph position="6"> In the actual implementation there are specialized edge slots for different kinds of impulses, in relation to their range: if they modify a structure in a different search space, or if they &amp;quot;fill an argument&amp;quot; for the present frame, etc. Introducing a new edge also means that these data structures are set consistently.</Paragraph>
    <Paragraph position="7"> The contents of a word edge are established and instantiated by the lexicon (or, more exactly, by the morphological processor and the lexicon)* When a word edge W is used to extend a given edge A, to produce the edge A' based on the configuration of A, then (without considering the further processing described below) its syntactic and semantic contents are inserted in the corresponding slots of A'. When any other inactive edge I is used to extend an active edge A, then (without considering the further processing described below) some of its contents are inserted in A' in this way: if I &amp;quot;plays an argumental role&amp;quot; in that space (information that comes from the SMTN arc transited for accepting I), then the value of its Main is placed among the pending candidates of A'. The likelihood figure of A' is the result of the application of a numeric function to the likelihoods of A' and I. The semantic net shred of I is copied to A'.</Paragraph>
    <Paragraph position="8"> What we shall call the Principle of the Good Clerk governs the unification activity: Before the arrival of the Main (the &amp;quot;Boss&amp;quot;) the processor adopts a lazy attitude, when the Main arrives all possible (unification) activity is done, when the Main is already there an (unification) action is performed as soon as a new event occurs.</Paragraph>
    <Paragraph position="9"> Accordingly, before the Main arrives, an extension of an edge causes very little more than the copying actions indicated above. The only check is for mood and tense to be in accordance with what was possibly expected.</Paragraph>
    <Paragraph position="10"> On the arrival of the Main, all the present impulses that either have only a syntactic aspect (like the case marking introduced by a preposition) or that are imported from subsumed edges (impulses that bind modifiers, in the linguistic sense) must be satisfactorily carried on and, for each candidate an expectation matching its characteristics must be found. If all this does not happen then the new edge A' supposed to be added to the chart is not added: the situation is recognized as a failure.</Paragraph>
    <Paragraph position="11"> After the arrival of the Main, each new candidate must find an impulse to merge with, and each incoming impulse must find satisfaction. Again, if all this does not happen, the new edge A' will not be added to the chart.</Paragraph>
    <Paragraph position="12"> The example in Section 1 informally shows the effect of the principle.</Paragraph>
    <Paragraph position="13"> What are the effects of unifying? The value of the linguistic function gets more precise with a reference to the filler node, a category (if not yet specified), a mark as prescribed by the considered alternative in the impulse and features of the two contractors merged together. If, for one feature, there is a set of values for one or both contractors, then the intersection of the values Computational Linguistics, Volume 15, Number 1, March 1989 7 Oliviero Stock Parsing with Flexibility, Dynamic Strategies, and Idioms in Mind is considered as the value. Of course, if the intersection is empty, and/or if any of the characteristics do not match, the merger is precluded, and unification fails.</Paragraph>
    <Paragraph position="14"> An implementation note: changes are always performed in the active edge, but nonetheless, using specialized data types only selected data need to be copied from underlying edges.</Paragraph>
    <Paragraph position="15"> The actual semantic unification happens in the semantic network, represented as a set of labelled propositions. With every entity x (variable or proposition label), Rx, a list reference pairs (proposition and argument number), is associated. In this way, merging x with y causes the substitution of all the occurrences of x, inferred from Rx, with y and the assignment to Rx of the union (without repetitions) of Rx and Ry. Should there be any specified side effects of a merging type (see Section 2), they are carried out in the same way.</Paragraph>
    <Paragraph position="16"> In languages with a variable order or, for instance, if subject gapping is admitted, possible mergings are not univocally determined by structure or position. Therefore there may actually be more than one complete merging combination for a tentative new edge A'. In this case more new edges must be added to the chart, one for each alternative unification. This establishes the second kind of task mentioned before---an insertion task, with self-explanatory execution behavior.</Paragraph>
    <Paragraph position="17"> By way of example, let us consider the simple sentence: I! saggio arna la ballerina, i.e., &amp;quot;The wise man loves the ballerina&amp;quot;. When the &amp;quot;loves&amp;quot; edge is taken care of in the S active edge that has already included the NP &amp;quot;the wise man&amp;quot;, it happens that two cases must be considered: in one case &amp;quot;the wise man&amp;quot; is the subject, in the other case it is a topicalized object. This is specified in the lexicon with the word, &amp;quot;loves&amp;quot;, and different likelihoods are associated with the two possibilities. Two different edges are therefore proposed: the proposals are in the form of insertion tasks specifying the measure of likelihood of the new proposed edges. Up to that point a single path had been followed, with no commitment before the verb was analyzed. Only as a consequence of the verb introduction are the two hypotheses set up, with all the possible commitments involved in each of them and resulting in stronger constraints in the following analysis.</Paragraph>
    <Section position="1" start_page="0" end_page="0" type="sub_section">
      <SectionTitle>
4.1 NON-DETERMINISM AND DISAMBIGUATION
</SectionTitle>
      <Paragraph position="0"> Various kinds of ambiguities can be present in the linguistic knowledge: a word may be semantically ambiguous; an impulse may have a number of alternatives; a case marking may be ambiguous. On top of this there may be idiomatic interpretations competing with the literal ones (see Section 6). And then, of course, the SMTNs express an infinite set of configurations. When a particular set of nondeterministic unifications are to be carried out, an insertion task is introduced in the agenda. The execution of the task leads to the inclusion of a new edge in the chart. Dynamically, apart from the general behavior of the parser, there are some particular restriction,; on its nondeterministic behavior which bring syntactic dynamic disambiguation to bear.</Paragraph>
      <Paragraph position="1">  1. The $EXP arc allows for a transition only if the cotffiguration in the active edge includes an impulse to link up with the Main of the proposed inactive edge.</Paragraph>
      <Paragraph position="2"> 2. The sleeping edge mechanism prevents spaces not appropriate to the left context from being established. null 3. A search space can be closed only if no impulse specified as having to be satisfied remains. In other words, if in a state with an outgoing EXIT arc, an active edge can lead to the establishment of an inactive edge only if there are no obligatory impulses left.</Paragraph>
      <Paragraph position="3"> 4. A proposed new edge A' with a verb tense not  matching the expected values causes a failure, i.e., A' will not be introduced in the chart.</Paragraph>
      <Paragraph position="4"> 5. As set out in greater detail in the previous paragraph, failure is caused by inadequate mergings, with relation to the presence, absence, or ongoing introduction of the Main.</Paragraph>
      <Paragraph position="5"> Making a comparison with the criteria established for LFG for functional compatibility of an f-structure (Kaplan and Bresnan 1982), the following can be said of the dynamics outlined here. Incompleteness recognition performs as specified in (3), and furthermore, there is an earlier check when the Main arrives, in case there were obligatory impulses to be satisfied at that point (e.g., an argument that must occur before the Main). Incoherence is avoided completely after the Main has arrived by means of the $EXP arc mechanism. Before this, it is recognized as specified in (5) above, and causes an immediate failure. Inconsistency is detected as indicated in (4) and (5). As far as (5) is concerned, though, the attitude is to activate impulses when the right premises are present and to look for the right thing and not to check if what was done is consistent.</Paragraph>
      <Paragraph position="6"> One hitherto fundamental, but only partially implemented, aspect is a mechanism for semantic disambiguation that would interact with WEDNESDAY 2. In fact, some aspects of the parser have been designed precisely in view of such interaction.</Paragraph>
    </Section>
  </Section>
  <Section position="8" start_page="0" end_page="0" type="metho">
    <SectionTitle>
5 LONG-DISTANCE DEPENDENCIES
</SectionTitle>
    <Paragraph position="0"> Sentences like &amp;quot;The programmer with the brother of whose colleague John said that Mary danced has left&amp;quot; require that something positioned far away in the sentence (in this example, &amp;quot;with the brother of whose colleague&amp;quot;) plays the role of filler for a certain gap (in this case bound to &amp;quot;dance&amp;quot;). Constructions of this type are called long-distance dependencies. It is generally difficult to relate them back to a context-free mechanism unless some particular device, such as metarules in GPSG, is introduced. In LFG, double arrows are used to denote this and are combined with particular rewriting rules in the grammar that derive c-structures.</Paragraph>
  </Section>
  <Section position="9" start_page="0" end_page="0" type="metho">
    <SectionTitle>
8 Computational Linguistics, Volume 15, Number 1, March 1989
</SectionTitle>
    <Paragraph position="0"> Oliviero Stock Parsing with Flexibility, Dynamic Strategies, and Idioms in Mind These rules have the empty element, e, on the right-hand side, and therefore gaps are foreseen by the grammar writer.</Paragraph>
    <Paragraph position="1"> We take a different approach. First, we do not want the parsing process to be overwhelmed by the appearance of spurious constituents, where a rule that prescribes the introduction of the empty symbol was applied. Of course, the use of empty rewriting rules is attractive to the linguist, because he/she can then give a clear description of the structures. But in parsing we have a problem of the explosion of possible constituents. Second, we want the lexicon to control also the problems and idiosyncrasies involved in limiting the allowed dependencies. To illustrate this point, the solutions given by Government Binding theory and by lexical-functional grammar are worth recalling. The former approach considers wh-movements as constrained by 'among other thing' the principle of subjacency, which is a universal structural principle that imposes bounding categories. The latter approach allows imposition of hounding nodes, which is a more flexible concept (see Bresnan 1982).</Paragraph>
    <Paragraph position="2"> We shall give here a very concise description of how relative clauses are dealt with in WEDNESDAY 2. Let us begin with lexical data. Entries like wh-words carry the information of being the focus of a relative clause, complemented by two kinds of constraints: a. constraints on the path that may be followed upwards before reaching the top-level space of the relative clause. The effect of this in the above example is to allow precisely the following structure: \[S'\[PP with \[NP the brother\[pp Of\[Np whose colleague\]\]\]\] .... \] and cause the main node of the NP &amp;quot;the programmer&amp;quot; to be referenced as the second argument of the relation &amp;quot;colleague&amp;quot;; b. constraints on the path that may be followed downwards from the implied S' space. The effect of this is that the impulse bound to &amp;quot;with&amp;quot; will be allowed to operate in the above example, modifying either &amp;quot;said&amp;quot;, or &amp;quot;danced&amp;quot;, while irregular sentences will be detected.</Paragraph>
    <Paragraph position="3"> We have so far discussed the effects before mentioning the other data implied in the matter: SMTNs. Let us introduce virtual arcs to the network. This is a concept that may sound reminiscent of ATNs (Woods 1970).</Paragraph>
    <Paragraph position="4"> Actually, the role is similar but the concept is quite different. In ATNs a virt arc (a.k.a. retrieve) specifies the particular kind of phrase that should be retrieved from an implicit register called hold. Of course, that particular kind of phrase (for instance, NP or PP) is specified in the central network and it becomes quite unnatural to deal with word idiosyncrasies. The grammar writer must describe all kinds of complex gappings that can occur with different words. WEDNESDAY 2's philosophy reverses the procedure. A virtual arc has nothing specified on it: it can be transited if there is an impulse or a candidate that, although displaced, has been made available by the parser's mechanism, and now is compatible with the situation determined by the Computational Linguistics, Volume 15, Number 1, March 1989  configuration of the present particular words. The virtual arc is transited if unification can be carried out. As an example of a fragment of SMTN that includes a virtual arc, see Figure 4.</Paragraph>
    <Paragraph position="5"> The cycling arc specifies that transitions can be made on whatever is expected by the left context, plus modifiers such as PPs or adverbials. The virtual arc indicates that a proposed displaced element compatible with present functional expectations may be used. Note that virtual arcs are basically inserted in the SMTNs after the Main has been taken care of and therefore unification is immediate. The implementation of this aspect of the parser requires the alteration of the basic chart with the inclusion of topicalized edges, i.e., edges that have a displaced candidate or impulse in evidence, which, by obeying the constraints discussed above, may find its place within that particular edge. If this happens and the parser manages to build an inactive edge, it will be an inactive edge that satisfies the topic. If the inactive edge, and all the inactive edges subsumed by it, do not make use of the topic, the topic simply does not appear at all in the edge. Of course, a necessary condition for the introduction of an inactive top-level topicalized edge (e,g., an edge corresponding to the search space S') is that it satisfies the topic.</Paragraph>
    <Paragraph position="6"> In our example, given that in the fragment \[s,\[ppwith the brother of whose colleague\]\[sJohn said that\[sMary danced\]\] ..... the inactive edge of \[sMary danced\] satisfies the topic, also \[s John said that\[sMary danced\]\] satisfies the topic, and therefore an S' inactive edge can be happily introduced in the chart. The mechanism discussed above guarantees one aspect that we have stressed throughout our work: the dynamic unification of partial interpretations. Furthermore, it seems to work quite well with Italian.</Paragraph>
  </Section>
  <Section position="10" start_page="0" end_page="10" type="metho">
    <SectionTitle>
6 IDIOMS
</SectionTitle>
    <Paragraph position="0"> Idioms are a pervasive phenomenon in natural languages. Linguists have proposed different accounts for idioms, which are derived from two basic points of view: one point of view considers idioms as the basic units of language, with holistic characteristics, perhaps including words as a particular case; in the other point Oiiviero Stock Parsing with Flexibility, Dynamic Strategies, and Idioms in Mind of view the emphasis is instead laid on the fact that idioms are made up of normal parts of speech that play a definite role in the complete idiom. An explicit statement in this approach is the Principle of Decompositionality (Wasow, Sag, and Nunberg 1982): &amp;quot;When an expression admits analysis as morphologically or syntactically complex, assume as an operating hypothesis that the sense of the expression arises from the composition of the senses of its constituent parts&amp;quot;. The syntactic consequence is that idioms are not a different thing from &amp;quot;normal&amp;quot; forms.</Paragraph>
    <Paragraph position="1"> We hold the latter view. We are aware of the fact that the flexibility of an idiom depends on how recognizable its metaphorical origin is. Within flexible word order languages the flexibility of idioms seems to be even more closely linked to the strengths of particular syntactic constructions.</Paragraph>
    <Paragraph position="2"> Let us now briefly discuss some computational approaches to idiom understanding. Applied computational systems must necessarily have a capacity for analyzing idioms. In some systems there is a preprocessor delegated to the recognition of idiomatic forms. This preprocessor replaces the group of words that make for one idiom with the word or the words that convey the meaning involved. In ATN systems, specially if oriented towards a particular domain, there are sometimes instead sequences of particular arcs inserted in the network, which, if transited, lead to the recognition of a particular idiom (e.g., PLANES, Waltz 1978). LIFER (Hendrix 1977), one of the most successful applied systems, was based on a semantic grammar, and within this mechanism idiom recognition was easy to implement, without considering flexibility. Of course, there is no intention in any of these systems to give an account of human processing. PHRAN (Wilensky and Arens 1980) is a system in which maximum emphasis is placed on the role of idioms. A particularly interesting aspect is that the system had also a twin generator, called PHRED (Jacobs 1985), that reversed the process, as both are based on the same linguistic representation.</Paragraph>
    <Paragraph position="3"> Idiom recognition, following Fillmore's (1979) view, is considered the basic resource all the way down to replace the concept of grammar-based parsing. PHRAN is based on a data base of patterns (including single words, at the same level) and proceeds deterministically, at least in its basic implementation, applying the two principles &amp;quot;when in doubt choose the more specific pattern&amp;quot; and &amp;quot;choose the longest pattern&amp;quot;. The limits of PHRAN lie in the capacity to generate various alternative interpretations in case of ambiguity (though it must be reported that a nondeterministic implementation was also eventually realized) and in running the risk of having an excessive spread of nonterminal symbols if the data base of idioms is large. A recent work on idioms with a similar perspective and with particular attention to the problem of learning idioms is Dyer and Zernik (1986).</Paragraph>
    <Paragraph position="4"> The approach we have followed is different. The  goals pursued in our work must be stated explicitly: 1.</Paragraph>
    <Paragraph position="5"> to yield a cognitive model of idiom processing; and 2. to integrate idioms in our lexical data merely as further information concerning words (as in a traditional dictionary). Idiom understanding is based on normal syntactic analysis with word-driven recognition in the background. When a certain threshold is crossed by the weight of a particular idiom, the latter starts a process of its own, that may eventually lead to a solution. Some of the questions we have dealt with are: How are idioms to be specified? When are they recognized? What happens when they are recognized? And what happens afterwards? null Note that a morphological analyzer, WED-MORPH (Stock et al. 1986), linked to WEDNESDAY 2, plays a substantial role, especially if the language is Italian.</Paragraph>
    <Section position="1" start_page="10" end_page="10" type="sub_section">
      <SectionTitle>
6.1 SPECIFICATION OF IDIOMS IN THE LEXICON
</SectionTitle>
      <Paragraph position="0"> Idioms are introduced in the lexicon as further specifications of words, just as in a normal dictionary. They may be of two types: a. canned phrases, that just behave as several-word entries in the lexicon (there is nothing particularly interesting in that, so we shall not go into detail here); b. flexible idioms; these idioms are described in the lexicon bound to the particular word representing the &amp;quot;thread&amp;quot; of that idiom; in WEDNESDAY 2 terms, this is the word that bears the Main of the immediate constituent including the idiom. Thus if we have an idiom like &amp;quot;to build castles in the air&amp;quot;, it will be described along with the verb &amp;quot;to build&amp;quot;.</Paragraph>
      <Paragraph position="1"> After the normal word specifications, the word may include a list of idiomatic entries. Figure 5 shows a BNF specification of idioms in the lexicon. The symbol + stands for &amp;quot;at least one occurrence of what precedes&amp;quot;. Each idiom is described in two sections: the first one describes the elements that characterize that idiom, expressed coherently with the normal characterization of the word; the second one describes the interpretation, i.e., which substitutions should be performed when the idiom is recognized.</Paragraph>
      <Paragraph position="2"> Let us briefly describe Figure 5. The lexical form indicates whether passivization (which, in our theory, as in LFG, is treated in the lexicon) is admitted in the idiomatic reading. The idiom-stats, describing configurations of the components of an idiom, are based on the basic impulses included in the word. In other words constituents of an idiom are described as particular fillers of linguistic functions or particular modifiers. For example &amp;quot;build castles in the air&amp;quot;, when &amp;quot;build&amp;quot; is in an active form, has &amp;quot;castles&amp;quot; as a further description of the filler of the OBJ function and the string &amp;quot;in the air&amp;quot; as a further specification of a particular modifier that may be attached to the Main node. MORESPECI-FIC, the further specification of an impulse to set a filler for a function includes: a reference to one of the possible alternative types of fillers specified in the normal impulse, a specification that describes the fragment that is to play this particular role in the idiom, and Computational Linguistics, Volume 15, Number 1, March 1989  the weight that this component has in the overall recognition of the idiom. IDMODIFIER is a specification of a modifier, including the description of the fragment and the weight of this component.</Paragraph>
      <Paragraph position="3"> CHANGEIMPULSE and REMOVEIMPULSE allow normal syntactic behavior to be modified. The former specifies a new alternative for a filler for an existing function, including the description of the component and its weight (for instance, the new alternative may be a partial NP instead of a complete NP, as in &amp;quot;take care&amp;quot;), or a NP marked differently from usual). The latter specifies that a certain impulse, specified for the word, is to be considered to have been removed for this idiom description.</Paragraph>
      <Paragraph position="4"> There are a number of possible fragment specifications, including string patterns, semantic patterns, morphological variations, coreferences,tetc.</Paragraph>
      <Paragraph position="5"> Substitutions include the semantics of the idiom, which are supposed to take the place of the literal semantics, plus the specification of the new Main and of the bindings for the functions. New bindings may be included to specify new semantic linkings not present in the literal meaning (e.g., &amp;quot;take care of &lt;someone&gt;&amp;quot;, if the meaning is &amp;quot;to attend to &lt;someone&gt;&amp;quot;, then &amp;quot;&lt;someone&gt;&amp;quot; must become an argument of &amp;quot;attend&amp;quot;).</Paragraph>
    </Section>
    <Section position="2" start_page="10" end_page="10" type="sub_section">
      <SectionTitle>
6.2 IDIOM PROCESSING
</SectionTitle>
      <Paragraph position="0"> Idiom processing works in WEDNESDAY 2 by being integrated in the nondeterministic, multiprocessing-based behavior of the parser. As the normal (literal) analysis proceeds and partial representations are built, impulses are monitored in the background, and a check made for possible idiomatic fragments. Monitoring is carried out only for fragments of idioms not in contrast with the present configuration. A dynamic activation table is introduced with the occurrence of a word that has some associated idiom specification. The occurrence of an expected fragment of an idiom in the table raises the level of activation of that idiom, in proportion to the relative weight of the fragment. If the configuration of the sentence contrasts with one fragment, then the relative idiom is removed from the table. So all the normal processing continues, including the possible nondeterministic choices, the establishing of new processes, etc. The activation tables are included in the edges of the chart.</Paragraph>
      <Paragraph position="1"> When the activation level of a particular idiom crosses a fixed threshold, a new process is introduced, dedicated to that particular idiom. In that process only that particular idiomatic, interpretation is considered.</Paragraph>
      <Paragraph position="2"> So, in the first place, an edge is introduced in which substitutions are carried out; the process will proceed with the idiomatic representation. Note that the process begins at that precise point, with all the previous literal analysis acquired by the idiomatic analysis. The original process goes on as well (unless the fragment that caused the new process is nonsyntactic and peculiar to that idiom alone); the idiom is merely removed from the active idiom table. At this point there are two working processes and it is up to the (external) scheduling function to decide priorities. Relevant points are: a. the idiomatic process may still result in failure: further analysis may not confirm what has been hypothesized as an idiom; and b. a different idiomatic process may start from the literal process at a later stage, when its own activation level crosses the threshold. Overall, this yields all the analyses, literal and idiomatic, with likelihoods for the different interpretations. But it also provides a reasonable model of how humans process idioms. Some psycholinguistic experiments have given support to this view, which seems also compatible with the model presented by Swinney and Cutler (1978).</Paragraph>
      <Paragraph position="3"> Here we have disregarded the situation in which a possible idiomatic form occurs and its role in disambiguating. The whole parsing mechanism in WEDNESDAY 2 is based on dynamic unification, i.e., at every step in the parsing process a partial interpretation is provided; dynamic choices are performed by scheduling the agenda on the basis of the relation between partial interpretation and the context.</Paragraph>
      <Paragraph position="4"> Computational Linguistics, Volume 15, Number 1, March 1989 11</Paragraph>
    </Section>
    <Section position="3" start_page="10" end_page="10" type="sub_section">
      <SectionTitle>
6.3 AN EXAMPLE
</SectionTitle>
      <Paragraph position="0"> As an example, let us consider the Italian idiom prendere il toro per le coma (literally: &amp;quot;to take the bull by the horns&amp;quot;; idiomatically: &amp;quot;to confront a difficult situation&amp;quot;). The verb prendere (&amp;quot;to take&amp;quot;) in the lexicon includes some descriptions of idioms. Figure 6 shows the representation of prendere in the lexicon, but includes only information relevant to our example. The stem representation will be unified with other information and constraints coming from the affixes involved in a particular form of the verb. The first portion of the representation is devoted to the literal interpretation of the word and includes the semantic representation, the likelihood of that reading, and functional information, including the specification of impulses for unification.</Paragraph>
      <Paragraph position="1"> The numbers are likelihoods of the presence of an argument or of a relative position of an argument. The secowd portion, after idioms, includes the idioms involving prendere. In Figure 6 only one such idiom is specified. It is indicated that the idiom can also occur in a passive form (the first t specification) and the specification of the expected fragments is given. The numbers here are the weights of the fragments (the threshold is fixed at 10). The substitutions include the new semantic representation, with the specification of the main node and of the binding of the subject. Note that the surface functional representation will not be destroyed after the substitutions, only the semantic (logical) representation will be recomputed, imposing its own bindings.</Paragraph>
      <Paragraph position="2"> As mentioned, Italian allows great flexibility. Let the input sentence be L'informatico prese per le corna la capra (literally: &amp;quot;The computer scientist took by the horns the goat&amp;quot;). When prese (&amp;quot;took&amp;quot;) is analyzed, its idiom activation table is inserted. When the modifier per le coma (&amp;quot;by the horns&amp;quot;) shows up, the activation of the idiom referred to above crosses the threshold (the sum of the two weights goes up to 12). A new process starts at this point, with the new interpretation unified with the previous interpretation of the subject. Also, semantic specifications coming from the suffixes are reused in the new partial interpretation. The process merely departs from the literal process--no backtracking is performed. At this point we have two processes going on: an idiomatic process, where the interpretation is already &amp;quot;The computer scientist is facing a difficult situation&amp;quot; and a literal process, where, in the background, yet other active idioms monitor the events. In  the form of semantic networks (they correspond to two sets of propositions).</Paragraph>
      <Paragraph position="3"> When the last NP, la capra (&amp;quot;the goat&amp;quot;), is recognized, the idiomatic process fails (it needed &amp;quot;the bull&amp;quot; as object). The literal process yields its analysis, but  Computational Linguistics, Volume 15, Number 1, March 1989 O|iviero Stock Parsing with Flexibility, Dynamic Strategies, and Idioms in Mind another idiom crosses the threshold, starts its process with the substitutions, and immediately concludes positively. This latter, unlikely, idiomatic interpretation is that the computer scientist confused the goat and the horns.</Paragraph>
    </Section>
  </Section>
  <Section position="11" start_page="10" end_page="14" type="metho">
    <SectionTitle>
7 THE EXPLORATIVE ENVIRONMENT
</SectionTitle>
    <Paragraph position="0"> Computer-based environments for the linguist are conceived as sophisticated workbenches, built on AI workstations around a specific parser, where the linguist can try out his/her ideas about a grammar for a certain natural language. In doing so, he/she can take advantage of rich and easy-to-use graphic interfaces that &amp;quot;know&amp;quot; about linguistics. Of course, behind all this lies the idea that cooperation with linguists will provide better results in NLP. To substantiate this assumption it may be recalled that some of the most interesting recent ideas on syntax have been developed by means of joint contributions from linguists and computational linguists. null Instances of the tools introduced above are the LFG environment, which was probably the first of its kind, an environment built by Ron Kaplan for iexical-functional grammars, D-PATR, built by Lauri Karttunen (Karttunen 1986) and conceived as an environment that would suit linguists of a number of different schools all committed to a view of parsing as a process that makes use of a unification algorithm.</Paragraph>
    <Paragraph position="1"> The environment we have developed has a somewhat different purpose. Besides a number of tools for entering data in graphic mode and inspecting resulting structures, it provides a means for experimenting with strategies in the course of the parsing process. We think that this can be a valuable tool for gaining insight into the cognitive aspects of language processing as well as for tailoring the behaviour of the processor when used with a particular sublanguage.</Paragraph>
    <Paragraph position="2"> In this way an attempt can be made to answer basic questions in the course of a nondeterministic approach: what heuristics to apply when facing a certain choice point, what to do when facing a failure point, i.e., which of the pending processes to activate, taking account of information resulting from the failure? Of course this kind of environment makes sense only because the parser it works on has some characteristics that make it a psychologically interesting realization.</Paragraph>
    <Paragraph position="3"> Psychologically motivated parsers may be put in three main categories. First, those that embody a strong claim on the specification of the general control structure of the human parsing mechanism. The authors usually consider the level of basic control of the system as the level they are simulating and are not concerned with more particular heuristics. An instance of this class of parsers in Marcus's parser (Marcus 1979), based on the claim that, basically, parsing is a deterministic process: only sentences that we perceive as surprising (the so-called &amp;quot;garden paths&amp;quot;) actually imply back-Computational Linguistics, Volume 15, Number 1, March 1989 tracking. Connectionist parsers are also instances of this category. The second category refers to general linguistic performance notions such as the Lexical Preference Principle and the Final Argument Principle (Fodor, Bresnan, and Kaplan 1982). It includes theories of processing like the one expressed by Wanner and Maratsos for ATNs in the mid-'70s. In this category the arguments are at the level of general structural preference analysis. A third category tends, at every stage of the parsing process, to consider the full complexity of the data and the hypothetical partial internal representation of the sentence, including, at least in principle, interaction with knowledge of the world, aspects of memory, and particular task-oriented behavior. Worth mentioning here are Church and Patil (1982) and Barton et al. (1987), who attempt to put some order in the chaos of complexity and computational load.</Paragraph>
    <Paragraph position="4"> Our parser lies between the second and the third of the above categories. The parser is seen as a non-deterministic apparatus that disambiguates and gives a &amp;quot;shallow&amp;quot; interpretation and an incremental functional representation of each processed fragment of the sentence, and allows choices to be made amOng alternatives in a cognitively meaningful way. We shall briefly describe the environment and, by way of example, illustrate its behavior by analyzing oscillating sentences, i.e., sentences in which one first perceives a fragment in one way, then changes one's mind and takes it in a different way, then, as further input comes in, goes back to the previous pattern (and possibly continuing like this till the end of the sentence).</Paragraph>
    <Section position="1" start_page="10" end_page="14" type="sub_section">
      <SectionTitle>
7.1 OVERVIEW OF THE ENVIRONMENT
</SectionTitle>
      <Paragraph position="0"> WEDNESDAY 2 and its environment are implemented on a Xerox Lisp Machine. The coverage of WEDNESDAY 2 with the present Italian data includes subordinates, complex relative clauses, interrogatives, and so on, all this possibly mixed with idioms. The current lexicon is about 1,500 stems, resulting in about 30,000 forms, but it is very easy to extend it, with the tools built around the parser. An effort is currently being made to revisit linguistic specifications in the lexicon.</Paragraph>
      <Paragraph position="1"> Moreover, we have been reversing the process for the purpose of generation (Caraceni and Stock 1987). We think that, beside the interest of generation per se and the usefulness of having a generator based on the same principles and data of the parser, this is, paradoxically, the real test for our parser. The results, so far, have been very encouraging, including the fact that it was not hard to make good the deficiencies we detected in our parser when dealing with problems the other way round.</Paragraph>
      <Paragraph position="2"> The environment is composed of a series of specialized tools, each one based on one or more windows (Figure 8).</Paragraph>
      <Paragraph position="3"> Using a mouse the user selects a desired behavior from menus attached to the windows. We have the following windows:  ,, , .... &amp;quot; * ?&amp;quot; *: ';': : :',';'i'&amp;quot; &amp;quot;' '. '( ' ':'.. ,&amp;quot;, '.' '.'.' :.:.: :.:.:.:.'.'.:+:.:.'+:'..:.: : '.:.:.:+:::':':.:':':: '.':&amp;quot; ;';';'; : :'&amp;quot;: : :'&amp;quot; &amp;quot;&amp;quot;'~ '. &amp;quot; =rf-~lr,l~itl,~f~Q ffc, r'f-, n,~l~le ~,~' t,~a~L~ to t&amp;quot;lo(~e ~{~ pF~El~rd/tt II~j ~F: i L 'r' i.,:11,= .... ~1 F * r :iii:i!iii!!ii:.ii!ii!ii:!:iii!!i~ii:iiii:~il I i i:::i i;:iiii!'!iiii !:.~' ;;' re a c .' * d u ~ el* , t~'~ * f o o., n c~ .'a t u * .- .~n ...... -:::= = =~,'~&amp;quot;--- .- :?~i!i!iiiii~iiii!iii!i ?.i i!i!!i :i'!iiii i 51!ii!i ~ii: : :i :::::: :::!:: !:~::~ i::.</Paragraph>
      <Paragraph position="5"> * the main WEDNESDAY 2 window, in which the sentence is entered. Menus attached to this window specify different modalities (including &amp;quot;through&amp;quot; and &amp;quot;stepping&amp;quot;, &amp;quot;all parsings&amp;quot; or &amp;quot;one parsing&amp;quot;) and a number of facilities; * a window where one can view, enter, and modify transition networks graphically (Figure 9).</Paragraph>
      <Paragraph position="6"> * a window where one can view, enter and modify the lexicon. As a word reading is a complex object for WEDNESDAY 2, entering a new word can be greatly facilitated by a set of subwindows, each specialized in one aspect of the word, &amp;quot;knowing&amp;quot; what it may be like and facilitating editing.</Paragraph>
      <Paragraph position="7"> The lexicon is a lexicon of stems: a morphological analyzer and a lexicon manager are integrated in the system. Let us briefly describe this point. A lexicalist theory such as ours requires that a large quantity of information be included in the lexicon. This information has different origins: some comes from the stem and  some from the affixes. All the information must be put into a coherent data structure, through a particularly constrained unification-based process. Furthermore, we must emphasize the fact that, just as in LFG, phenomena such as passivization are treated in the lexicon (the Subject and Object functions and the related impulses attached to the active form are rearranged). This is something that the morphological analyzer must deal with. The internal behavior of the morphological analyzer is beyond the scope of the present paper. We shall instead briefly discuss the lexicon manager, the role of which will be emphasized here.</Paragraph>
      <Paragraph position="8"> The lexicon manager deals with the complex process of entering data, maintaining, and preprocessing the lexicon. One notable aspect is that we have arranged the lexicon on a hierachical basis according to inheritance, so that properties of a particular word can be inherited from a word class and a word class can inherit aspects from another class. One consequence of this is that we Computational Linguistics, Volume 15, Number 1, March 1989 Oiiviero Stock Parsing with Flexibility, Dynamic Strategies, and Idioms in Mind Tr~o:,*tions from noO~. 14P. AIZ'J tO no@- ~ I4P 11 )rb. )lt,:,=~.eeJ Ul'lcler thff followirl~l '~ ;)10.1~5;  can introduce a graphic aspect (Figure I0) and the user can browse through the lattice (the lexicon appears as a tree of classes where one has specialized editors at each level). What is even more relevant is the fact that one can factorize knowledge that is in the lexicon, so that if one particular phenomenon needs to be treated differently, the change of information is immediate for the words concerned. Of course, this also means that there is a space gain: the same information does not need to be duplicated--complete word data are reconstructed when required.</Paragraph>
      <Paragraph position="9"> There is also a modality by which one can enter the syntactic aspects of a word through examples, ~ la TEAM (Grosz 1984).</Paragraph>
      <Paragraph position="10"> * a window showing the present configuration of the chart; * a window that allows zooming into one edge, showing several aspects of the edge, including its structural aspects, its likelihood, the functional aspect, the specification of unrealized impulses, etc.</Paragraph>
      <Paragraph position="11"> * a window graphically displaying the semantic interpretation of an edge as a semantic net or, if one prefers (this is usually the case when the net is too complex), in logical format; * a window where one can manipulate the agenda (Figure I 1).</Paragraph>
      <Paragraph position="12"> Attached to this window we have a menu including a set of functions that the tasks included in the agenda to be manipulated: &amp;quot;move before&amp;quot;, &amp;quot;move after&amp;quot;, &amp;quot;delete&amp;quot;, &amp;quot;switch&amp;quot;, &amp;quot;undo&amp;quot;, etc. One just points to the two particular tasks one wishes to operate on with the mouse and then to the menu entry, thus obtaining the desired effect. The same effect could be obtained by applying a different scheduling function: the tasks will be picked up in the order here prescribed by manual scheduling. This tool, when the parser is in the stepping modality, provides a very easy way of altering the default behavior of the system and of trying out new strategies. The manual scheduling mode is supplemented by a series of counters that provide control over the penetrance of these strategies. (The penetrance of a nondeterministic algorithm is the ratio between the  steps that lead to the solution and the overall steps carried out in trying to obtain the solution. Of course this measure lies between 0 and 1).</Paragraph>
      <Paragraph position="13"> Dynamically, it is attempted to find sensible strategies, by interacting with the agenda. When, after formalizable heuristics have been tried out, they can be introduced permanently into the system through a given specialized function. This is the only place where some knowledge of Lisp and of the internal structure of WEDNESDAY 2 is required.</Paragraph>
    </Section>
  </Section>
  <Section position="12" start_page="14" end_page="16" type="metho">
    <SectionTitle>
7.2 AN EXAMPLE OF EXPLORATION: OSCILLATING
SENTENCES
</SectionTitle>
    <Paragraph position="0"> We shall now briefly discuss a processing example that we have been able to understand using the environment described above. The following example is a good instance of flexibility and parsing problems present in Italian: A Napoli preferisco Roma a Milano.</Paragraph>
    <Paragraph position="1"> The complete sentence reads &amp;quot;while in Naples I prefer Rome to Milan&amp;quot;. The problem arises during the parsing process with the fact that the &amp;quot;to&amp;quot; argument of &amp;quot;prefer&amp;quot; in Italian may occur before the verb, and the locative preposition &amp;quot;in&amp;quot; is a, the same word as the marking preposition corresponding to &amp;quot;to&amp;quot;.</Paragraph>
    <Paragraph position="2"> The reader/hearer first takes a Napoli as an adverbial location, then, as the verb preferisco is perceived, a Napoli is clearly reinterpreted as an argument of the verb (in the sense of surprise). As the sentence proceeds after the object Roma, the new word a causes things to change again and we go back with a feeling of surprise to the first hypothesis. When this second reconsideration takes place, we feel the surprise, but this does not cause us to reconsider the sentence, we only go back to adding further to an hypothesis we were already working at. It should also be noted that the surprise seems to be caused not by a heavy computational load, but by a  sudden readjustment of the weights of the hypotheses.</Paragraph>
    <Paragraph position="3"> In a sense it is a matter of memory, rather than computation.</Paragraph>
    <Paragraph position="4"> We have succeeded in getting WEDNESDAY 2 to perform naturally in such situations, taking advantage of the environment. The following simple heuristics were found: a. try solutions that satisfy the impulses (if there are alternatives consider likelihoods); b. maintain viscosity (prefer the path you are already following); and c. follow the alternative that yields the edge with the greatest likelihood chosen among edges of comparable length.</Paragraph>
    <Paragraph position="5"> The likelihood of an edge depends on: 1. the likelihood of the included edges; 2. the level of obligatoriness of the filled impulses; 3. the likelihood of a particular relative position of an argument in the string; 4. the likelihood of that transition in the network, given the previous transition.</Paragraph>
    <Paragraph position="6"> The critical points in the sentence are the following (note that we distinguish between a PP and a &amp;quot;marked NP&amp;quot; possible argument of a verb, where the preposition has no associated semantics): i. At the beginning: only the PP edge is expanded, (not the one including a marked NP) because of static preference for the former expressed in the lexicon and in the transition network.</Paragraph>
    <Paragraph position="7"> ii. After the verb is detected: on the one hand there is an edge that, if extended, would not satisfy an obligatory impulse, on the other hand, one that possibly would. The marked NP alternative is chosen because of (a) of the above heuristics.</Paragraph>
    <Paragraph position="8"> iii. After the object Roma: when the preposition a comes in, the edge that can extend the sentence with a PP on the one hand, and on the other hand a cycling active edge that is a promising satisfaction for an impulse are compared. Since this relative position of the argument is so favorable for the particular verb preferisco (.9 to. I for this position compared to the preceding one), the parser proceeds with the alternative view, taking a Napoli, as a modifier, and so on, after reentering the working hypothesis. The object is already there, analyzed for the other reading and does not need to be reanalyzed. So a Milano is taken as the filler for the impulse and the analysis is concluded properly.</Paragraph>
  </Section>
class="xml-element"></Paper>
Download Original XML