File Information

File: 05-lr/acl_arc_1_sum/cleansed_text/xml_by_section/metho/79/j79-1045_metho.xml

Size: 90,073 bytes

Last Modified: 2025-10-06 14:11:10

<?xml version="1.0" standalone="yes"?>
<Paper uid="J79-1045">
  <Title>SYNTAX</Title>
  <Section position="2" start_page="0" end_page="3" type="metho">
    <SectionTitle>
S REL
S NP PRO I
FEATS NU SG
AUX TNS PAST
VP V GIVE
NP **NP**
PP PREP TO
NP PRO YOU
FEATS
</SectionTitle>
    <Paragraph position="0"> The fourth element of every arc in a MATN is a small integer whlch is called the wei~ht of the arc. This weight was originally conceived of as a rough measure of either (a) how likely the arc is to be taken when the parser is in that state or (b) how much information is likely to be gained from taking this arc, i.e. whether the parse path will block quickly if the arc is wronn. That these two schemes are not equivalent can be seen by the following example. In a given state, say just after the maln verb of the sentence has bean found, the arc which accepts a particle may be much less likely than the arc whdch Jumps to another state to look for complements. However if a particle which agrees with the verb is found in the input stream at this point, then the particle arc is more likely to be correct. Since it is not at all clear how to measure or even inbuit how much inforrnatiop is likely to be gained from taking an arc, it was decided that the weights would reflect relative likelihoods. The actual weights which have been used in the speech grammar reflect an Tntuitive, though experienced guess as to how likely the arc is to be correct if it is taken, assuming the state itself is on the corr2ct path.</Paragraph>
    <Paragraph position="1"> Two grammars which will figure predominantly in the remainder of this paper have been written in the MATN formalism.</Paragraph>
    <Paragraph position="2"> One is an extensive grammar which can handle ,many questions, declaratives, noun phrase utterances, imperatives, active and passive forms, relative clauses (reduced and unreducad) , complements, simple quantifiers, noun-noun modifiers, varb-particle constructions, numbers, and dates (but not conjunctions). It began as a modification of the grammar for the Page 14 LUNAR system [361 but has been considerably adapted and expanded. This grammar is. called SPEECHGRAMMAR, and is listed ina[4].</Paragraph>
    <Paragraph position="3"> Exampled are given below which ware produced using this pramnar.</Paragraph>
    <Paragraph position="4"> For some illustrative purposes, SPEECHGRAMMAR ip too hir nnJ complex, so we have produced a UINIGHAEIMAR which 1 be ~IF~V to show the basic operation of the speech parser. A detailed  listing is given in Appendix I, but the diagram In Flpu~e ?...</Paragraph>
    <Paragraph position="5"> probably shows the structure mor-c clearlv. The serious rrodrr is encouraged to sketch n ccpy of this grammar for r-rfrrencc later on.</Paragraph>
  </Section>
  <Section position="3" start_page="3" end_page="3" type="metho">
    <SectionTitle>
CAT AD3 CAT N PUSH PP/
PUSH NP/ POP
PP/PREP PP/ NP
</SectionTitle>
    <Paragraph position="0"> Since the work reported here was finished, the author has written another grammar, called SMALLGRAM which uses the 1IATN formalism but which embodies a great deal of semantic and pragmatic, information specific to the domain of discourse currently baing used by the BBN speech understand in^ project.</Paragraph>
    <Paragraph position="1"> Page 15 In or for the parser der to move from right to left (to predict what could precede that first given word), it must be able to determine for any state which arcs can enter it, and for any arc which state it comes from. Since the ramm mar is organized for normal parsing in just the oppoeite fashion, i.e. for any state one can determine what arcs leave it and for any arc (except POP) one can determine which state it terminates on, it was necessary to build an index into the granlmar. This index consists of a number of tables centaining pre-computed informationwhich in effect inverts the grammar.</Paragraph>
    <Paragraph position="2"> Section 4 Uverview of SPARSER The input to SPARSER is assumed to be a set of words together with their boundary points (which may or may not be related to points in time). A word together with its boundaries Is termed a word match. A word match also includes a score which indicates how well the ideal phonemic representation of the word matched the acoustic analysib of the utterance (but as we shall see the parser has little need of this information). Since the same word may match at several sets of boundary points or may match in deveral ways between the same boundary points, each word ~t~~ is also given a unique number to help identify it. Thus the structure for a basic word match is: (number word leftboundary rightboundary lexicalscore) e.8. (4 TRAVEL 5 11 94), or (4 TRAVEL 5 11 (94 110)) where the score is given as a pair of numbers representing the actual and maximum scores, or (4 TRAVEL 5 11) where the score is omitted. How is the input to the parser to be constructed? We assume that acoustic processing and lexical scanning components can operate on a digitized waveform to produce a number of word matches such as prev.iously shown in the word lattice of Figure 1.1. (That this is possible has bean demonstrated by Woods [33]). Allowing the parser to operate unrestricted on the entire word lattice would probably not be fruitful because of the large numbe~ of locally syntactically correct combinations of words, but one possibility for input to the parser would be to take a set of the best-matching, non-overlapping word matches in the lattice, such as those in Figure 4.1.</Paragraph>
    <Paragraph position="3"> A set of non-overlapping word matches is a hypothesis about the content of the utterance. In order to avoid creating large numbers of such sets which are put together combinatorially with no basis except local acoustic match, semantic or pragmatic processes can be used to group word matches based on what is meaningful or likely to be heard. For example, if a dialogue has been about various nickel compounds, the combination &amp;quot;nickel analysesw may be more likely than &amp;quot;chemical analysesff even though the word match for 'tchemicalff has a higher score than that for mnickelfff'. We will not attempt to detail here how this semantic grouping could.be done and how the sets could be scored, since it has been described elsewhere [15].</Paragraph>
  </Section>
  <Section position="4" start_page="3" end_page="22" type="metho">
    <SectionTitle>
DO MANY PEOPLE DONE CHEMICAL ANALYSES ROCK
</SectionTitle>
    <Paragraph position="0"> Using more terminology from the BBN speech system, the word theorv to denotes a set of word matches such as we have just described together with (possibly empty) slots for information from each of the possible knowledge sources in the system. From the point of view.of SPARSER, usually only the word match portion of a theory is of fnterest, hence we shall fall into the habit of using the word &amp;quot;theoryv to refer to the word match set it contains. When speaking of the syntactic component of a theory, however, we are refering to the information slot for syntax whicn accompanies each word match set.</Paragraph>
    <Paragraph position="1"> Theories have the fallowing characteristics: 1) They contain a set of basic, nondverlapping word matches.</Paragraph>
    <Paragraph position="2"> 2) They tend at first to contain long content words and not many shdrt function words. This is because long words are more reliably acoustically verified and content words are easier to Page 18 relate semantically and pragmatically. Since small words such as tam r~d~v , rrtherl, rr~nall , &amp;quot;have&amp;quot;, rr Of 11 11 in rr I T , etc. may be reprssented by very little acaQstic information, they would tend to match at many places in the utterance where they do not really occur. Consequently they ase not searched for? by the initial word match scan, nor are they proposed in the semantic stages of hypothesis formation.</Paragraph>
    <Paragraph position="3"> 3) They need not (and generally do not) completely span the utterance, but have numerous gaps of va~ving sizes (a.p. for the function words).</Paragraph>
    <Paragraph position="4"> 4) They tend to contain some sequences of contiguous word matches. Such a sequence is called an island.</Paragraph>
    <Paragraph position="5"> That such a set of theories can be created has been demonstrated by the BBN SPEECHLIS system. ?he syntactic component, SPARSER, is expected to process these theories one at a time. In certain circumstances which will be detailed later, the input to SPARSER will be a theory together with one or more word matches which are to be added in order to create a new larger theory which is then to be syntactically analyzed.</Paragraph>
    <Paragraph position="6"> We will assume that there exists a cantrol component which presents SPARSER with theories to process and to which SPARSER can communicate predictions and results.</Paragraph>
    <Section position="1" start_page="3" end_page="22" type="sub_section">
      <SectionTitle>
Preliminaries
</SectionTitle>
      <Paragraph position="0"> Given a theory, what is to be done with it? We begin by considering a subset of the question: Given an island of word matches, what is to be done with it? The answer is to create one Page 19 or more parse patbn through tho island and to predict what words or syntactic classes could surround the island. A parse path is tho Sequehce of arcs in the grammar which would be usad by a conventional ATN parser to process the words in the island, if the island were embedded in a complete sentence.</Paragraph>
      <Paragraph position="1"> For example, consider the way a parser might process an island of word match'es such as (1 CHEMICAL 14 22) (2 ANALYSES 22 30) using the MINIGRAMMAR of the previous section. Beginning in state NP/ of the grammar (omitting for the moment the problem of how it is known that NP/ is the rieht place to begin) the sequence of s arcs which would bh taken to parse &amp;quot;chemical analysesw as a noun phrase is that shown below in  Portion of MINIGRAMMAR needed to parse %hemica1 analysesm Let us define a confiatmation to be a representation of the parser being in a given state (say NP/QUANT) at a given point in the utterance (say 14). We will write configurations as STATE:POSITION in text (e.g. NP/QUANT:14) and schematically as a box within yhich are written the state and the position. If a configuration represents a state which is either the initial state of the grammar or a stake which can be PUSHed to (i.e a  state which can begin the parsing of a constituent),  filled-id semi-circle attached to the left edge of the box. Note that a confi~uration N~/QUANT:I~ is quite distinct from a configuration NP/QUANT:22 since they are at different positions in the input. In SPARSER, each configuration .is also assigned a unique number which is a convenient inte~nal pointer.</Paragraph>
      <Paragraph position="2"> The process of traversin~ an arc of the grammar using a particular word is represented by a transition from one config.uration to another. A transition ean be made only if the arc type is compatible with the current item of input and if the context-free test on the arc is satisfied. (The context-sensitive tests are evaluated later.) A transition carries with it information about the arc which it represents and the item of input it uses. The item of input is usually the word match which the arc uses, but it is NIL in cases such as JUMF arcs which do not use input, and it is a complete constituent fop PUSH ams. A unique identifying number and the list of features, if any, which is associated with the input word or constituent are el30 recorded on the transition in SPARSER, but they are not shown schematically. A transition is represented schematically by an arrow from one configuration to a~~other with an abbreviated form of the arc written above the arrow and the item of input under it.</Paragraph>
      <Paragraph position="3"> The syntactic part of any theory which SPARSER processes contains, among other things, lists of the transitions and configurations which are created or used by the theory. Thus wheh we talk about creating a configuration or transition it is implicitly understood that SPARSER also adds it to the appropriate list, and when we talk of adding an ax is tin^ configuration or transition to a theory we mean adding it to the npprbpriate list. Therefore, removing a confi~uration or transition from a theory means removing it from the syntactic part of the thebry, not removing it entirely from SPARSER s data base.</Paragraph>
      <Paragraph position="4"> Like confi~urations, transitions are unique, so only one transition is ever constructed from point A to point B for arc X and input Y. We will frequently speak of creating a transition or a configuration, but the reader must bear in mind that if such a confi~uration or transition already exists, this fact will be recognized and the pre-existing configuration or transition will be used. (Timing, measurements indicate that it takes about ,052 seconds to create a configuration and only .01 seconds to test if a particular configuration already exists. For transitions, creation takes about .54 seconds and recognition .012 seconds. The sequence of configurations and transitions which would parse the above example is displayed in Figure 4.3.</Paragraph>
      <Paragraph position="5"> A conpected sequence of transitions and configurations is called a at. If the sequence begins with an initial configuration and ends with a transition representing a POP arc,  it is a complete path, otherwise it is a partiaL path. Paths are assumed to be partial unless otherwise specified.</Paragraph>
      <Paragraph position="6"> darrinnina Paraa a Island SPARSER processes an island of words by beginning with the leftmost word and determining its possible parts of speech. Then the arcs of the grammar which can process the word arc fpund (by looking in the previoRdly constructed Rrammar index). For each arc, two confi~urations are constructed one for the state at the tail of the arc and one for the state at the head, using th$ left and right boundary positions of the word match, respectively, and a transition for that arc using the current word match is also built. Schematically, we have for our example a situation which looks like that of Figure 4.4 (such a display of all or some of the transitions and co~fi.qul*at ions which the parser has constructed is called a map). Notice that a configuration may have any number of transitions entering or leaving it.</Paragraph>
      <Paragraph position="7"> Figure 4.4 Initial map for parsing llchemical analysestr Page 23 The idea of this process is to begin t6 set up paths which my be used to parse the island. However it is not necessarily the case that the only donfigurations which could start paths throu~h the island are those which have just been obtained, since it may be possible to oreatr transitions which enter them via JUMP arcs or TST ahca. For each state, the sequence of arcs which can reach it without using the previous word of input have bean be pre-calculated by the grammar indexing package so the appropriate configurations and transitions may be constructed. These transitions ape cplled lead-in transitions. Thus the map becomes that in Figure 4.5 Note that any of the configurations (except for NP/ADJ:22  J J* and NP/N:22) could actually be the correct leftmost configuration for this island, depending upon what the (currently unknown) left  By looking in the grammar index, SPARSER can determine, for t~c each configuration which could start the island, just what sort of left context could be appropriate. For example, th'e CAT ADJ arc in MINIGRAMMAR which enters state NP/QUANT implies that an adjective could precede the island and, if it did, the tra~sition which would proceas it would terminate on configuration NP/ADJ:14, Because the initial configuration NP/:14 could start the isiand, anything which could precede n noun phrase could occur to the left; again the grammar index provides the information that the CAT PREP arc could lead to a configuration which could accept a noun phrase (via the PUSH NP/ arc), so a preposition could also prefix the island. If the index functions indicate that a constituent could be picked up by a PUSH arc which could terminate on the configuration under consideration, an indication is made in tho WFST so that any time a constituent of the desired type is built which ends at the proper location, it may be tried here.</Paragraph>
      <Paragraph position="8"> Because of the highly recursive nature of ATN grammars, it is vary likely that as we chain back through the possible sequences of PUSHas which could lead to tho beginning of tho current constituent (or the seauence of POPS which could be initiated by the completion of the current constituent) a large number of predictions will be made. Rather than make all these predictions automatically, beford we are even sure that there is in fact a constituent at the current level, the possible configurations which could make predictions on other levels are saved to be activated later if the predictions from the current set of active configurations are not sufficient.</Paragraph>
      <Paragraph position="9"> Page 25 The predictions which are made (not saved) are not acted upon at this time, but ard kept internally by SPARSER until all the islands of the theory have been prooessed. We shall see bdew what then becomes of the predictions.</Paragraph>
      <Paragraph position="10"> Island Once proceasing has proceeded this far, we can go back and consider the set of configurations which represent states the parser could be in just after processing the first word of the island. In our example, these are configwatiqns WP/ADJ:22 and NP/N:22. Configurations such as those whlch are waiting to be extended to the right are called active configurations. SPARSER selects a subset of the set of active co'nfi~urations (how this subset is selected will be discussed in the next section) and for each configuration tries to extend it by tryin8 to parse the rest of the island beginning in that confipuration. When the parser is considerin~ a configuration at some position, the input pointer is set to the word match of the island, if any, which begins ah the same position in the input.</Paragraph>
      <Paragraph position="11"> The grammar associates with the state of the configuration a list of arcs which may be tested (using the arc type, the context free test on the arc, and the Current input) to determine whether a transition can be made to extend the path. We will consider each type of arc in turn, since the effects of taking various types of arca are different, and explain for each case what happens if the arc is taken. Whether just one transition, or several, or all possible transitions are made from an active Page 26 conf,iguration is a matter to be discussed in Section Five. Soma JUMP arcs do not look at the current item, so they may be taken whether the input pointer is set to a word match or to NIL. The transition which results from taking an arc of this type has a null item associated with it, even if there is a word match in the theory at this point. The positions of the confipurations at each end of the transition are the same; this corresponds to the fact that an ATN parser would not move the input pointer as a ,consequence of taking this arc.</Paragraph>
      <Paragraph position="12"> Rarely, a JUMP arc may test the current iten in some way, for example, to make a feature check. If there is no word match for input, an arc of this type cannot be taken. If there is a word match, it is noted on the trahsition wh ch is created, but the configurations at each end of the transition have the same posit ion. (It is than the case that thainext input-using or input-consuming transition on the path including this transition must use the same word match.) These are TST, CAT, and WRD arcs which end in a (TO nextstate) action. The operation is exactly tho same as that above except that the configuration on which the transition terminates has the position of the right boundary of the current word match.</Paragraph>
      <Paragraph position="13"> Taking a POP arc results in the creation of a transition which has a null final configuration and a null item, because POP arcs are not permited to look at input.</Paragraph>
      <Paragraph position="14"> Page 27 When a PUSH arc is encountered, a monitor is placed in the Well-Formed Substring Table (WFST) at the current Dosition to await the occurrence of a constituent of the required type. If one or nore such constituents are already in the table, then for each one there are three possibilities: it may be composed of word matches which are in the current theory, it may be composed of word matches some of which are not in the current theory but which could be added without violating the non-overlapping constraint, or it may be composed of word matches some of which are incompatibld with the current theory.</Paragraph>
      <Paragraph position="15"> In the first caae a transition is set up using the constibuent as the current word. The transition terminates on a confleuration whose state is determined from the termination of --.</Paragraph>
      <Paragraph position="16"> the PUSH arc and whose position is that of the right boundary of the rightmost word match in the constituent.</Paragraph>
      <Paragraph position="17"> In the second case, a notice is created and sent to the control component. A notice is a request that SPARSER be called to enlarge a theory by addinp some new information, in this case, some additional word matches which form a constituent that the theory can use. SPARSER does not try to determine when (or even whether) the theory should be so enlarged. That is an issue for the main cqntroller to decide (see Rovner, et.al. C231). We will discuss below how SPARSER enlarges a theory if called upon to do so.</Paragraph>
      <Paragraph position="18"> In the final case, if there are no usable cohstituents in the WFST, a new configuration is set up to start looking for one and is added to the list of active configurations. Its state is Page 28 the state specified by the PUSH arc and its position is the same as the current configuration.</Paragraph>
      <Paragraph position="19"> There is a considerable amount of processinc that can happen any time one of the transitions lust discussed is rnndr. Whenever% an initial configuration is constructed, this fact is r*t)c@rded in the configuration. Whenever a transition is nade from such a confipuration, the information that there is a path I sene initial conf igur-at ion is recorded on the subsoquenf configuration. Similarly, whenever a POP tr-ansition is made, the c~n~figuration it emanates from and all previotls configuratims on any path which can terminate with the POF transition are marked to indicate that they can reach a POP transition. Whenever a transition is made which completes a path from an initial configuration to a POP transition, the path is executed, one transition at a time, and the rqister setting actions and context sensitive tests are executed. If a test fails or an arc aborts, the transitions and configurations f the path are removed from the list of configurati~ns and transitions which are in the syntactic part of the currant thaory (unless they are used by another path in the theory) but not removed from the nap. If the execution is successful, a deep structure tree is produced. That structure together with its features is given a score, whioh may include evaluations by other cornpodants such as semantics and prosodies, and is entered in the WFST.</Paragraph>
      <Paragraph position="20"> It is quite important that sources of knowledge other than syntax be called upon to verify and to rank syntactic constituents. This is because there are likely to be many Page 29 ombinations of plausible words from the word lattice which form ayntact%rally reasonable constituents but which may be ruled out om @%her grounds. To allow immediate use of this information which syntax cannot provide alone, SPARSER has an interface to he semantic component so that constituents can be vrrif ied dinctlg without going through the control component. It will be %trivial modification to insert verification aalls to pragmatics aad prosodies when they become available. In the meantime, even semantic knowledge can be turned off; if the parser gets no IngosnratIon from the call to semantics, it proceeds without it. Plaoement of a constituent in the WFST causes a number of Lbiags to happen. First, any monitors which have been set by the mmnt theory at that position aFe activated. That is, for each ebnfigaration which was waiting for this constituent, a PUSH tmaaitioa 3s made which uses the constituent as its input item. If ao rmonltors have been set which can use this constituent, it as treated exactly as if it were the first word of an island: a11 the PUSH arcs which can use it are found in the grammar index md appropriate configurations and transitions (including lead-in tnositions, if appropriate) are set up. Next, if there are any monitors for other theories which can use the constituent, patfees are created and output to Control as was described above is the section on PUSH transitiohs.</Paragraph>
      <Paragraph position="21"> Figure 4.6 shows SPARSER s map after our example island has ken completely processed. The parsing results-in the creation oi a CAT II transitio~ to configuration NP/N:30 using the word *~~alyses~ The PUSH PP/ arc at state NP/N would oause configuration PP/:30 to be created. Similarly, PP/:22 would be created when tho configuration NP/N:22 is picked up to be extended. The POP arc transitions from each of the configurations for state NP/N result in the formation of complete paths, resulting in the creation of two noun phrases (&amp;quot;chemical analysesw and ttchamicalft). Since there were no monitors for them, they result in the creation of configuration PP/PREP:14 and  It may be the case that no path can be found from one end of an island to the other, (This would occur when all active configurations block.) In this case, there is no possible way that the island could form part of a grammatical string, so SPARSER can inform the control component that the theory is wrong, When an active configuration is picked up to be extended and there is no word match at that point, the end of the island has been reached. That does not mean that no more transitions can be made, since arcs which do not test the input word can be taken as usual. Arcs which do use input cannot be taken, but they can be used to predict what sort of input would be acceptable at that position. For example, a GAT V arc which has a test requiring the verb to be untansed would allow SPARSER to predict an untensed verb beginning at the position of the ourrent configuration. CAT and WRD arcs cause the prediction of syntactic categories and specific words, respectively, modified by the Context-free test on the arc. TST arcs provide only the test which must be satisfied, and PUSH arcs cause a monitor to be set in the WFST as well as a TST monitor for the the look-ahead test (if any) on the arc.</Paragraph>
      <Paragraph position="22"> Bndinq 3 Theorv When k11 the islands of a theory have been processed in the manner just described, it is time to deal with the gaps between the islands. As we have seen, arcs in the grammar which can Page 32 enter configurations at the left end of an island or which can leave configurations at the right and of an island can be used to make predictions about warda that may be adjacent to the island. The prediction is a list of the arc, the confi~uration it would connect to, and an indication of whether the transition caused by the arc will enter the configuration from the left or leave it to the right.</Paragraph>
      <Paragraph position="23"> If a gap between two island8 is small enou~h that it may contain just one word, than it is likely that tho arc which would process that word may have caused a prediction from both tho left and right sides of the gap. If this is the case, and if the predictions intersect in a single possibility, it is highly probable that the word (or syntactic class) so predicted is correct. If the predictions do not intersdct, parsing is continued from the active cbnfigurations which were not triad earlier because of their scores and from the configurations which could begin constituents at the right and of an island. This continued parsing is an attempt to find a path which results in a common prediction acg#ss the gap. If that too fails, then the configurations which were saved because they could lead up a chain of PUSHes or POPS to new configurations are triad. If no possibilities are left to try and there is still no prediction to fill the gap, this information is noted, but it does not definitely mean that the islands are incompatible, since in some cases the gap could actually be filled by two words instead of one.</Paragraph>
      <Paragraph position="24"> SPARSER has two kind of predictions - those bhioh seem highly likely and those which seemaless likely. A highly likely prediction, such as one which is made from both side3 of a small gap, is output in the form of a prooosal, which is a request to  the rest of the system to find a word meeting the requirements of the proposal. A proposal contains: 1) the item being proposed, which is either a particular word or list of words (from a WAD arc), or a syntactic olass (from a CAT arc), @r NIL, meaning any word (from a TSI a~c) 2) tho loft and/or right boundary pointb) of the item 3) a test which the item must satisfy (the context free test from bhe arc) 4) the context of the proposal, i.e. the word match(es) on  the left apd/or right side of the item baing proposed. (This is to help the lexical retrieval component take into acceunt phonological phenomena which may occur across word boundaries.) All predictions whether or not they are confident enough to become proppsals are output oas monitoys. A monitor is a notification to the control component that if a word meeting the requirements of the monitor is somehow found (perhaps by the action of a proposal) , it may be added to the theory. Thus a monitor acts like a demon which sits at a particular point in the word lattice and watches for the appearance of a word match which it can use. A monitor contains:  1) the item being monitored for (generally a syntactic categcry, but may be a word or a test) 2) the left or right boundary position of the item baing monitored far 3) a. test which the item must satisfy (same as for proposals) 4) the thaory which generated the sonitor 5) the arc in the grammar which will process the item if found 6) the configuration from which the prediction was made 7) a score, indicating roughly how important the monitor is,  i.e. how much information is likely to be gained by processing an event for that monitor.</Paragraph>
      <Paragraph position="25"> (Notice that monitors which are sent to the control component are very much like monitors which are set in the WFST by the occurrence of PUSH arcs.) Once the proposals have been made and the monitors have been set., SPARSER bundles up the information it knows about the current theory, such as the configurations and transitions in the theory, any configurations which a still candidates for expansion, the constituents in the theory, the notices, proposals, and monitors which have been created, etc. and associates the bundle with the thaory number. This insures that SPARSER will be able to pickup where it left off if it is later given the thmry to process further.</Paragraph>
      <Paragraph position="26">  Thus far we have seen only the opSrations which SPARSER performs on a single theory, but we made the assumption that SPARSER would be given a number of theories to process in sequence. Let us now examine what will happen when the second (or nth) tkeo~y is processed.</Paragraph>
      <Paragraph position="27"> Page 35 SPARSER will no longer have a blank map and WFST; instead it will have all the configurations, transitions, and constituents which have been constructed by all previous theories. For concreteness, let us imagine that the theory (1 CHEMICAL 14 22) (2 ANALYSES 22 30) has been processed, result in^ in the map shown in Figure 4.6. Now we are going to process a theory containin8 the island (4 NICKEL 16 22) (2 ANALYSES 22 30), which results in the map of Figure 4.7 where the configurations and transitions added by this theory are shown in dotted lines. The process bagina as usual with the creation of conf,iguration #P/ADJ:16 and three possible lead-in tranbitions. Tho transitions for the two CAT N arcs, however terminate on configurations which already existed in the map, so the complete paths from configuration NP/:16 to configurations NP/N:30 and NP/N:22 will be discovered and processed, resulting in the construction of two new noun phrases. Those new constituents would then result in the creation of configuration PP/PREP:16 and two new transitions. Thus we have constructed only five new configurations and seven new transitions and have been able to take advantage of six old configurations and six old transitions. In this fash-ion any information which has once been discovered about a possible parse path is made available to any other path which can use it. - No reparsinq - is ever done SPARSER merely realizes the existewe of relevant configurations and transitions and incorporates them into the current theory.  Map after processing islaad for &amp;quot;nickel analysesif If the new word (or wokds) in a theory are at the and (or in the middle) of an i$land, when SPARSER begins to parse the island it will discover the existing configurations and transitions from the previous theory. Whenever a transition which can be used in the current theory is discovered in the map, it and its 'terminating configuration are added to the syntactic part of the current theory. This is callad tracinq the transition. In addition, all paths beginning with that transition which do not require the next word of input are also included in the syntactic part of the theory. This is accomplished by tracing from the terminating configuration all transitions which use either tho same word of input as the previous transition or no input word at all. (A similar process 19 used to trace backwards, i.e. right to left, when neessary.) Uhen a configuration is reached which has no traceable transitions emanating from it, the tracing process, stops. Since both transitions and configurations are stored in such a way as to facilitate tracin~ (for example, each transition has a code attached to indicate whether or not it consumes or tests input), this process is considerably faster than creating that portion of the map in the first place. (To illustrate this, a theory was processed twice, onca with an empty map and onca start in^ with the map previously created; the time required for processing the theory fall from 47.5 seconds td 16.5.) Configurations which can end trvaced paths are put on the active conCigurations list. If, when one oP them is picked up for extension, it is discovered that the next word of input was used on a transition already in the map, the tracin~ process is repeated. If the next word of in~ut is new (or at least has not caused any transitions from thi? con fipurat ion beinp considered ) then para in^ continues in' the normal manner.</Paragraph>
      <Paragraph position="28">  As-was mentioned earlier, SPARSER can be called upon to add some new word matches to a theory it has previously processed. In this case, SPARSER is said to process an event. An avant may be thought of rather abstractly as the discovery of a piece of information that has been syntactically proposed, monitored for,  Page 38 or noticed. Concretely, an event is a piece of data consistina of: 1) the old theory that proposed or set a monitor far the event 2) something to be added to the theory (a new ward natch or constituent) 4) the arc in the Rramnar which will process the new information 4) the corifipuration ih the old tbrory which will be at one end of the transition created by the above arc When SPARSER is ~iven an event, it retrieves from its tables  the bundle of configurations, transit ion, etc. in the old theory. Then using the arc and the new word or constituent in the event, it creates the appropriate transitioncs). Then processing continues as usual, that is, any complete paths are noticed and processed, and any new active confi~urations are exbended, if possible.</Paragraph>
      <Paragraph position="29"> New predictions may be made as a result of this increased information. (A record is kept of previous predictions so none are remade unless with a more liberal score.) Finally SPARSER returns the nGw, larger theory. This new thsory may be processed as part of another event at some later time, thus gradually reducing the number and size of the caps in the theory.</Paragraph>
      <Paragraph position="30"> If an event results in filling the final gap in a theory, and if the resultinp complete sequence of words can be parsed, SPARSER notifies the control component of this fact, since the entire utterance may have been discovered. Of course, this may not be the correct solution -- it is up to the control component to look atm the acoustic ~oodness, semantic meaningfulness, pragmatic likelihood, etc. of the result as well as the syntactic structure before daclarlng the utterance to have been understood. If for reasons other tham syntactic, the utterance appears to be bad, the control component of the system could ~o on to try to find anothar, more suitable, possibility.</Paragraph>
      <Paragraph position="31"> Section 5 More Details of the Parsinfi Process</Paragraph>
    </Section>
    <Section position="2" start_page="22" end_page="22" type="sub_section">
      <SectionTitle>
5.1 DEPTH vs BREADTH
</SectionTitle>
      <Paragraph position="0"> The parsing strategy just outlined works bottom up when beg inn in^ to parse an island and when a constituent is created which was not monitored for by the current theory. It works top dawn after an island has been started and to make syntactic predictions at the ends of islands. Both top down and bottom up techniques can be either depth or breadth first. Depth first processing takes at every step the ffl-st piece of information available and pursues its consequences. Breadth first processing considers at every step every possible next step of every alternative and pursues all paths in parallel. Breadth first processing generally takes much more space than the depth first many paths would have to be remembered at once Rags 40 instead of having just one stack which could be popped and reused when necessary.</Paragraph>
      <Paragraph position="1"> The breadth first process mi~ht save some computation steps and might produce several ambiguous parsin~s simultaneously while tba depth first process would find one before the others (the latter is a'small difference, since both processes would have to be run to exhaustion to insure that all possible parsings had been found). In parsing speech, some mixture of breadth first and depth first proc~ssing can be extremely useful.</Paragraph>
      <Paragraph position="2"> To illustrate an advanta~e of breadth first processing in the speech environment, consider what might happen if, durinp the processin8 of an island the parser picks up a confirtiration to extend which has several possible arcs emanating from it. If one arc is chosen and all the others are held as alternativss (imem depth first), but the chosen arc is wronp, all subsequent paths beg inn in^ with that arc would have to block before the alternatives would be tried. However, if the end of the island were reached before. the success or failure of the first choice were confirmsd, the only way that backup would ever take place would be to have one or more events add words to the thsory so that the path could be extended until it failed. Since the pap wou.ld be likely to be filled by (incorrect) words predicted by the erroneous path, or by no words at all if the (incorrect) predictions were not satisfied, it is not at all clear how the process would aver know to back up.</Paragraph>
      <Paragraph position="3"> This problem cannot be eliminated completely without pursuing all alternatives to thair Fulleat extent (a combinatorially unacceptable solution) but it can be modified to a praat extant by a judioious combination of depth and breadth first processing to find the best path, not just the first one, through the island. This &amp;quot;bast pathw is not ~uaranteed to be the correct one, so it is possible to continue process in^ by extending paths with were suspended earlier.</Paragraph>
      <Paragraph position="4"> SPARSER handles the problem by assigning a score &amp;o every configuration which reflects the likelihood of the path which terminates on that confi~uration to be correct. The score can also be thought of as a measure of how good that eonfiguration looks in relation to others as a candidate for extension. One question which was previously left unanswered, how a subset of the active confisurations is chosen for extension, can now be answered : the subset of maximally scoring configurations is chosen at each step until the maximal score of act ivt! configurations begins to fall. (The score on a configuration and the score of a path terminatinr on that configuration are the same thing -- we will use which ever terminology seems most natural at the time.) The result of this process is a sort of modified breadth first approach, where at one step all the alternatives are tried but at the next step only th&amp; best ones are chosen for further extension. This is similar to the best-first parser describad by Paxton in [I81 but it can be applied to the sort of partial paths which SPARSER generates rather then requiring the perfect Page 42 information resulting from a strictly left to ripht approach.</Paragraph>
      <Paragraph position="5"> The auccesg of this method is directly dapandent on the ralativz accuracy of the scores which are assigned to the paths.</Paragraph>
    </Section>
    <Section position="3" start_page="22" end_page="22" type="sub_section">
      <SectionTitle>
5.2 SCORING PATHS
</SectionTitle>
      <Paragraph position="0"> Sevdral attempts have b~en made to d*velop rigorous systems for parsing arrorful or spedch-like input baaed on probabilities [I, 14, 271. These attempts have all simplified the problem to such an axtent that it is no lonper realistic or extendible, e.p.</Paragraph>
      <Paragraph position="1"> by assuming the input is a sequence (rather than a lattice) of probability distributions, by assuming that all the neclsbary information is present in the searhh space to begin with so the only problem is to find an optimal path throuph ths spacz, by requirine a small vocabula~y, and/or by limiting the Rranmar to be context free.</Paragraph>
      <Paragraph position="2"> The ideal scoring mechanism for SPARSER would be one which accurately reflected at every step the probability that the path 8 correct. Bayas rule could be Used, but it would ba necessary to know, at any point in ths parsing process, what the probability is that th~ next arc under consideration is correct, given that the entire path up to the current step is correct. In order to use this application of Bayas rule it would be necessary to pra-calculate the probabilitiss for evary possible path and partial path which could be generated -- a clearly impossible task sincs there are an inftnite number of such paths.</Paragraph>
      <Paragraph position="3"> Givzn that we cannot calculate the probabilities we need exactly, what is the next best option? If we ignore the effect of tila path traversed up to the current point, but can say for ray ~iven state how likely each arc em an at in^ from that state is to ba correct, we would have a model which uses only local Paforaation rather than one which takes into account accurately all1 tbe Left context which is available.</Paragraph>
      <Paragraph position="4"> Since it was not practical to run large amounts of data tbm~h a parser in order to obtaid accurate measurements even Sor the limited model, the author re'lied on considerable experfeace with ATN grammars to assign a weight to each arc of tba grqmmar representing tho intutive likelihood that the arc fii it can be taken) is the correct cane to choose from that state. These weights are small integers (0 throu~h 5) -- the jarger the weight the more likely the arc.</Paragraph>
      <Paragraph position="5"> The question might arise as to why the score oG the word sate4 used by an arc should not be used to influence the score of tbu path using it. SPARSER tries to treat each theory as ladrependently as possible and tq assign scores based only on the syntactk information which is available. The one exception to this rule is the semantic information which is used to score constituents. If lexical Qord match scores were used, the cbmtrol component would not be able to separate the lexical goudwss from the syntactic goodness of the theory and make Judgments aa to their relative importancz. In a syntax-driven apeecb understanding system, however, it would probably bz usaful to combine lexical scores with syntactic information.</Paragraph>
      <Paragraph position="6"> Page 44 As was described in the previous section, when SPARSER begins to parse an island each possible partial path is begun by creating a configuration at the head of a transition for an arc which can use th@ current word. Rather arbitrarily, it was decided to giva this confi~uratioq a score of one. This starts all partial paths out equally, a technique which is not quite accuratb, since some contexts are more likely than othms. For example, the words tttom and &amp;quot;forw are more likely to occur In prepositional phrases than in sentantial complements. If this simplification appaarv to harm the overall performance of SPARSER, it coula be remadikcf by giving eaah state an a priori score similar to the weights on arcs. Configurations on lead-id paths are also given a score of one.</Paragraph>
      <Paragraph position="7"> After the initial step, whenaver a transition (othar than a PUSH or POP) is made, the score of the subsequent configuration is influenced by the score of the conf'i~uration being extended and the weight on the arc beins us$d. If the scores were actual probabilities; they would be multiplied; since they are not, it was arbitrarily decided to add them.</Paragraph>
      <Paragraph position="8"> When attempting to create a configuration which already exists (a situation encountered whenever two or more parse paths for the same theory merge), ths configuration is given the maximum of the sxisting score and the score which would haw been assigned had the configuration been created anew.</Paragraph>
      <Paragraph position="9"> Whsn a PUSH arc is encountered and a configuration created to begin the search for the required constituznt, the score of that configuration is set to be the sum of the scora of the configuration causing the PUSH, and the value (If any) of the look-ahead test on the PUSH arc. For example, upon encountarinr an arc such as (PUSH NP/ ((NPSTART) T T) ... ) the look-ahead function NPSTART returns a high integer valua ir the next word is a noun and a lowen valua if it is a verb (e.g. ltaccounting coststt). Of course, if tha look-ahead funcrion fails altogether, the c~nfipuration is not set up, althou~h hhe monitor in the WFST remains.</Paragraph>
      <Paragraph position="10"> When a constituent is completed (or found in the WFST) end a PUSH tranaitiqn is about to be made, the score of the confi~uration on Ghich the transition terminates is a function ~f the score of the confi~uration heinc extended the weipht on the arc, an@ the score of the constituent itself. The score of tht* constituent is currently very ad hoc, bdng a function of the number of words in th* constituent (lass a function of the number of sub-constituents subsumed by this constituent, boosted if the constituent is a mador one) and the score which is determined by semantic verification. Thus semantically ''roodIf constituents will st the scores of the paths which use them more than semantically &amp;quot;badw ones.</Paragraph>
      <Paragraph position="11"> Due to the level of effort required to gather accurate statistics on the relative frequencies of arcs, the current scores are admittedly ad hoc. It is not clear whether different scoring mechanisms would be better, however it is clear that the current scoring strategy is better than no scoring at all, as praiiminary measurements indcate that the number of transitions created (as well as the number of confi~urations and predicions) is reduced about 25% by thz current strategy.</Paragraph>
      <Paragraph position="12"> (It is rzasonable to ask why semantic scores are used to influance parse paths, sincd it was tust argued that lexical Ycores should not be ushd in this way Semantic scores may be more reliable than lexical ones because we are assuming that the utteran&amp; is semantically maanibaful. Under this ausumption, a constitudnt like &amp;quot;range remainder&amp;quot; as a noun-noun modifier analogous to ltaurplus moneyff should be ruled out as early as possible. Since such con8tituents cannot be ruled out on syntactic rounds alone, since prosodic information (which might help to rule them out) is not available (see discussion in Section 7.2), and since they would seriouslv overrun the parser with a plethora of false paths if they wwe not reJected, it seems reasonable to permit semantics to influence the parser.)</Paragraph>
    </Section>
    <Section position="4" start_page="22" end_page="22" type="sub_section">
      <SectionTitle>
5.3 SCORING PREDICTIONS
</SectionTitle>
      <Paragraph position="0"> The previous section discussed three ways in which SPARSER can make predictions about what could fill in gaps between islands. Monitors wait for the occurrence of a word in the word lattice (or a constituent in the WFST), proposals request a search for a particular set of words, and notices indicate the presence of a usable word in the word lattice (or a constituent in the WFST). Since the processing of a typical theory is likely to result in a number of predictions it is necessary t'o be able to order them so that predictions most likely to bz correct or most likely to yield important information will be acted upon first. For example, it is more important to fill a Rap between two islands than to extend a sin~le islahd, since by filling the Rap one can chzck the consistency of information which was locally good in aach island individually but may not be consistent when they are joined. Since two words can occur to~ethar in (usually) many contexts but lon~er szquances arl generally more restrictiv~, addine a word to a one word island is likely to be leas profitable in terms of the number a,f possible paths which are sliminatod by the addition than add in^ a word to a multi-word island.</Paragraph>
      <Paragraph position="1"> It is up to the syntactic component to indicate to the control component the relative importance attachsd to each notica and monitor; the hipher the score, the stron~er the prediction.</Paragraph>
      <Paragraph position="2"> Several factors influence the score attached to predictions.</Paragraph>
      <Paragraph position="3"> One is the length of the island to which the prediction is attached. One word islands, if they are processed at all, yield very little information anc many pradictions, bznca the predictions are not scored high. Proposals are lass important if there is already a noticsable word in the word lattice (since that word is acoustically better than the word to be proposed, else it would have bean found earlier. Howevsr, if a proposal fills a gap between two islands, it is given a higher score.</Paragraph>
      <Paragraph position="4"> Notices are boosted in importance if an entire constituent may be added and penalized if they will add onto a one word island.</Paragraph>
      <Paragraph position="5">  interaction with tha other components (in order to make the scores of syntactic prediction8 commensurate with those of semantic predictions) and may be chanazd considerably au the atire sy~tem evolves.</Paragraph>
      <Paragraph position="6"> Small syntactic classes em. detePml'naru and prepositions) are proposed in their entirety (that is, their elurnants are to be dnumrratad and ~ivdn to the lexical matchinc componlnt for verification) if the island which monitored for them is more than one word lon~. If a gap batwsen two island3 is small enouch for iust one word and if a syntactic class has been monitored for  SPAR3ER is written in INTERLISP and runs on a PDP-10 und~r the TENEX operating system . The program and initial data structures occupy approximately 90000 words of virtual memory.</Paragraph>
      <Paragraph position="7"> (The other componants of the BBN speech undzrstandicp system occupy separate forks from the syntactic component.) Page 49 At the time the 'examples in this section were run, the al~orithm controlling tha dacision-making process in the control component was under~oing reviaion and was not solidified into a function which could operate automatically. Rathar, there ware a number of primitlvd operations such as scanning an utterance (or some specified partion of it), creating thoorieu, call in^ SPARSER with a theory or event, calling f&amp;quot;or the processing of proposals, etc., which could be invoked by a human simulator The follow in^ examples were produced in this mode, with the user act in^ as the control component in a way which could be modelled by later imple,mentation.</Paragraph>
      <Paragraph position="8"> Several convention8 have been used in tracin~ the operation of SPARSER. Conf ieurations are rspresented as NUMBER : STATE : POSITION (SCORE). For example, the configuration written as 30:NP/HEAD:23(39) is the confi~uration for state NP/HEAD at position 23 which has been given the (unique) numbe~ 30 and which currently has a score of 39. The creation of a transition is indicated by naming the type of arc causing the transition, the (unique) number of the transition, and the configurations at each and of the transition. For example, CAT N TRANS #9 FROM 14:~PhET:6(1) TO 15:NP/DET:19(4).</Paragraph>
      <Paragraph position="9"> Annotations havc been inserted within brackets { 1; typeout in upper case was produced by the program.</Paragraph>
      <Paragraph position="10">  (In this version or the system, ra~ular inflectional endings are included in word matches after the element representing the score, hence the somewhat peculiar word match for the word &amp;quot;trips&amp;quot;.) Two theories were constructed, one for word matches 2 and 3, the other for 1 and 3. What follows is an annotated (but otharwisc! unedit,ed except for considerations of spacing) transcript of SPARSER processing these two theories in sequence, using the MINIGRAMMAR of Figura 3.3 and Appendix I.</Paragraph>
    </Section>
  </Section>
  <Section position="5" start_page="22" end_page="22" type="metho">
    <SectionTitle>
SPARSER PROCESSING THEORY'I:
</SectionTitle>
    <Paragraph position="0"/>
    <Paragraph position="2"> (This is a linear representation of the thepry baing processed. The endpoints are 0 and 30, but the words occupy-only the middle part of the utterance.)</Paragraph>
  </Section>
  <Section position="6" start_page="22" end_page="22" type="metho">
    <SectionTitle>
STARTING AN ISLAND
</SectionTitle>
    <Paragraph position="0"> (Now the lead-in transitions are being created, along with the monitors far syntactic categories which may precede the newly constructed configurations.</Paragraph>
    <Paragraph position="1"> Configurations along the lead-in path are all assigned a  index table for &amp;quot;winterw. Th* lead-ln transitions to configuration 1 havd already been Constructed, so they are not remade. Now we are ready to choose configurations to extend. The pool of candidates for extension contains confi~urations 2 and 6.)</Paragraph>
  </Section>
  <Section position="7" start_page="22" end_page="22" type="metho">
    <SectionTitle>
SELECTED CONFIGS (6) FOR EXTENSION
</SectionTitle>
    <Paragraph position="0"> [Only this one is chosen because it has a hi~har score than configQration 2, since the use of a noun as a head noun of a noun phraac! is more likely than its use au a  {We must ba~ln axecutinp the path at the first transition, because no part of it has been executed before. Later we will see that it is possible to begin execution of a path in the middle, since th* repister contents are stored at each step.]</Paragraph>
  </Section>
  <Section position="8" start_page="22" end_page="22" type="metho">
    <SectionTitle>
DOING JUMP ARC FROM 5:NP/ TO 4:NP/ART
DOING JUMP ARC FWM 4:NP/ART TO 3:NP/QUANT
DOING JUMP ARC FROM 3:NP/QUANT TO I:NP/ADJ
DOING CAT ARC WITH WINTER FROM 1 :NP/ADJ TO 6: NP/N
DOING POP ARC FROM 6:NP/N
TEST FAILED
</SectionTitle>
    <Paragraph position="0"> {Tha test failed because thzre is no determiner, and MINIGRAMMAR requires that singular, undetermined nouns can be complete noun phrases only if they are mass nouns. &amp;quot;Winterv is not marked as a mass noun in our dictionary, hence it will not parse as a complete noun phrase. )  {No monitor exists in the WFST for a NP/ at this placz, so the arcs (in NINIGRAMMAR there is only one) which could push for a NP are processed bottom up in exactly the same manner as the two arcs which couqd use a noun at the beginninp of the island.)</Paragraph>
  </Section>
  <Section position="9" start_page="22" end_page="22" type="metho">
    <SectionTitle>
ALL ARCS TRIED AT THIS CONFIG
</SectionTitle>
    <Paragraph position="0"> (Now the thewy has been procdssed. There followa a summary of the proposals, monitors, and notlces constructed. The syntactic wore assigned to the theory is ~ivon -- here just the acore of the conutituent constructed. Then there is a summary of ~tatistics.)  {This time there are monitory in the WFST, one which is looking for a NP start in^ at position 12 and one which is looking for a NP ending at position 21. One transition is sufficient to satisfy both of these, ant! the preposition needed to complete a PP/ is monitored</Paragraph>
    <Paragraph position="2"/>
  </Section>
  <Section position="10" start_page="22" end_page="23" type="metho">
    <SectionTitle>
NPI MAY LEAD TO CONFIC 11
</SectionTitle>
    <Paragraph position="0"> {This is caused by the fact that there was a monitor Tor a , noun phrase ending at confipurat ion 11 -- the one craated when constituent 1 was made. The transit ion which would bd 8at up is the transition Just created, so it is not remade.</Paragraph>
    <Paragraph position="1"> All of the processinp which resulted from the completion of a constituent is finished; however there are honitors still to be set for configurations alone the path.)</Paragraph>
    <Paragraph position="3"> for, its associated test (if any) th~ theory which is to be notified when the monitor is satisfied, and the configuration and arc causing the monitor, monitors must be mad* anew each time one of the elements changes, although some of the list structure can be shared, hence thz seeming proliferation ofmonitors.)  part of the currant theory. It doas not compllte a pqth or cause any further action. If it had a trrminatinff configuration, i.e. if a transition other than a POP tranai tion hao been traced , the terminatinp configuratdon would have been placed on the list of possible confipurations to extend.}</Paragraph>
    <Paragraph position="5"> FINISHED THEORY 2 WITH SYN SCORE 10 {The processing of this theory tbok approximately 4.5 seconds. ) This example has shown the trace produced by runninp SPARSER on input which is analogous to the example presented with illustrations of the map in Section Four. The Interested reader is urged to draw his own maps while reading the follow in^ Pact! 56 examples in order to best understand the dynanic operation of SPARSER.</Paragraph>
    <Paragraph position="7"> This example is more realist ic t ban the previous cntb -- it shows the operation of SPARSER in the context of an pt terance which has been 2utornatically segmented and labelee, ~ith the lexical rr;trieval and match component in operat ion. It demonstrates how SPARSER can help to select the best set vf words from a or complex word lattict. This example uses the SPEECHGRAMMAR described in [4].</Paragraph>
    <Paragraph position="8"> The utterance &amp;quot;What is the registration fee?&amp;quot; was spoken by an adult male speaker in a quite room and was record on tape.</Paragraph>
    <Paragraph position="9"> Tha ut teranca was automatically diqit ized rind passed through the warnentat ion and labelinr routines of the BBN speech understandins ~ystem. The initial scan of the utterance, usinp the lexical retrieval component, produced a w~rd lattice of fifteen entries, includin~ several for inflectional endings. (In this version of thl system, they were not combined with the root form into a single word hatch, and hsncr? could match evan without a root word.) Tha format for a word match is:  The two best matches, for &amp;quot;what1! and ftrzgistrationll, appear to be good candidates for a theory, so we begin by build in^ and procasring that theory.</Paragraph>
  </Section>
  <Section position="11" start_page="23" end_page="23" type="metho">
    <SectionTitle>
STARTING AN ISLAND
STARTIHG AT LEFT END OF SENTENCE
</SectionTitle>
    <Paragraph position="0"> (Itno~ing that it is not necsssary to go through the usual startup procedure for islands when beginning an island at position 0, SPARSER starts with a configuration for state 3/ at position 0.)  (Hare all the words which can start quantifiers, like &amp;quot;a hundredw or &amp;quot;point fivam, ard proposed. The grammar does not preclude a quantifier following a que$tion-determihzr, e-g. &amp;quot;What three men traveled to Spain'llI. ) {MONITORING [ INTEGER ZERO NO POINT A] For considerations of space, long listings of monitors and proposals in this example will be compacted as shown here. Such alterations to the actual trace produced will be surrounded by brackets.)  {This is an ex~mple of he fallibility of using only context free tests on partial paths, The parser thinks it has succassfully reached state NP/HEAD, while in fact tbis cannot be the case because no head nouh has bean dlscovarad for the noun phrase. Thus it is incorrect to predict relativz clauses at this point. This issue will be discussed in more detail be1ow.J</Paragraph>
  </Section>
  <Section position="12" start_page="23" end_page="23" type="metho">
    <SectionTitle>
TEST FAILED
</SectionTitle>
    <Paragraph position="0"> {A question-determiner alone cannot bs a complete noun phrase; although this is permitted by considering &amp;quot;whatw as a QWORD as in tranaition #2.)</Paragraph>
  </Section>
  <Section position="13" start_page="23" end_page="23" type="metho">
    <SectionTitle>
STARTING AN ISLAND
</SectionTitle>
    <Paragraph position="0"/>
    <Paragraph position="2"> {This notice is in response to the look-ahead test on the push arc to atate R/NIL. Since 'vfee'l can start a reduced re&amp;ative clause, it is noticed, but there is not a specific monitor set up becauue the arc within the relativd clause network which will actually process the  {Processing the proposal-s just made results, notably, in the detection of the word lvotherv between vv~hatvl and wre~istrationlv, but the word match score is very low. Word matches for *isw and Itare&amp;quot; from position 3 (next to &amp;quot;whatvv) to position 4 are also found, but since they do Page 61 not fill the pap, the event scores are low. The bzst event is that for the word &amp;quot;faen. Procassin~ it is fairly uninteresting, since it completes no constituent, so we will omit the trace of that event. After it has bean processed, however, the best event is that for the word &amp;quot;thevf and the theory just created. ]  ha format of this noun p5rase is slightly different from that in the previous example because tpe ~tructure building action for noun phrases in SPEECHGRAMMAR is different from that in MINIGRAMMAR.</Paragraph>
    <Paragraph position="3"> There ara many places in the SPEECHGRAFIMAR which push for noun phrases, and since there were no monitors in the WFST which can us&amp; thiv constitusnt, all of them must be tried, resulting in a numbdr of predictions and notices. )  {There arc two arcs entwine state R/WH which use the words l'whichn and llwhomv. There is a check made to see that duplicate proposal8 alre not actually communicated to the control component, although they appear to be duplicated in the trace. }  first two have bean executed, resulting respaatively jn failure and the completion of a constituent with all the processing that entails. Now the third path is still pending and is about to be executed.]  36,38,40,41,43,46,47,49, and 51 because of the monitors set when the first constituent was found.) SELECTED CONFIGS (55 54 39 37) FOR EXTENSION {Because these ara the maximally scoring configurations from ths large pool of possibilities.)  {Here is an example of a constituent which has features attached to it. The feature NPU can be tested by tne semantic component to determine that the constituent is a noun phrasa utterance. If nscessary, it could also be tasted on a PUSH S/ arc in the prammar; since there are aome times , e.~. durin~ the construction of a sentantial complement, when an embedded sentence must  discovery of the noun phrases which were not monitored for. 1 (Processing the proposals from this theory results in the bavt event being the one for &amp;quot;isw in the last gap. The word &amp;quot;arevt also fills the gap, but the lower lexical acore prevents ths event for it from surfacin~. If it were syntactically procasued, however, no new theory would be created since tha completed string would be ungrammatical.)  This example was run with a vary simple, mechanical control structure. After the processing of the initial theory, the proposals which had bean made by SPARSER were processed by th+ lexical retrieval component and the results added to the word lattice -- a process which can sat off monitors and result in the creation of event notices. The ~v*nts are scored by a combination of the monitor score assigned by SPARSER and the lexical score asslpnad by the word match component. In this sentence, syntax and lexical score alone ware suffhient to make ths besb scorinp event at each step be one which resulted in a correct extension of the theory.</Paragraph>
    <Paragraph position="4"> Vz now ahow how the same utterance use@ in the prdvious example cah be recognized when dkfferznt theories qre crsated and when avants and theories are processed in a diffarent orda? from that in Example 2. Suppose that after the initial scan of the Ptterance the semantic component created two thzorias, one for the words llwhatw and &amp;quot;feeff and the other fori the wobds I'vhat1' and nrzristrationtl Let us see what happen9 in SPARSER when we hepin by procassipg these two tbsories in sequence.</Paragraph>
    <Paragraph position="5">  {The processing of this thzory iq very similar to that of the first theory in the previous ~xample, and will not be commented upon here. The purpoae in show in^ it is to provide a map, part of which the next call to SPARSER will trace. )</Paragraph>
  </Section>
  <Section position="14" start_page="23" end_page="23" type="metho">
    <SectionTitle>
STARTING AN ISLAND
STARTING AT LEFT END OF SENTENCE
</SectionTitle>
    <Paragraph position="0"/>
  </Section>
  <Section position="15" start_page="23" end_page="23" type="metho">
    <SectionTitle>
STARTING AN ISLAND
STARTING AT LEFT END OF SENTENCE
SELECTED CONFIGS (1) FOR EXTENSION
</SectionTitle>
    <Paragraph position="0"> {Upon picking up this configuration to extend it, SPARSER finds thd transitions which were created durin~ the processing of the word I1whatfr by tha previous theory . It tttracesll them all, that is, it does not recreate them but simply puts the transition numbers on a list which will form part of the syntactic infornation associated with the current theory. The tracing process also involves the creation of monitors (and notices, uherr applicable) for constituants along the path.</Paragraph>
    <Paragraph position="1"> These monit~rs and notices must be remade, since the previous monitors will activate only th6 previous t heary .</Paragraph>
    <Paragraph position="2"> Due to the recursive nature of the tracin~ process, the transitions a not necessarily followad in t.he same order that they were originally created, nor are the monitors made in exactly the same order.</Paragraph>
    <Paragraph position="3"> Notice that the many arcs which we tried but which did not result in the creation of transitions in the previous theory are not retried here.}  [This does not mean that confiauration 6 was just created. Since it already existed in the map, having been craated during the p~ocessinc of the previous theory, the configuration number is merely put on the list of configurations in the current theory.)  {No proposals were made here because proposals are not theory dependent; that is, the word proposals which were made during the processing of the previous theory resulted in some words baing placed in the word lattice whLch were noticed here. Remaking the proposals would not lead to the discovery of any new information.) TRACfNG POP TRANS 8 FROM 10:NP/HEAD:3(29) {The processinp of thz island for llregistrationw is idBntical to that in the last example, so the remainder of tha trace will be omitted. Thz total processing took 12.2 seconds.] {Let us now process the event which adds the word &amp;quot;theu to ths thleory just processed. This will result in the creation of a conutituent event.)  his constituent cannot ba used immediately by this theory because it contains a word (&amp;quot;feevf) which is not in the thaory. Therefore a noticd is sdnt to Control which may be turned into an event at some later time. Nothing further is done with this constituznt at this time, i.e., no transitions using it are created. It is, howevsr, placed in the WFST for later use.)  {Tljis constituent is completely consistent with the current theory, that is, it is composed only of word matches already in the theory, and there are no nonitors in the WFST for it, so it is proce~sed bottom up as we have swp before. )  {Now we will process the constituent event for the theory just created. Because of the constituent for &amp;quot;the registrationm there art3 now monitors in the WFST for a noun phrase beginning at position 4, so ths appropriate transitions are made.)</Paragraph>
  </Section>
  <Section position="16" start_page="23" end_page="23" type="metho">
    <SectionTitle>
SYNTAX PROCESSING EVENT FOR THEORYiI3 WITH CONSTITUENT #1
TO GET NEW THEORYf4
0 WHAT 3 4 THE 6 REGISTRATION 19 FEE 23
</SectionTitle>
    <Paragraph position="0"> {Processing begins exactly where it left off when the constituent was made -- thz constituent is semantically evaluated with rwspact to this theory so that the constituent weight may be altered. In this case, howevar, Semantics has been turned off, so them is no increment in the score.)</Paragraph>
  </Section>
  <Section position="17" start_page="23" end_page="23" type="metho">
    <SectionTitle>
SYN WEIGHT + SEM WT = 15 + 0 = 15
NP/ WAS PUSHED FOR AT CONPIG 34
</SectionTitle>
    <Paragraph position="0"> PUSH NP/ TRANS #51 FHOll 34:FOR/FOH:I+(l) TO 96:T0/:27('11) {Similar transitions are set up for all 9 other confipurationu where an NP/ was used in the previous theory. The monitors set by these path8 are copied from the previous theory, 80 there is no indication h+rz of a  {This avant took only 9.7 seconds.) The procrssinp of the final event, that which adds the word &amp;quot;isw to the theory Just created, will not bz shown. ThesB examples have shown that SPARSER is a u~oful tool in th* automatic reco~nition df speech. The t iminp measurements indicate that considerable procesain~ iu done when the parser is forced to work in bottom ~p mode, especially with a large Eranmar Of course there is some implementaion ovdrhead involvdd in doins the timin~s themselves. If the paruinc al~orithm were to be carefully recoded in assembly langua~e a speed up of at least a factqr of 20 (and perhaps much nore) could be achieved. Another way to cut down the time-con sum in^ processing mipht be to at-tempt to obtain gar6 semantic ~uidance. For sample, if the semantic hypothesis asuociated with a theory indicates that a particular noun is likely to be used in a noun phrase modifier (ern&amp; lttomorrowtt), than SPARSER should be able to take advanta~r of this information by scorinp the PUSH NP/ transition from a confi~uration for atata PP/ (i.e. to pet something like 'Iby tomarroww) hi~har than those PUSH NP/ transitions for other syntactic slots. In fact, th* others may not need' to be constructed at all. The Erammar could also be further tuned to eliminate soma spurious predictions and reduce the time spent following erroneous paths.</Paragraph>
    <Paragraph position="1">  One of the weak points of the current system is the fact that some context Information is not uued until a path is complete, result in^ in the creation of false paths and predictions which should not have been nade. This is partly miti~ated by the fact that this avoids a too rea at dependence on left context and allows the creation of partial paths which may be followed if an earlier word is changed.</Paragraph>
    <Paragraph position="2"> It is important, however, to rininize the number of predi'ctions which are made and to ~ake the predictions as accurate as possible. In this r~~ard, it is unfoqtunatc that the currqt system makes predictions on the left of an island solely on the bavis of the first word in the island and a makes predictions 6n the right end from confi~urations which, if context sensitive tests had besn done alone the path, would never nave been created.</Paragraph>
    <Paragraph position="3"> One way to help tighten the predictions would be to take each context free path threu~h an island and walk it in a special mode after the island has been processed but before predictions are communicated to the cohtrol component, This mode would set and check registers, assuming that any tests which require unknown left context are true. Only if the path did not fail under this mode of operation would the pr~dictions at aither end pf it be made. If a really efficient way of handlinp unknown left context and of storinp this informafion ware developed, it could be uued in place of the context free pass in the first place, thus a1 lminatinp a 11 inconsistent paths.</Paragraph>
    <Paragraph position="4"> The problem with st or in^ all possible contexts is that they must be ~ecomputed each time a new atep is added to the path.</Paragraph>
    <Paragraph position="5"> Nia ie relatively easy if th* next step is taken to the ripht of an 2xistin~ path, since ATN s are mora auitdd to left to ripht processing, but it becomes extremely complex when a transition is added to the left end of a path (or set of paths) or when a transition Joins two sets of paths togethzr. To be absolutely sure that no contexts haw been misued, all the paths would have to be walked and their contexts reprocessed and copied in whole or in part (since the new step may be wronR, the old context rnuvt be preserved.) Of eoursa this is not the only approach which could be .used -- a merninp technique like that of Earley's algorithm might be feasible, if the structure of the crammar were also chanped to make it less left to ri~ht oriented.</Paragraph>
    <Paragraph position="6"> One ~reat ytren~th of the system is its ability to store and merge information in such a way that it does not hava to be redone when tha context is changed. For example, once an arc has bean tried with a particular word match, a transition will be created if the arc may be taken and the arc will ba removed from further consideration if it may not be taken. Then, if the configuration should ever be reached with the same word match apai'n (perhaps in a later theory) not only will any relevant transitions be reco~nized without havinp to po throuyh the work of r*-creatin~ them, but also ne arc8 whic_h had prev,Lou3+ly failed, Y~ ec hk ~&amp;~kd* Another feature of SPARSER. in the fact that it \~i4~ ~!trsiyned and implemented with many unsolved problar~ :jnd unsviiiluble bats in mPnd, and therefore many wl~oleall have been lef't on which to lfhooku further developments. For exnmle, a1 t h~urh pro::~l!ic verification of constituents is not yet available, the scorinr rntrchani:~~ for ~wnst ituenty is structured in such a way that it would be easy to include the results of verification by prcsodics (or any other component). Th* oripinsl implenen'tation of SPAHSEH used a depth first search but Qas implemented in such a way that the chanre to dodified breadth first was quite simple. This foresight has paid off in a flexible systen rrhich has shown that it can be readily experimented with in o~der to explore many still unsolved problems ccrncoerning the nature and use of syntactic information in understandinp A tremendous amount of information in speech is conveyed by proscdic features: stress, intonation duration, loudness, pauses, pitch. For example, if John mumbles to Bill, ??The mailman left something for you,&amp;quot; Bill may reply aithar IrWhat?&amp;quot; with much energy and a sharply risih'p intanation or &amp;quot;What?I1 with a flat o'f. falling intonation. In th~ first case John is vzry likely to shout 111 said, Ths mailman left something for you interpreting lvWhat?lv to mean &amp;quot;What did you say?&amp;quot; whereas in the second case he is likely to say something like *A package from youp rnothw,&amp;quot; interpreting ltWhat?vv to mean &amp;quot;What is it?&amp;quot; To i~nora prosudics iu to ipnore a sourca of information which has bssn shown repeatedly to be an extremely important factor in human underatandinp.</Paragraph>
    <Paragraph position="7"> Consider the following examples of sentences and sentence fhpmehts which illustrate some of the ways prosodie8 are used:  semantically correct, syntactically consistent phrases which are nonethsless wrong. If the constituent &amp;quot;speech understandingtf were identified and relied upon, it might be very difficult to produce a correct analysis of the utterancs: wBd~a~~d of peculiarities in his speech, understanding Joa is not easymN Besides indicatinp syntactic boundaries and/or providinp intonation contours for certain constituents, prosodic features can be used to mark emphasis, introduce new topics, cpnvey information about the speaker s internal mental and t:notion:il state (e.p. whether hl is teasinp or ~erioua), nnd probably more. It is particularly interestin~ to note that some well known phenomena ~uch us l'pranouns are almost never atrt*suedU :in3 &amp;quot;ip discourse wh~n new topic word 1:: gention~bd it is alcost always stressedm hhve very naturtil explainat ions in ll~ht of what we know about acoustic processing. Stressed war-da are ~anerdlly easier to identify because there is less acoustic aztbicuity, but unstressed words may differ creatdy f rorn their ideal pronunciation and hence are harder to reliabl y identify Pronouns refer to antecedents which are presu~ably known to the listener, so he can anticipate them or at leaut verify them easily, hence they need not have mod acoustic characteristics.</Paragraph>
    <Paragraph position="8"> A new topic may not have been anticipated, so the listener will have to depend heavily on identify in^ the word from acoustic information alone and the speaker can provide this extra reliable information by stressine the word.</Paragraph>
    <Paragraph position="9"> Unfortunately, not a przat deal is known about either the acoustic correlat~s of prosodic features or the ways in which they are used. Many- of the rules which have bezn developed thus far are speaker dependent and are snfficient for convey in^ information but are not necessary. This makes them difficult to use in thz analysis mode. Althou~h a good start has been made in exploring pr-osodies (see, for example, Lea [52, 131 and Eates and Wolf [8]), much more work remains to be done before prosodic Page 83 information can be reliably used by speech understandinc systems. SPARSER could use prosodic information in several ways.</Paragraph>
    <Paragraph position="10"> Verification of constituents would be a prest help, but local proaicid information could be used ev*n earlier in the parsing p~oceas. For example, if maJor constituent boundaries could be accurately determined, then inatead of both POPing a constituent and continuing it in parallel, as is done now, one alternative could be chosen inatdad of the other on the badis of prosodic informal ion. If, as is more likely, yome major boundariey could be reliably detected, then it would be easy to revise SPARSER to begin procsssinp at such places even #within an island at states which can begin constituents. This would a~ain reduce the number of partial paths created when pars in^ an island.</Paragraph>
  </Section>
  <Section position="18" start_page="23" end_page="23" type="metho">
    <SectionTitle>
7.3 EXTENSIONS AND FURTHER RESEARCH
</SectionTitle>
    <Paragraph position="0"> One of the obvious extension8 to a basic speech underatandinp . . system is to relax the restrictions on the input to the system. Syntactically, this can mean removing the requirement that the initial utterance be prammatical. Since people frequently speak unprammatically in informal discourse, this is a natural step to want to take.</Paragraph>
    <Paragraph position="1"> In order to extend SPARSER to handle such input, sevaral approaches are possible. Certain types of errors may be called errors of style (and may not be called errors at all by some people) such as the use of tlain'tlt and the occurrence of a prepositjon at thd and of a sentence. These resularit ies r?:ly simply be declared ~ramrnatical by ~0difyin.r the pr&amp;+nnur to accept them. Hany speech errors have hern shown to fQcllc-\: rt~\:l~~:* piitterns and hence gay ht? :~vn:tble to t!lis :ippl*~;~ck.</Paragraph>
    <Paragraph position="2"> t her PC~V~!OII 1 spec- i 1&amp;quot;ic t tbst s \~?:ic!? :jp;-kS:+!- p .! the arcs at' the r, f 1 t ?r*~hikit i+t~:~-l~net is n t 0 check for nu~ber a~~trtwven t bet wet'11 s1:5.-~vt :iz:q verb or between determiner itnd noun (c.~. *I is SPY Y=~~*V severe restrictions on this rule.&amp;quot;). In this case, rather t!?ap renovinp the tests from the yramar it w~uld ~CJ rcre suita?le tc &amp; modify the^ SP th;it if t 1 t -arc 1 still hr L~.&lt;C*Y, thourh it h a :nvch reeuce~ w~~ir!~t or \:it h ;in icbicat icn in sere re~ister that an error has occurred. Cne way to in2lzxent this would be to have all tests return a nu~bttr as their value indicating how well they succeeded on some sca&amp; fr~a &amp;quot;perf&amp;quot;ect1y&amp;quot; to ltnot at all&amp;quot;.</Paragraph>
    <Paragraph position="3"> Not all arc tests are of this reIaxablc! nature, however, since certain types of errors are so rare, if they occur at all, that they may be Judged ~nacceptati~. Examples of such tests are the case checks for pronouns (2.~. *&amp;quot;I pavz it to he1'] and the requirement that a verb modifyinp. a noun must be in either the present or past participle form (2.g. Itthe sincinc brookn vs.</Paragraph>
    <Paragraph position="4"> *&amp;quot;ths sinp brook1').</Paragraph>
    <Paragraph position="5"> These methods would not allow all tvpes of qran~atical errors to be handled (in particular it irnores the problen of constituent ordering errors such as ItThrow !lama from the train a kissw), but uould handle many of the most common syntactic errors.</Paragraph>
    <Paragraph position="6"> &amp; experiment hoepinr in mind that SPARSER is not intended to be a mo'dd ~f human svntactic analysis, it is nonetheless reasonable to ask whether there are anv ~imilarities which may be seen. The followinp experiment is suppest ed with the hypothesis tkit it will indicate that people do considerable processing at the end of svntactic constituenks in a way similar to-some repister set tin^ and testing actions and szmantic (or othw) verification The experiment is this: a subject is seated in front of a switch which hi is asked td press whenever ha is surd that he is hear in^ an anomalous sentence, He is then presented with a pulaber of recorded utterances, some of which are incorrect, e.~. {'The cat and dog which live3 next door are friendly.</Paragraph>
    <Paragraph position="7"> I sawma red big barn on f he farm.</Paragraph>
    <Paragraph position="8"> 1 hypothesize that the subject will indicate the presence of an ewer at a point shortry after the end of the constituent in uhich the error occurred more often thand shortly after the earliest possible place where the error coul be detected.</Paragraph>
    <Paragraph position="9"> In conclusion, it is obvibds t~gf there is much work yet to be done in the problem.of speech undzrgtanding, but it is hoped Dha( the system presented be- has no.t, only advadczd our current  This appendix lista the 351 words which wwa fn the dictionary of the BBN speech understanding ayatem wtkn the examples in Chapter Six ware run (July 19n). (A 569, word dictionary and one with 1000 entries arc now available. After the listing of the wards in the dictionary, Obey ara.broken into syntactic classes, with the number of words in. each alas3 indicated beside the class name. Finally, the ~yntactic features are given together with a list of the word8 which carry each feature. Features may be of the form FEATURE, (FEATURE), or (FEATURE VALUE).</Paragraph>
    <Paragraph position="10"> This is not a listing of tha dictionary as dt appears to the systerrt, but rather a derived crosv reference which indicates the various parts of speech and ~yntactic features for each word Tha words:</Paragraph>
  </Section>
  <Section position="19" start_page="23" end_page="23" type="metho">
    <SectionTitle>
(A ABOUT ABOVE ACL ACOUSTICAL ACOUSTICS ADDITIONAL AFFORD AFTER
A1 AIR AIRPLANE ALL ALREADY ALSO AM AMHERST AMOUNT AN AND
ANTICIPATE ANY ANYONE ANYWHERE APRJL ARE ARPA ARRANGE ASA ASK
&amp;SSOOIATION ASSUME ASSUMPTION AT ATTEND AUGUST AVAILABLE BATES 3E
BECAUSE BEEN BEFORE BEGINNING BEING BIG BILL BpNNfE BOSTON BDTH
BREAKDOWN BUDGET BUS BY CALIFORNIA CAN CANCEL CAR CARMEGIE CENT
CHANGE CITY COLARUSSO COMPUTATIONAL CONFERENCE CONTXWUE COSELL
COST COSTING COSTS COUNTRY CRAIG CURRENT DATE DAY-E IAY DECEMBER
DENNIS DID DIVISION DO DOES DOLLAR DONE DUE@TO DURING EACH EIGHT
EIGHTEEN EIGHTEENTH EIGHTH EIOHTY EITHER ELEVEN ELEVENTH END
ENGLAND E~OUGH ESTIMATE-N ESTIMATE-V EVERY EVERYONE EXPECT
EXPENSE EXPENSIYE FALL FARE FEBRUARY FEE FIFTEEN FIFTEENTH FIFTH
FIFTY FIGURE FINAL FIRST FISCAL FIVE FOR FORTY FOUR FOVRTEEN
FOURTEENTH FOURTH GET GETS GETTING GIVE GIVEN GIVES GIVItJG GO
GOES CO3'NG GONE GOT GOTTEN CROUP WAD HALF HALVES HAS HAVE HAVING
HE HER HIM HIS HPw HOWMANY HOWMUCH HUNDRED 1 IF IFIP IJCAI IN
INTERNATIONAL IS IT JANUARY JERRY JOHN JULY JUNE KNOW L.A.
LAST LATE LEFT LINDA L,INGUISTICS LIST LONDON LONG LOSeANGELES LYN
LYNN. MADE MAKE .MAKES MAKMOUL MA~INC MANY MARCH MASSACtIUSETTS MAY
ME HEETING MEMBER MISCELLANEOUS MONEY MONTH MORE MOST MUCH MY
NEED NEW@YORK NEXT NINE NINETEEN NINETEENTH NINETY NINTH NQ NOT
NOTE NOVEMBER NOW. OCTOBER OF ON ONE ONLY OH OTIIER OUT OVERHEAT,
PAJARRDPDUNES PARTICIPAIJT PAUL PAY PENNSYLVANIA PEOPLE PER PEHSON
PHONOLOGY PITTSBURGH PLACE PLEASE PLUS PRINT PROJECT-N PROJECT-V
PURPOSE QUARTER REGISTRATION REMAIN REST REVISE RICH RICHARD
ROUNDeTRhIP SANTAeBARBARA SCHEDULE SECOND SEND SENDING SENDS SENT
SEPTEMBER SkVEN SEVENTEEN SEVENTEENTlI SEVENTH SEVENTY SHE SINCE
SETE SIX SIXTEEN SIXTEENTH SIXTH SIXTY SO SOCIETY SOME SOMEONE
SPEECH SPEND SPENDING SPENDS SPENT SPRING ST.LOUIS START STATUS
STOCKHOLH SUMMER SUPPOSE SUPPOSED SUPPOSITION SUN SWEDEN TAKE
TAKEN TAKES TAKING TEN TENTH THAN THANKQYOU THAT THE THEIH THEM
THERE THESE THEY THIRD THIRTEEN THIRTEENTH THIRTIETH THIRTY THIS
THOSE THOUSAND THREE TIME TO TOO TOOK TOTAL TRAVEL TRIP TWELFTH
TWELVE TWENTIETH TWENTY TWO UNANTICIPATED UNDUDCETED UNSPEilT
UNTAXEN NUS VARIOUS VISIT WANT WAS WASHINGTON WE WEEK WENT WERE
WHAT WHEN WHERE WHICH iJHO WHOM WHOSE WILL WINTER WISCONSIN WITH
WITHIN WORKSHOP YEAR YES YOU)
</SectionTitle>
    <Paragraph position="0"> The syntactic categories:</Paragraph>
  </Section>
  <Section position="20" start_page="23" end_page="23" type="metho">
    <SectionTitle>
(ADJ 23 (ACOUSTICAL ADDITIONAL AVAILABLE BIG COI-IPUTATIONAL
CURRENT EACII ENOUGH EXPZNSIVE FINAL FISCAL INTERNATIONAL
LATE LEFT LONG MANY MISCELLANEOUS OTHER UNANTICIPATE~
UNBUDGETED UNSPENT UNTAKEN VARIOUS))
</SectionTitle>
    <Paragraph position="0"/>
  </Section>
  <Section position="21" start_page="23" end_page="23" type="metho">
    <SectionTitle>
BEGINN-ING BREAKDOWN BUDGET BUS CAR CENT CHANGE CITY
CONFERENCE COST COUNTRY DATE DAY DIVISION END ESTIMATE-N
EXPENSE FALL FARE FEE FIGURE GROUP HALF HALVES LINGUISTICS
LIST MEETING MEMBER MONEY MONTH MUCH NEED NOTE OVERHEAD
PARTICIPANT PEOPLE PERSON PHONOLOGY PLACE PROJECT-N
PURPOSE QUARTER REGISTRATION REST ROUNDQTRIP SCHEDULE SITE
SOCIETY SOME SPEECH SPRING STATUS SUMMER SUPPOSITION
</SectionTitle>
    <Paragraph position="0"/>
  </Section>
  <Section position="22" start_page="23" end_page="23" type="metho">
    <SectionTitle>
BOSTON CALIFORNIA CARNEGIE C:ILARUSSO COSELL CRAIG DECEMBER
DENNIS ENGLAND FEBRUARY IFIP IJCAI JANUARY JERRY JOHN JULY
JUNE L.A. LINDA LONDON LOS@ NGELES LYN LYNN MAKHOUL MARCH
MASSACHUSETTS MAY NEWCYORK NOVEMBER OCTOBER PAJARROeDUNE-S.
PENNSYLVANIA PITTSBURGH RICH RICHARP SANTA~BARBAHA
SEPTEMBER ST.LOU1S STOCKHOLM SUR SWEDEN WASHINGTON
</SectionTitle>
    <Paragraph position="0"/>
  </Section>
class="xml-element"></Paper>
Download Original XML