File Information

File: 05-lr/acl_arc_1_sum/cleansed_text/xml_by_section/metho/79/j79-1026_metho.xml

Size: 64,620 bytes

Last Modified: 2025-10-06 14:11:10

<?xml version="1.0" standalone="yes"?>
<Paper uid="J79-1026">
  <Title>American Journal of Computational Linguistics</Title>
  <Section position="1" start_page="0" end_page="0" type="metho">
    <SectionTitle>
JUNCTION GRAMMAR
AS A BASE FOR
NATURAL LANGUAGE PROCESS I NG
</SectionTitle>
    <Paragraph position="0"/>
  </Section>
  <Section position="2" start_page="0" end_page="0" type="metho">
    <SectionTitle>
Abstract
</SectionTitle>
    <Paragraph position="0"> Junction Grammar, a model of language structure developed by Eldon Lytle, is being used to define the interlingua for a machine-assisted translation project. Junction Grammar representations (called junction trees) consist of word sense information inter-related by junctions, which contribute syntactic and semantic information. The first step of the current translation system is interactive analysis. During this step, the program interacts with the human operator to resolve ambiguities and then produces a junction tree representation of the meaning of the input text. The second and third steps of the translation process are automatic transfer and synthesis into one or more target languages. For each target language the transfer step makes adjustments on each junction tree, if needed, before sending it to the synthesis program for that language. This translation system is currently under development at Brigham Young University in Provo, Utah. Present lexicons for English analysis and Spanish, German, French, and Portuguese synthesis contain about 10,000 word senses each.</Paragraph>
  </Section>
  <Section position="3" start_page="0" end_page="1" type="metho">
    <SectionTitle>
TABLE OF CONTENTS
</SectionTitle>
    <Paragraph position="0"> Assuming that semantics must be taken into account in a translation system, there are at least two major classes of intermediate representations to choose from for use in a translation system. One class of repre.sentations is non-linguistic. In this class syntax is ignored and use is made of various conceptual units and relations (e. g. Schankl s conceptual dependency) . Another class of representations is linguistic. In this class syntax and semantics are combined into a syntacto-semantic representation which is connected to surface structure in a fairly direct fashion (e.g. Montague grammars [I] and Junction Grammar [2]).</Paragraph>
    <Paragraph position="1"> We have chosen to test Junction Grammar as the interlingua for a machine-assisted translation project. The basic concepts of this model of language were developed by Eldon Lytle in the late 196OVs, but, as with many theories, it has been under continual development ever since.</Paragraph>
    <Paragraph position="2"> For subject-verb-object sentences, junction trees look much like the familiar phrase markers. Let us then preview the steps of the Junction Grammar translation process using an extremely simple sentence.</Paragraph>
    <Paragraph position="3"> Given the English surface string, &amp;quot;He needs it&amp;quot;, English analysis would produce the junction tree : Note that the tree has an order independent of the surface word order of the input text. The function m(x) produces the word sense or sememe of its argument. The values of this function have been implemented as positive integers called semantic indices. The plus signs under the categories PV (verbal predicate) and SV (verbal sentence) stand for adjunction, one of the basic types of junction recognized by Junction Grammar.</Paragraph>
    <Paragraph position="4"> Suppose we want to synthesize this junction tree into French.</Paragraph>
    <Paragraph position="5"> Transfer would adjust this tree because French expresses the meaning of 'needt as 'have need oft. Then French synthesis would perform lexical selection and ordering to produce the string &amp;quot;Ila besoin de ca&amp;quot;, or &amp;quot;11 en a besoinn.</Paragraph>
    <Paragraph position="6"> The following chapters will discuss junction trees in more detail and then explain how the analysis, transfer, and synthesis steps are currently implemented.</Paragraph>
    <Paragraph position="7">  11. Introduction to Junction Grammar.</Paragraph>
    <Paragraph position="8"> A. Basic Relative Modifiers.</Paragraph>
    <Paragraph position="9"> Consider the following sentence : (1) Every monkey that swallows that, gets indigestion.</Paragraph>
    <Paragraph position="10">  Utilizing a modification junction * called subjunction, consider the following as a junction marker for (1) : (Consider the terminal words as names of sememes and consider quantifiers to be of a noun- type category) .</Paragraph>
    <Paragraph position="11">  swallows that The problem with the above junction marker is that it is not clear which that is the relative pronoun. Following Montague (see also Partie [3] and Gabbay [4]) we could subscript the subjunction operation and the appropriate that with respect to which the modification occurs: Every /&amp;quot;. gets indigestion</Paragraph>
    <Paragraph position="13"> swallows that The above tree would then be lexicalized as sentenc (1). If we had subscripted just the other - that then the mark~r would be lexicalized not as sentence (1) but rather as: (1A) Every monkey that that swallows gets indigestion. A readable way of handling such clauses is the method used in Junction Grammar: The proform with respect to which the modification occurs is directly joined to the noun being modified. Thus the structure takes the place of subscripts.</Paragraph>
    <Paragraph position="14">  swallows that The reader will notice that the above diagrams are not normal binary trees, but we call them trees or intersecting trees because they can be drawn as sets of cross-referenced trees.</Paragraph>
    <Paragraph position="15"> The follawing tree illustrates the notational variations of (a) drawing intersecting trees as cross-referenced sets of normal, trees, (b) drawing the junction symbol between the brothers (subordinate nodes) instead of next to the father (superordinate) node, and (c) optionally leaving out the junction symbol when it is  It is interesting to note that one simple method of generating in a context-free way such intersecting trees as those used in Junction Grammar is to employ context-sensitive rules as node admissability conditions. This method is essentially in harmony with McCawleyV s [S] proposal. For example, the rule A+BC allows the well-formed structure A Further, the rule</Paragraph>
    <Paragraph position="17"> ly dominating the sequence BC is a well-formed subtree provided that elsewhere in the tree a node F immediately precedes A and A immediately precedes a node G. Thus, as an example, the following tree is well-formed from the rules indicated.</Paragraph>
    <Paragraph position="18">  Though the tree is well-formed if the rules are node admissability conditions, FBCGH cannot be derived from this gramma? in a standard rewrite derivation way.</Paragraph>
    <Paragraph position="19"> If such production rules were to be used, their form would be: W --3 XY Such production rules are type 0 and not type 1.</Paragraph>
    <Paragraph position="20"> Bm Noun Complements.</Paragraph>
    <Paragraph position="21"> It is instructive to contrast the above Junction Grammar treatment of relative clauses with its treatment of noun complements. null It would seem that there are serious theoretical problems associated with the phrase structure rules *NP?rNP S and</Paragraph>
  </Section>
  <Section position="4" start_page="1" end_page="1" type="metho">
    <SectionTitle>
NP-N S of Transformation Grammar [6] .
RELATIVE CLAUSE RULE : NP+NP S
COMPLEMENT RULE: NP-N S
</SectionTitle>
    <Paragraph position="0"> (2) The embarrassing fact that John was able to learn frustrated his adversaries.</Paragraph>
    <Paragraph position="1">  able to learn to learn TG Treatment of Noun Complements and Relative Clauses. These rules propose to account for restrictive relative clauses and noun complements, respectively, both embedding an entire sentence to a nominal antecedent, encompassing both constituents with brackets labelled NP. It seems clear that the two structures in question are in fact related, i.e. in some sense similar, but not in the way suggested by the P-rules proposed to generate them. Specifically, in each case there seems to be an overlapping of constituents in the main clause with those in the subordinate clause: In the relative clause, an NP of the main clause coincides referentially with an NP of the dependent clause ; in the complement, a noun, or potentially, a noun phrase, of the main clause is equated referentially with the entire dependent clause. Thus sentence (2) is ambiguous over the relative clause and the complement readings. (Notice that one can replace that with which for the relative clause reading but not for the complement reading.] There is nothing in the P-rule formulation to make this overlapping of constituents explicit, however. Hence, a mechanism for checking the coreference of NP1s is required so that the relative clause transformation could apply to sentences embedded by NP-NP S and produce the appropriate relative pronoun. It is not clear, however, whether this mechanism is supposed to establish a coreference relation between the head N and the complement S in NP -+N S, since no T-rule seems to depend upon such a check. Moreover, there is no clear justification for using N rather than NP to the right of tlie arrow in the complement rule, since the head of a complement can have articles and modifiers too.</Paragraph>
    <Paragraph position="2"> Still more serious is what appears to be implied by the relative clause rule. Namely, the entire clause is bracketed with a nominal category (NP), suggesting that it, like the complement, is functioning in its entirety as a nominal constituent. The structural symmetry of these two rules results in a false generalization (the illusion that both clauses were nominalized) while failing to make explicit the generality which actually exists and is semantically crucial (the referential overlap between constituents in the main and dependent clauses). What is needed are structural representations which reflect the overlapping of constituents without violating what seem to be the correct categorizations. null The Junction Grammar solution is as follows:</Paragraph>
    <Paragraph position="4"> fact which fact</Paragraph>
    <Section position="1" start_page="1" end_page="1" type="sub_section">
      <SectionTitle>
Relative Clause Noun Complement
</SectionTitle>
      <Paragraph position="0"> In the case of the relative clause, the intersection occurred on categorially homogeneous nodes (on Nls) , whereas in the complement structure the intersection occurred on heterogeneous nodes (N /S) so that the entire subordinate clause intersected with an NP of the main clause.</Paragraph>
      <Paragraph position="1"> Subjunction can thus be seen to be of two basic types--fullsubjunctions, as in noun complements, and interjunctions, as in relative clauses. In full-sub junction the modifying constituent node is ~omp~letely subjoined, e.g. number five, or the boy John as</Paragraph>
      <Paragraph position="3"> On the other hand, in interjunction the modifying constituent is just interjoined, e.g. the book (that) I read as  In both cases, however, the nodes which intersect referentially are directly joined to each other.</Paragraph>
      <Paragraph position="4"> The schematic expression for full-subjunction is Z+X*Y. The following are representative members of this schernq:  - ing read the book C. Other Modifiers as Relative Statements.  Returning now to the discussion of relative modifiers, we continue our description of interjunction, which, as the reader will recall, entails intersecting trees corresponding to the following schema : Basic relative modifiers are defined to be relative clauses of which the modifying node is of noun category, the relative marker for such clauses being in some cases null (e.g. the boy I saw was crying), or such words as which, who -9 and that. Non-basic relative modifiers are defined to be all others which entail interjunction. Some of these relatives have not in some cases been recognized in the literature as relative constructions at all, but were identified during the elaboration of junction theory by deducing from the interjunction schema specific possibilities not hitherto noted by observation of random data. This deduction was done in terms of the constituent categories extant in the junction grammar system of diagramming at that time. Sentences (3-8) below illustrate some of the non-basic relative modifiers in question. Notice that one might search available data a long time before finding an instance of some of the types exemplified, as each would require a special context. Yet, any grammar of English would be incomplete without them. The fact that some rules are used less in no way invalidates them; in a number of cases, deductions which at first seemed most ridiculous were found to have instances that were perfectly acceptable.</Paragraph>
      <Paragraph position="5"> The perspective deriving from this experience suggests that any grammar which is based on rules arrived at inductively, i.e. simply set down in a list as available data suggests them, will be incomplete at best, and most probably not motivated by any significant generalizations. Diagrams for sentences (5) - (8) will be supplied in the discussion of non-verb cored relatives.</Paragraph>
      <Paragraph position="6">  (3) Verb phrase. I hate war, like you.</Paragraph>
      <Paragraph position="7"> V N hate war (4) Verb. I peeled the peach, like you the apple.</Paragraph>
      <Paragraph position="8"> ach peeled like A the apple (5) Prepositional Phrase. My kite is on the roof, like your ball.</Paragraph>
      <Paragraph position="9"> (63 Preposition. Fred is above a store, like you were a bakery.</Paragraph>
      <Paragraph position="10"> (7) Adjective. He is like [such as) Bill is.</Paragraph>
      <Paragraph position="11"> (8) Noun. A soldier, as I am, seeks adventure.</Paragraph>
      <Paragraph position="12">  Junction Grammar claims that some statements occur without a verb in the relative modifier. One way to represent such sentences in a natural way is to allow non-verb cored statements as follows:</Paragraph>
      <Paragraph position="14"> at school old fireman We will indicate that a predicator (e.g. adjective or adverb) becomes a predicate without a direct object either by the notation , where E - means empty slot.</Paragraph>
      <Paragraph position="15"> Preposition cored: My kite is on the roof, like your ball. Fred is above a store, like you a bakery.</Paragraph>
      <Paragraph position="16">  (9) A delegate from Canada proposed a compromise. (10) The old man answered the question.</Paragraph>
      <Paragraph position="17">  The modification occurring in these sentences can be expanded out to relative clauses and represented as follows:  One benefit of considering modifiers such as these to entail entire statements is that various types of ambiguity occurring with them can be explained as types of ambiguity also characteristic of clausal statements. For instance, consider the following ambiguaus sentence : (11) John clumsily stepped on the snail.</Paragraph>
      <Paragraph position="19"> stepped on the snail The ambiguity in question hinges upon whether or not we consider the relative modification to be restrictive or non-restrictive, a contrast which in junction grammar is expressed by subcategorizing the sub junction operation (*) .</Paragraph>
      <Paragraph position="20"> Note also that semantic complexities which occur with the non-clausal modifiers of sentences (9) and (10) can also occur with basic relative clauses. For example, if John is tall (as a person) and John is a basketball player, we cannot conclude that John is a tall basketball player. And similarly, if John is an expert and John recommends Gillette, we cannot necessarily conclude that John is an expert who recommends Gilletfe (the reason being that he might be a blimp expert who just recommends Gillette products to his friends, while saying that he is an expert who recommends Gillette implies that he is an expert on the types of products that Gillette produces) .</Paragraph>
      <Paragraph position="21"> Further, if modifiers such as those in (9) and (10) are considered to be relative statements, it explains our intuition that obviously, surprisingly, etc., are usually modifiers at statement level. For example, a natural way, it would seem, to represent charitable man is as follows:  But this represents the reading in which obviously is taken as a manner adverb, and it seems that because we have not represented the modification as a relative clause, we are prevented from using obviously as a sentence level modifier.</Paragraph>
      <Paragraph position="22"> Now consider the following sentence with its Junction Grammar  By expanding the rich relative clause to include a verb, the sentence might be lexicalized as: The boy who obviously was rich, bought the Mercedes.</Paragraph>
      <Paragraph position="23"> But also, obviously can be used as an adverb of manner. Taken in this way (12) would mean that the ostentatiously rich boy bought the Mercedes. As such, it would be diagrammed as follows:  Finally, consider these two relative modifier sentences: John runs as slowly as Bill walks.</Paragraph>
      <Paragraph position="25"> as as Bill brought as many books as John bought paintings.</Paragraph>
      <Paragraph position="27"> as as paintings It should be pointed out that the above discussion has omitted major areas of Junction Grammar, such as specializations of sub junction which indicate flow of information in &amp;quot;compilingu a junction tree, conjunction, and the function and application of lexical rules to junction trees.</Paragraph>
      <Paragraph position="28"> 111. Implementation.</Paragraph>
      <Paragraph position="29"> A. Analysis.</Paragraph>
      <Paragraph position="30"> The analysis step must accept natural language input and produce the appropriate junction tree for each segment of text (normally, but not necessarily, a segment is a sentence) , according to the context of that segment.</Paragraph>
      <Paragraph position="31"> Two types of ambiguity can be distinguished: word sense ambiguity and syntactic ambiguity. Without some logical processing, an utterance such as &amp;quot;John bought some ink for the pen&amp;quot; is word sense ambiguous as to pen, between the writing utensil and the enclosure senses. Also, the sentence &amp;quot;The boy flipped the coin by the book,&amp;quot; is syntactically ambiguous as to the point of intersection between the prepositional phrase and the rest of the sentence.</Paragraph>
      <Paragraph position="32"> Several English analysis systems (e. g. W. Woods, Y. Wilks , T . Winograd, etc .) have seen considerable success in resolving word sense and syntactic ambiguities automatically on texts that stay within some restricted vocabulary or context. However, no one has attempted to apply such principles to a large scale system (10,000 to 20,000-word lexicons) which can analyze a wide variety of structures and types of texts. At present, it is not known whether any system is expandable to such a degree or, even if it were, how many years would be needed to complete such a project and how much it would cost to run. Clearly, research and development in the area of automatic analysis should be continued. In the meantime, however, work is needed op the processing that must follow automatic analysis to form a zomplete translation sys tem.</Paragraph>
      <Paragraph position="33"> Therefore, our research group has decided to develop a machine-assisted translation system consisting of interactive analysis (where the human operator resolves difficult ambiguities via his video screen) followed by automatic transfer-synthesis into multi ple target languages with junction trees as the interlingua. This configuration, which has also been proposed by Kay [7] , has some attractive features. Supposa that the Junction Grammar trapsfersynthesis system produces acceptable translations of real-world text from the output of the interactive analyzer. The analyzer could then be replaced by a more automatic version without disturbing the rest of the system.</Paragraph>
      <Paragraph position="34"> On the other hand, if the whole system fails, we will not have invested nearly as much emfort as if we had tried to develop a large-scale automatic .analysis to junction trees.</Paragraph>
      <Paragraph position="35"> In addition, during the period of development and testlng, counters can easily be set up to keep statistics on what interactions are most common and therefore should be given priority in a more automatic version or during the testing phase of autamatic routings to see if they disambiguate the same way the human does.</Paragraph>
      <Paragraph position="36"> Finally, regardless of the particular version of analysis used, the system divides the effort expended in doing the analysis by the number of target languages being translated into from the same junction tree.</Paragraph>
      <Paragraph position="37"> Now let us describe the current five -phase implementation of interactive analysis. The first phase performs limited morpheme splitting (e.g. possessives and contractions). As each word is identified it is matched against an index-sequential dictionary stored on disk. If there is more than one word sense associated with the word then several definitions are chained to the entry. These several meanings appear on the video screen and the human is asked to choose the meaning for this particular context. When a choice is entered, the information associated with that particular use of the word (or phrase of which it is the keyword) is retrieved from the dictionary. The dictionary contains semantic indices, categories, and binary features. Information specific to the source language and information considered to be language independent are represented in these features. This information is put into a two dimensional array (Reference table) and will be available during the other phases. If a word is not in the dictionary, the person monitoring has the option of picking a synonym, or passing the word through untranslated as he would a proper noun, signaling only the category of the word (noun, verb, adjective, adve,rb, preposition). It is by this interactive process that we temporarily solve the word sense ambiguity problem.</Paragraph>
      <Paragraph position="38"> The second phase of analysis, called ffPHRASER&amp;quot;, logically segments the reference table (sentence) into &amp;quot;phrasesff containing a maximum of one noun each. The first word of the sentence starts the first phrase. Thereafter, a noun or comma ends a phrase, and the succeeding phrase begins with the next word. The senfence &amp;quot;The boy in the car ate a very good hamburger while driving down the street&amp;quot; is symbolically segmented as (The boy) (in the car) (ate a very good hamburger) (while driving down the street) .</Paragraph>
      <Paragraph position="39"> During the segmenting operation PHRASER assigns a global reference category to each element (word) of the reference table. The main purpose of PHRASER is to divide the sentence into &amp;quot;phrasesl1 or segments which are individually processed by the infix generattion routines as separate units, thus simplifying their task. The third phase of analysis is the syntactic resolver and infix builder (Figure 1). This phase produces an infix representation equivalent to normal junction trees. For example, N1$(P+N2) is equivalent to The basic philo-</Paragraph>
      <Paragraph position="41"> sophy of this routine is to process phrases generated in PHRASER by calculating their internal structure, then placing them in their proper structural position in the total sentence. This is facilitated by the introduction of what we call &amp;quot;order-rules. 11 The Order-Rule As discussed earlier, a simple sentence can be represented in a tree diagram as SV  be called a clause order-rule. Notice that the order rule has the terminal nodes (N + (V + N)) as well as sub j verb ob j the father node markings (parentheses).</Paragraph>
      <Paragraph position="43"> post post sentence predicate predicate sentence This gives seven slots of layers for the actual entries of the order-rule. An order-rule has the order portion, i .em (N+(V+N)) , and the entries that fill the slots of the order: subject  There are three basic constraints on order-rules . First, they can never represent discontinuous word order. The discontinuous orders (V, N subject, N object and N object, N subject, verb) cannot be represented directly but must use variant order rules. The second basic constraint in the use of order rules is that when order rules are interjoined, the first terminal node element of the subordinant order-rule must be the intersect node. For example: &amp;quot;I saw the boy that likes youv has two order rules. must be first terminal node of the subordinate order-rule</Paragraph>
      <Paragraph position="45"> When these are combined, the following interjunction will be created: (N + (V + (N $ (V + N)))) The third constraint is that all subordinate order rules must indicate their antecedent nodes in some other order-rule.</Paragraph>
      <Paragraph position="46"> To review, phase three routines build order-rule representations; phase four inserts the slot or layers up into the orderrule, then it conflates all the order-rules into one linear infix string. This linear infix string cross references the information table produced in phase one (cross-referencing is achieved via doubly- linked- list processing) .</Paragraph>
      <Paragraph position="47"> The reader will recall f~om Section I that Junction Grammar relative clause modification involves an interj unction and phrasal modifiers (prepositional phrases, adjectives, and adverbs) have the same basic form as the clause.</Paragraph>
      <Paragraph position="48">  Phase three programs can generally identify the subject, predicate, and, if present, indirect objects of clauses, both independent and subordinate. There are, however, certain areas of processing that present difficulties for the program. A simple sentence like, &amp;quot;I threw the ball in the roomw illustrates such problems because the program does not know without context whether in the room identifies which ball it was that I threw, the direction of its path (motion), or where I and the ball were during the action. As stated before, the key to this problem is the context. As a matter of fact, almost every prepositional phrase is ambiguous in the computer environment. For this reason, we have the program query the person monitoring the program and ask,</Paragraph>
      <Paragraph position="50"> I threw ball in room and it would identify which ball was thrown. If PV 2 is entered, it is interpreted to modify the predicate (throw + ball) ; then the infix would be (N + (V + N) $ (P + N)) and implies I threw ball in room that the phrase tells where the action took place. If &amp;quot;ZU(throw) is entered, the infix would be (N +(V $ (P + N) + N)) and the phrase modifies the verb. Note that in addition to identifying the ambiguity, we can signal the intended meaning and represent any of those meanings in a systematic manner.</Paragraph>
      <Paragraph position="51"> Another &amp;quot;conte~tual~~ problem in English is the identification of the antecedents for relative pronouns. &amp;quot;I saw the boy in the car that the girl loves .If The question, of course, is: &amp;quot;What does the girl love--the boy or the car?&amp;quot; This is solved by having the program ask &amp;quot;Does 'that1 refer to X?If where X is popped off of a push-down stack of previously encountered nouns. The relatives that, who, whom, and which all invoke new order-rules. The result - of this interaction is to point the subordinate order-rule to the antecedent identified by the human s response.</Paragraph>
      <Paragraph position="53"> Combining t-he prepositional phrase and antecedent problems, we produce a sentence like this:  The program will ask what by the church modifies--we will answer &amp;quot;6&amp;quot; (car). Then it will ask &amp;quot;Does which refer to church?&amp;quot; This is really asking &amp;quot;Did the birds eat the bread on the church?&amp;quot; If we say &amp;quot;No&amp;quot; then it will ask, &amp;quot;Does which refer to car?&amp;quot; &amp;quot;Yes.&amp;quot; Next the program asks &amp;quot;What does 'on-which1 modify?&amp;quot; This is somewhat difficult because we have a prepositional phrase whose antecedent is in one clause yet the phrase modifies something in another. In the example the prepositional phrase modifies the predicate 'birds ate the bread' by telling where the action hap-</Paragraph>
      <Paragraph position="55"> Passive constructions may have an ambiguous element in them: *'The cake was eaten by the dog.'1 The question is: did the dog do the eating or does the prepositional phrase tell where the cake was eaten. Interaction indicates which reading should be represented. null Complement constructions can be identified in this same way (through interactio-n) . &amp;quot;It surprised Aunt Jane that the girl loves the car.&amp;quot; In this instance the program asks &amp;quot;Does 'that' refer During the layer insert phase the pre PV will be inserted to the left of the predicate parenthesis, making the topic the first node encountered.</Paragraph>
      <Paragraph position="56"> to Aunt Jane?&amp;quot; Reply &amp;quot;No.&amp;quot; &amp;quot;Does 'that1 refer to it?&amp;quot; This time the reply is &amp;quot;YesCu where the &amp;quot;C&amp;quot; indicates that the subordinate clause is the complement of - it.</Paragraph>
      <Paragraph position="57"> This structure is not interjoined but fully sub j oined .</Paragraph>
      <Paragraph position="58">  The infix notation for the above is : (N * (N + (V + N)) +(V + N)) . It girl loves car surp. A.Jane The order rule representat ion :</Paragraph>
      <Paragraph position="60"> A variation of the complement response is: &amp;quot;He came after we ate dinner .'I The program asks &amp;quot;What does after-we ' modify?&amp;quot; This is erroneous, and is so indicated by answering instead what the true prepositional phrase &amp;quot;after-it (that) ff modifies and putting a &amp;quot;Ctt for complement at the end of the reply: &amp;quot;PV 2C.&amp;quot; The &amp;quot;C&amp;quot; will remove the &amp;quot;we&amp;quot; as object of the preposition and replace it with a noun tt it tl A new order-rule is then invoked, pointing to the noun llit.w The underlying structure is represented &amp;quot;after it that we</Paragraph>
      <Paragraph position="62"> or, in infix: (N + (V + N) $ (P + N * (N + (v+ N)))).</Paragraph>
      <Paragraph position="63"> The details of the other interactive situations will be omitted. Interaction is necessary for the adjectival like as in &amp;quot;1 saw a man with a dog like you.&amp;quot; The question is, &amp;quot;What is like you, the man or the dog?&amp;quot; So the program asks, RDoes 'like' refer to dog?&amp;quot; If the response is &amp;quot;no,&amp;quot; then the program asks, &amp;quot;Does 'liket refer to man?&amp;quot; Participles and gerunds are very ambiguous. For example, &amp;quot;I like chewing gum.&amp;quot; Is it &amp;quot;1 like to chew gumv or &amp;quot;1 like chewinggum&amp;quot;? Again, &amp;quot;John's hobby is shocking me .&amp;quot; Is the &amp;quot;shocking&amp;quot; a progressing verb or is it a gerund? At this point we have the program ask, &amp;quot;What type of participle is X ing&amp;quot;; the replies can be: N Adj for the chewing-gum type, use of an ing form V for the progressive verb Adj X (where X* is&amp;quot;the segment modified) for adjectival participles Adv X (where X is the segment modified) for adverbial participles Of course, the infinitive form also is often ambiguous. For example, &amp;quot;I like to buy candym is much different from &amp;quot;He sent me to buy candy.&amp;quot; There are several answers to the program's question, &amp;quot;What type of infinitive is 'to X '?I' where X is the verb.</Paragraph>
      <Paragraph position="64"> Other questions that the programs are designed to ask all have two points in common: first, they are stimulated when the surface string does not map into the underlying semantic string on a one-to-one basis; second, when the decision is made by the human as to its use and the reply given to the program, the processing continues in standard existing programs and no ad hoc routines are needed.</Paragraph>
      <Paragraph position="65"> The output of phase three is a series of order-rules. Each order-rule has the form:</Paragraph>
      <Paragraph position="67"> Phase four will make one continuous infix string of these order-rules. First, the rules are checked for those situations where both of the father nodes of an interjunction are available for further junctions. If such a situation exists, then the operations involved are marked. Next, all the layers are inserted into the order-rule, giving the following: Notice that the second rule still points to the antecedent in the first order rule. Next all order-rules are conflated into one, by performing the necessary interjunctions between order-rules .</Paragraph>
      <Paragraph position="68"> The above infix string is now ready for phase five.</Paragraph>
      <Paragraph position="69"> Phase five converts the infix notation into postfix notation to be used as input to the transfer-synthesis steps.</Paragraph>
      <Paragraph position="70"> As a summary of this section, we will follow a sample sentence through the five phases of analysis.</Paragraph>
      <Paragraph position="71"> Figure 2 shows the processing involved to disambiguate the reference for some of the words in the input string* In this sentence there were several potentially ambiguous sequences (See Figure 3) &amp;quot;TPUTV1 identifies the computer programs question of the operator and &amp;quot;TGET&amp;quot; shows the ~eply. Most of these are self-explanatory. This interplay between man and program makes it possible for the computer to build a structure depicting its syntax and semantics for the given context, provided of course that the human responds correctly to the questions. The output of phase three is several order rules which are merged into one infix structure (Figure 4).</Paragraph>
      <Paragraph position="73"/>
    </Section>
  </Section>
  <Section position="5" start_page="1" end_page="19" type="metho">
    <SectionTitle>
ENTER MEANING FOR: COACH (3)
</SectionTitle>
    <Paragraph position="0"/>
    <Paragraph position="2"> Withih the context of Junction Grammar, a transfer grammar from a source language S to a target language T is an algorithm which inputs a junction tree from an analysis grammar of S and outputs the junction tree, with any needed adjustments, to a synthesis grammar of T.</Paragraph>
    <Paragraph position="3"> In the present implementation, we are developing transfer grammars from English into Spanish, German, French, and Portuguese (abbreviated SSP, GER, FRN, and POR) . Figure I is a general diagram of the flow of information in the transfer system.  In the preparation phase, each junction tree is scanned for junction rules and semantic indices which stimulate transfers going into a given language. This scanning is accomplished by consulting files which associate key junction rules and semantic indices with transfer numbers. The list of transfers is then associated with the tree.</Paragraph>
    <Paragraph position="4"> After the preparation phase, the transfer language interpre ter phase (Phase two) executes those transfer programs whose numbers have been recorded. Of course, as the transfer executes, it may check the context and decide not to change the junction tree. All the transfer programs are in one file so that several languages can share common algorithms. The third phase of transfer converts the adjusted junction trees to synthesis representation, adding to them features from the target language lexicons. None of the three phases of transfer contain any language-specific coding; only the files are language-specific. Thus, a transfer grammar can be made machine executable simply by loading records into the key files and the transfer program files.</Paragraph>
    <Paragraph position="5"> Transfer programs are written in a linguistic programming language called Transfer Language (TL) (see [8]). The TL loader compiles the programs into a pseudo-machine language which is stored in the transfer program file and is interpreted at execution time in the second phase of transfer. Before describing TL in detail, we present an entry from a simple transfer grammar. Figure 2 shows the necessary input to the TL loader and the semantic key file loader. This figure shows the form of entries loaded under the current transfer system implemented on an IBM 360.</Paragraph>
    <Paragraph position="6">  A simple transfer program and its key.</Paragraph>
    <Paragraph position="7"> The details of TL needed to fully understand Figure 2 will be explained later. For now, let us consider the simple linguistic principle behind the example. When translating &amp;quot;1 miss my brotherff into French, brother becomes the syntactic subject, giving &amp;quot;man frere me manque,&amp;quot; which, taken literally back into English, is &amp;quot;my brother (to) me misses.&amp;quot; In terms of verb orientation, the English verb to miss is normally used as a &amp;quot;subject is interested person&amp;quot; verb while in French the sense wish someone were here is a &amp;quot;subject is interesting person&amp;quot; verb. Note, however, that the other sense of to not hit is a &amp;quot;subject is interested&amp;quot; verb in French, the same as English, i.e. &amp;quot;1 threw the ball but I missed him&amp;quot; is translated &amp;quot;J1ai lance la balle mais je 1' ai manque, f I and receives a different semantic index in analysis. Therefore, it does not stimulate a transfer as does t-he first word sense.</Paragraph>
    <Paragraph position="8"> In Figure 2, the desired adjustment from &amp;quot;X misses Y1' to IrY misses Xfl is accomplished by a transfer program which has been numbered 19. In transfer 19, the first two statements (LET ...; LET.. .) set pointers to crucial nodes of the junction tree and the next three statements (REPLACE ...; REPLACE ...; JOIN ...) make the direct object into the subject and the subject into the indirect object. The transfer will be executed at the appropriate time because at the bottom of Figure 2 we associate the semantic index 5916 (to miss someone, i.e. wish he were there) with transfer 19 going into French (FRN@5916.@@g 19). The qualifier attached to the semantic index is a sememic refinement code used to make very fine distinctions if needed. Work on this aspect of our lexicon is in its formative stages and will be reported later.</Paragraph>
    <Paragraph position="9"> With the preliminary example of Figure 2 in mind, let us now examine transfer language as a programming language.</Paragraph>
  </Section>
  <Section position="6" start_page="19" end_page="19" type="metho">
    <SectionTitle>
GENERAL STRUCTURE OF TL
</SectionTitle>
    <Paragraph position="0"> TL is a free-format, block-structure, list-processing language similar in general form to PL/I or ALGOL. Internal to a transfer program are DO-END blocks and external are subroutine blocks. Computer scientists may be interested to know that TL contains no GOTO-type statement. The central data structure is a junction tree, represented as one doubly-linked list whose internal structure is logically a set of intersecting binary trees in postfix notation. Of course, TL shields the user from the details of linked list manipulation through the use of high-level commands.</Paragraph>
    <Paragraph position="1"> The user data types are: (1) address variables (e.g. '=4') which point to nodes of the junction tree, (2) condition variables (e.g. T4') which can be true-false or contain integer values (such as semantic indices) , (3) parameters (e.g. 'P2') which contain optional information associated with a sememe key in addition to the number of the transfer program to invoke, and (4) integer constants.</Paragraph>
    <Paragraph position="2"> As previously mentioned, a transfer grammar is a set of keys and transfer programs. The transffer programs are not ordered relative to each other; .their execution is stimulated by the presence of a key sememe or junction rule in a junction tree. When a given top-level transfer (i.e. one that is invoked directly by sememe or rule, not invoked by another transfer) is executed, add-ress variable one (1 is pointed at the sememe or junction rule which stimulated the transfer program. Normally, a top-level transfer progrdm will begin with some statements which derive (from the address in =1) the addresses of some other crucial nodes in the junction tree. Then, except in the simplest cases, this context will be tested and appropriate manipulations, if any, will be performed on the junction tree. If errors are detected during the execution of one of the transfers on a list, control is normally returned to the transfer supervisor, which continues processing the sentence by invoking the next transfer on the list. However, the user may choose to set up &amp;quot;ON UNITS&amp;quot; which intercept error conditions and invoke some specified transfer designed for error recovery.</Paragraph>
  </Section>
  <Section position="7" start_page="19" end_page="19" type="metho">
    <SectionTitle>
TRANSFER LANGUAGE STATEMENT TYPES
</SectionTitle>
    <Paragraph position="0"> Let us now consider the most important statement types in TL, beginning with assignment, which is accomplished with the LET statement.</Paragraph>
    <Paragraph position="1"> LET. - One form of LET is used to move around the junction tree. For example, if =3 is set to the address of the verb of a clause, we can point =5 to the direct object, skipping over any verb-level moiiifiers, by the statement: LET =5 BE Y(A(=3)). The built-in fanction A moves to the next level of adjunction (e.g. from a V to its PV) and Y moves down the tree to the right brother Other built-in functions include X [which moves down to the left brother) , L (which moves up to the father) , B (which moves down successive sons, or left brothers, until reaching a terminal node), S (which moves from the father of a point of intersection in a ranking tree to the corresponding father node in the subordinate tree) , R (which moves from subordinate trees to ranking trees), and C (which moves to the SV governing the start node). In the example LET =5 BE Y(A(=3)) , the movement is relative to an initial node pointed to by -3. In this initial address slog we can also use the symbol H, which stands for the topmost node of the highest ranking subtree of the junction tree, or '=' , which stands for the address of the most recently created node in the tree.</Paragraph>
    <Paragraph position="2"> Another form of LET assigns values to condition variables.</Paragraph>
    <Paragraph position="3"> When an address variable is assigned to a condition variable (e.g.</Paragraph>
    <Paragraph position="4"> LET C7 BE =S) , the condition variable receives the semantic index of the node pointed to by the address variable. Also available in TL is an easily expandable set of built-in functions to perform such tasks as indicating the category of a node, checking for the presence or absence of certain semantic features on a node, or reporting on the structural context of a node. For example, LET C7 BE INODCAT(=8) sets C7 to a number indicating the category of the node pointed to by =8. Other special forms of the LET statement set features and the special pointer , which indicates the top of the junction tree.</Paragraph>
    <Paragraph position="5"> JOIN. The JOIN statement allows the user to insert any desired structure into the junction tree. The dollar sign ($) is used to indicate an interjunction; otherwise, the symbols used in the JOIN statement are a straightforward infix representation of a segment of a tree. Thus, if we assume the fragmented tree: and execute the statement JOIN =4 $ (Plfl267&lt;12&gt; + =5) 9 we obtain the tree: whose P node contains the semantic index 10267 and the semantic featu~e number 12. Also, the /=9 indicates that the primary result of the last performed operation should be pointed to by =9 UNJOIN. If we then execute UNJOIN =9 on the junction tree just obtained, we u~do the effect of the interjunction, obtaining the trees : REPLACE. If, using the same tree, we execute the statements LET =7 BE X(=4), LET =8 BE Y(=4), and REPLACE =4 WITH =2, we set some pointers and switch subject and the object, obtaining: Finally, if we replace the left conjoined part of the subject with nothing, i.e. by the statement REPLACE =7 WITH g, then we obtain: Note that the unneeded operation and label node are deleted automatically. null IF. The IF statement with optional ELSE clause is simply  borrowed from PL/I, but one form of the IF statement deserves comment. An expandable set of KEYWORDS is available, each of which indicates that a built-in function is to be called and its value compared against a predefined constant associated with the KEYWORD. For example, IF =14 ISA VERB--- is equivalent (using C49 as a work variable] to the two statements : LET C49 BE INODCAT(=3) IF C49 EQ 5 THEN--since 5 is the number which stands for the category V.</Paragraph>
    <Paragraph position="6"> SKIP and HALT. SKIP exits from the current transfer program, while HALT returns control to the supervisor, which moves on to the next top-level transfer.</Paragraph>
    <Paragraph position="7"> TRANSFER. The statement TRANSFER 11 is simply a call to t-ransfer number 11 as a subroutine, while the statement TRANSFER C4 is a call to that transfer program whose number is in C4 at the time that statement is interpreted. There sqems to be no need for an algebraic do-loop in transfer but a while-loop is useful. For example, TRANSFER 12 WHILE C2 executes transfer number 12 repeatedly until some context Lest within transfer 12, or some other transfer called by it, sets condition variable 2 to false (Lea to zero). ON. The statement ON CONDITION(N0-LABEL) TRANSFER 1 sets  up an ON UNIT so that if at any time during transfer on the current node an attempt is made to move up from a node without a label, transfer 1 will be executed. If an error occurs in transfer 1, control is then passed to the supervisor which processes the next top-level transfer.</Paragraph>
    <Paragraph position="8"> A few final details should be mentioned. When a transfer is called, it executes to its end or until a SKIP statement is executed. Comments can be scattered through a transfer program by enclosing them by t/* ----- */' or beginning the record with I*' Figure 3 summarizes the selected statement types we have discussed. null</Paragraph>
  </Section>
  <Section position="8" start_page="19" end_page="19" type="metho">
    <SectionTitle>
SAMPLE TRANSFERS
</SectionTitle>
    <Paragraph position="0"> Having discussed the major statement types in TL, let us consider a few specific transfers. This will be but an indication of what is done in transfer, but later reports are expected to present substantial transfer grammars between language pairs.</Paragraph>
    <Paragraph position="1"> Languages frequently differ in their preferred wnceptualization of a given action. For example, English prefers eat supper to sup, while French prefers souper to prendre le souper. Similarly, to watch television has become the single sememe fernsehen in German. Alternatively, a verb and non-verbal participle [9] may correspond to a single verb (e.g. make happy goes to alegrarse in Spanish). Several examples of such patterns are listed in Figure All these correspondences are handled in TL by writing one simple transfer for each general pattern and then loading the appropriate keys. For example, the first case (V + N+V + E) could be treated by keying on the verb in each case and loading as parameters with the key the crucial object or objects. This will cause collapse and the composite sememe to be formed. In order to make this and the subsequent examples easier to follow we will use the English words themselves instead of their semantic indices.</Paragraph>
    <Paragraph position="2"> Thus, we might load the following keys:  to eat lunch ---r dejeuner (FRN) to eat supper * souper [FRN) to eat breakfast desagunar (SPN) to give thanks - agradecer (SPN) 50 take a walk- spaziergehen (GER) to take a bath baden (GER) to have an accident 4 verunglticken (GER) to write a poem- dichten (GER) to point rnontrer du doigt (FRN) to clap- battre des mains (FRN) V$ (A* (A+E) +E) +E--+V+E to make sad * attrister (FRN) to make angry -+ encolorizarse (SPN) to stare at Xj regarder X fixQment (FRN) to turn the pages of a book + fculletcr un liver (FRN)  The built-in function SMATCH checks the parameters of the stimulating key against the index of the node pointed to by the first argument (52) . It either returns FALSE (no match) or TRUE and points the second argument (-3) to the proper composite node made available through the auxiliary JOIN parameters of the transfer key.</Paragraph>
    <Paragraph position="3"> In this simple transfer, any modifiers on the object would cancel the transfer by blocking a match (e.g. eat a big supper 4 manger un grand souper), but if the transfer writer so desired, he could easily UNJOIN an adjective on the object and JOIN it onto the predicate as an adverb (e. g. SOUPER GRAND) . This could be done by expanding the first statement to the following:</Paragraph>
    <Paragraph position="5"> Now let us consider a transfer stimulated by a difference in productivity of a junction pattern between two languages. English and French both have possessive adjectives (his, my+son, sa, ses, mon, ma, mes) but on non-pronouns the English adjectival form becomes prepositional in French, e. g. my friend's mother's kitchenla cuisine de la mere de mon copain (the kitchen of the mother of my friend).</Paragraph>
    <Paragraph position="6"> Fortunately, in TL we need not consider the embedded cases separately from the simple case; rather we can write one transfer which will be involved automatically by the transfer supervi si en for each occurrence of the possessive. The basic transfer should modify structure in the following way:</Paragraph>
    <Paragraph position="8"> In the case of one embedding the boy's mother's kitchen, the transfer would be executed on boy's and then on mother's, changing structure as follows :  Before leaving this second example, it should be remarked that in this transfer, as with all transfers, we do not consider questions of differing word order or inflectional patterns between languages. In transfer we simply modify the structure, and synthesis later produces the proper word order and inflection through the lexical-rule system.</Paragraph>
    <Paragraph position="9"> As a final example we will consider a simplified version of an extensive transfer written as a major component of a thesis which considered several forms of the English passive [lo]. In our simplifiad version we do not make many of the needed checks to weed out unwanted modifiers, etc.</Paragraph>
    <Paragraph position="10"> Consider the three closely related sentences:  (1) John gave him a bopk.</Paragraph>
    <Paragraph position="11"> (2) He was given a book by John.</Paragraph>
    <Paragraph position="12"> (3) A book was given to him by John.</Paragraph>
    <Paragraph position="13">  Many grammars would consider these sentences to have the same meaning and give them the same underlying representation. Junction Grammar, taking a slightly different approach, chooses to represent them in such a way that indicates both their similarities and their differences. This attitude is based on two assumptions of Junction Grammar.</Paragraph>
    <Paragraph position="14"> (1) The path between deep and surface structure should generally be direct, not tortuous, and (2) Any difference in structure implies some difference, however minute, in meaning, since meaning (in Junction Grammar) is defined as the compositidn of structure, reference, and context.</Paragraph>
    <Paragraph position="15"> However, the philosophical gulf between Junction Grammar and other grammars is not as wide as might appear at first, for if in any patticular application area it is useful to ignore such differences as those between the passive and the active, standard junction trees can easily be passed through a special normalizing transfer grammar. One transfer in such a normalizing grammar would be the following, which is stimulated by the rule A * PV and transfers sentences of type (2) and (3) into active sentence9 of type (I), examples of which are shown in Figure 5. (Going into this transfer, it is assumed that =1 is set to the label node of the rule</Paragraph>
  </Section>
  <Section position="9" start_page="19" end_page="19" type="metho">
    <SectionTitle>
*PASSIVE-ACTIVE TRANSFER
*SEE SWLE JUNCTION TREES IN FIGURE 5
</SectionTitle>
    <Paragraph position="0"/>
  </Section>
  <Section position="10" start_page="19" end_page="19" type="metho">
    <SectionTitle>
*SWITCH THE PASSIVE CONSTITUENT (IT MAY BE DATIVE OR OBJECTIVE)
WITH ITS CORE FERENT , THE SYNTACTI C SUBJECT
</SectionTitle>
    <Paragraph position="0"> The above description of Transfer Language and the sample transfers should convey some idea of transfer within Junction Grammar, but it should also point out the difficulty of comparing Junction Grammar transfer with the intermediate adjustment phase of some other translation system (e.g. [ll], [12], [13]). The reason for the difficulty is simply that a transfer phase is tightly interlaced with the analysis and synthesis phases and the theoretical base. For example, we do some adjustments in transfer which otner systems neutralize in analysis, while some other systems consider aspects of word order and word choice in transfer which we handle in analysis and synthesis.</Paragraph>
    <Paragraph position="1"> C. Synthesis.</Paragraph>
    <Paragraph position="2"> The input to this program is a junction tree or J-tree. The program applies lexical ordering rules, lexical hiatus rules, lexical matching rules, lexical agreement rules, and graphological and phonological adjustment rules. Figure 1 shows the synthesis mainline in a simplified form which omits some details pertaining to the processing of intersecting trees and discontinuities. Basic processing follows the figure. First a J-tree is read in to the computer. The perspective routine recognizes some essential relationships which are implicit in the J-tree and makes them available to the synthesis program in an explicit form. Then, beginning with the topmost node of the J-tree, the cycle processes the nodes one at a time. Label nodes are processed by the hiatus and ordering routines. The LHO algorithm determines the ordering procedure which should be employed and takes the steps necess.ary to implement it. These steps always include designating the order in which the operands of the label node should be processed. They may include the marking of hiatus (understood) elements, discontinuous elements, and insertion points.</Paragraph>
    <Paragraph position="3"> Processing continues through the J-tree in the sequence designated by the LHO algorithm until the entire J-tree has been interpreted. Terminal nodes are interpreted by the.LMA (lexical matching and agreement) routines and appropriate lexical man'ifes tations are generated.</Paragraph>
    <Paragraph position="4">  Two types of lexical ordering rules are used in the synthesis programs. There are: (1) left-right and right-left ordering of operands, and (2) discontin~ous ordering. These will be considered individually (1) When the LHO (lexical hiatus and ordering) algorithm interprets a label node, it designates one of its operands as the first operand and the other as the second operand. Processing then moves to the operand designated as the first. At the top of the chart the flow of processing is controlled by whether a node is terminal or a label. At the bottom of the chart, the order of processing is determined by whether the node which has just been processed was designated as a first or a second operand. Processing moves from the first operand to the second and from the second back to the label node.</Paragraph>
    <Paragraph position="5"> This seemingly simple ordering procedure gives the junction grammar system great flexibility and power. A single junction which has two operands has two ordering possibilities. However, if one of those operands i's itself a junction with two possible orders, then the string is capable of four orders. In a similar manner, a string of three junctions can be ordered in eight distinct ways. In general, a tree composed of N junctions can be ardcled in 2&amp;quot; distinct patterns.</Paragraph>
    <Paragraph position="6"> (2) Discontinuous ordering exists when sentences elements which are most closely related structurally are not contiguous.</Paragraph>
    <Paragraph position="7"> The sentences &amp;quot;It surprised me that he cameM and &amp;quot;It is so big that I can't lift it&amp;quot; have discontinuous order. With continuous ordering they would read &amp;quot;It that he came surprised me&amp;quot; and &amp;quot;It was so that I can't lift it big.&amp;quot; The synthesis program handles discontinuous ordering by redirecting the flow of processing in such a way that the discontinuous element is omitted at its normal position and processed instead at a predetermined insertion point. Figure 2 shows the synthesis mainline with the reset routines and their skip points.</Paragraph>
    <Paragraph position="8"> Figure 3 shows the reset routines in detail.</Paragraph>
    <Paragraph position="9"> If the flow of processing is followed as shown in the figures, the operation of the reset routines is relatively simple. On each cycle the reset routines check the node which is being interpreted to determine if it has been designated by the LHO algorithm as an insertion point or a disconf inuous element. If it has not, no action is taken. If it has, normal processing is interrupted to achieve the desired order. For instance, if the reset routine at the top of the diagram detects that a node is a discontinuous element, it causes the flow of p~ocessing to jump to the second skip point. This bypasses the LHO and LMA routines, in effect omitting the discontinuity. At the bottom of the chart processing resumes just as if the discontinuity had not been omitted. In a similar manner the reset routines direct processing from the insertion point to the discontinuity and from the discontinuity back to the insertion point.</Paragraph>
    <Paragraph position="10"> The ordering of intersecting structures is quite similar to the ordering of discontinuous elements. When the point of intersection is reached, the flow of processing shifts to the top of the subordinate structure and that entire structure is processed.</Paragraph>
    <Paragraph position="11"> Then processing shifts back to the main structure which is completed in the usual manner. Thus, the point of intersection has some similarity to an insertion point while an intersecting structure resembles the discontinuous element. This same procedure is followed when there are several intersections.</Paragraph>
    <Paragraph position="12"> The lexical matching rules access the target language lexicon.</Paragraph>
    <Paragraph position="13"> The lexical agreeement rules examine the junction and feature environment in which the particular sememe appears in orde~ to deterinine which inflected form should appear.</Paragraph>
    <Paragraph position="14"> At present, most inflected forms appear in the lexicon in full.</Paragraph>
    <Paragraph position="15"> Others, however, are generated by the synthesis program from stems and endings. The two approacE.es are theoretically equivalent since most of the inflected forms in the lexicon are generated from stems and endings by the lexicon building routine anyway.</Paragraph>
    <Paragraph position="16"> Thus, the synthesis program interprets a J-tree to generate an output string. When this output string has been adjusted by the graphological and phonological rules, which shape the final form of the output, the target language text appears.</Paragraph>
    <Paragraph position="17">  It is proposed that junction schemata of the type illustrated constitute a universal pool of structural patterns available to all languages.</Paragraph>
    <Paragraph position="18"> Presumably, most languages use the majority of them but no one language need use them all. Perhaps no two languages use exactly the same subset, nor any two persons speaking the same language, for that matter. Languages can be expected to vary in their manner of realizing junction patterns as surface strings. Specifically, the lexical stock, word order, concord, inflection and conjugation paradigms, etc., are known, for the most part, to be language specific.</Paragraph>
    <Paragraph position="19"> Our experience to date supports such expectations. For example, the basic word order for Japanese can be obtained by shifting the left-right options on subjunction used by English to right-left. While English uses heterogeneous interjunction freely (&amp;quot;John passed, which surprised us&amp;quot;, etc .) , Japanese does not ; it either uses the corresponding full-sub junction (&amp;quot;That John passed surprised us&amp;quot;) , or two conjoined sentences (&amp;quot;John passed, and this surprised us&amp;quot;). Of course, the lexical stock and morphophonemic patterns of Japanese are radically different from those of English. But the conclusion we have been forced to repeatedly is that while surface phenomena are vastly disparate from language to language, junction phenomena are more alike than different .</Paragraph>
    <Paragraph position="20"> While we use the familiar ANALYSIS/TRANSFER/SYNTHESIS scheme as a general framework for our translation system, the design of these components is rigidly governed by the junction grammar model, with junction trees serving as the interlingua.</Paragraph>
    <Paragraph position="21"> Analysis, operating in an interactive mode, produces junction trees from the source text. The data thus obtained is then passed to the transfer segment of the system.</Paragraph>
    <Paragraph position="22"> Transfer from a source language A to a target language B begins with inspection of every junction rule and semantic index appearing in the junction trees produced by analysis. If some rule or semantic index of A does not belong to the subset used by language 3, then a transfer subroutine adjusts the junction tree. The library of A-B transfer subroutines is a partial contrastive grammar of fanguages A and B. However, since word order and other lexical phenomena are abstracted away from the semantic component in our model, these cto not enter into transfer at all, but are handled independently by lexical synthesis for language B. The adjusted trees output by transfer are passed to a synthesis program. The four steps of the synthesis process correspond to four lexical rule types--ordering, matching, ellipsis (or hiatus) , and agreement.</Paragraph>
    <Paragraph position="23"> Our development group is currently engaged in developing English analysis and transfer-synthesis for translation into Spanish, French, and German, using the University's IBM 360/65.</Paragraph>
    <Paragraph position="24"> Our system's dictionaries are largely based on a vocabulary drawn from a data base ofmaterials published for the membership of the LDS (Mormon) Church, and hence they are not technically oriented.</Paragraph>
    <Paragraph position="25"> at present. Currently, each lexicon contains entries for over 10,000 semantic indices and we hope to double this size by the end of 1976.</Paragraph>
    <Paragraph position="26"> We expect that this prototype will provide a gauge for the utility of various on-line interaction techniques for compuberassisted translation systems and for the feasibility of achieving commercial grade computer-assisted translation via the junction grammar model of language described herein. Work is also in progress to develop a phonological component of Junction Grammar that accounts for the connection between prosodic features (pitch, amplitude, duration, pause, etc.) and semantic considerations of sentences.</Paragraph>
  </Section>
class="xml-element"></Paper>
Download Original XML