File Information

File: 05-lr/acl_arc_1_sum/cleansed_text/xml_by_section/metho/94/a94-1038_metho.xml

Size: 8,192 bytes

Last Modified: 2025-10-06 14:13:38

<?xml version="1.0" standalone="yes"?>
<Paper uid="A94-1038">
  <Title>Integration of example-based transfer and rule-based generation</Title>
  <Section position="4" start_page="196" end_page="196" type="metho">
    <SectionTitle>
2 Example-based transfer
</SectionTitle>
    <Paragraph position="0"> The transfer module outputs the TE structure from an input sentences by using translation examples.</Paragraph>
    <Paragraph position="1"> In the proposed translation method, translation examples are classified based on the string pattern of their SE and are stored as empirical transfer knowledge, which describes the correspondence between an SE and its TEs. The following is an example of the transfer knowledge about &amp;quot;X to Y&amp;quot; at the verb-phrase level.</Paragraph>
    <Paragraph position="2"> X to Y --~ Y' e X' ((go, Kyoto)...) Y' ni X' ((pay, account)...) Y' wo X' ((refer, announcement)...) The first possible TE is &amp;quot;Y' e X&amp;quot;, with the example set ((go, Kyoto)...). Within this pattern, X' is the TE of X, which expresses a variable corresponding to some linguistic constituent. (go, Kyoto) are sample bindings for &amp;quot;X to Y&amp;quot;, where X =&amp;quot;go&amp;quot;, and Y : &amp;quot;Kyoto&amp;quot;.</Paragraph>
    <Paragraph position="3"> Patterns in transfer knowledge are classified into different levels according to the scale of their linguistic structure in order to restrict the explosion of structural ambiguity, and an input sentence is decomposed into chunks by applying SE parts of transfer knowledge in a top-down fashion.</Paragraph>
    <Paragraph position="4"> Suppose that an input sentence is &amp;quot;Please come to our company.&amp;quot; SE parts of transfer knowledge are applied in the order, &amp;quot;please X&amp;quot; (simple sentence), &amp;quot;X to Y&amp;quot;(verb phrase), &amp;quot;our X&amp;quot; (compound noun), &amp;quot;come&amp;quot;, &amp;quot;company&amp;quot; (surface word), yielding the following SE structure: (Please ((come) to (our (company)))) For each chunk of the SE structure, the most appropriate TE is selected according to the calculated distance between the input words and the example words. The distance calculation method of (Sumita, 91) is adopted here. The distance between words is defined as the closeness of semantic attributes in a thesaurus.</Paragraph>
    <Paragraph position="5"> The SE structure chunks of &amp;quot;Please come to our company&amp;quot; are transferred to &amp;quot;X' tekudasai&amp;quot;, &amp;quot;Y' e X'&amp;quot;, &amp;quot;watashi-tachi no X'&amp;quot;, &amp;quot;kuru\[come\] 1, and &amp;quot;kaisha\[company\]&amp;quot;. By combining these TE chunks, the following TE structure is obtained, which will be the input of the composition module of the rule-based generation model: (((watashi-tachi no (kaisha)) e (kuru)) tekudasai) In the above structure, the honorific word &amp;quot;irassharu\[come\]&amp;quot; is more adequate than the neutral word &amp;quot;kuru&amp;quot; from the point of view of politeness. The replacement of &amp;quot;kuru&amp;quot; with &amp;quot;irassharu&amp;quot; will be done in the adjustment module of the rule-based generation model.</Paragraph>
  </Section>
  <Section position="5" start_page="196" end_page="196" type="metho">
    <SectionTitle>
3 Rule-based generation
</SectionTitle>
    <Paragraph position="0"/>
    <Section position="1" start_page="196" end_page="196" type="sub_section">
      <SectionTitle>
3.1 Composition
</SectionTitle>
      <Paragraph position="0"> The composition module checks whether a transferred sentence is grammatically appropriate or not, and corrects grammatical errors produced by the structural gap. The composing method is almost the same as the syntactic analysis method. However, the process is much simpler, because the input string has the correct Japanese structure and the corresponding English expressions.</Paragraph>
      <Paragraph position="1"> The procedure is as follows: 1) Divide the sentence into clauses, using not only Japanese grammatical features but also the TE structure and its corresponding English expressions. 2) Analyze each clause using the Japanese syntax rule. 3) Check on Japanese grammatical constraints. If the process finds violations, it corrects them by using Japanese linguistic knowledge.</Paragraph>
      <Paragraph position="2"> Japanese sentences have a peculiar grammatical constraint. Some expressions cannot appear in a subordinate clause. For example, the postpositional particle &amp;quot;wa(topic marker)&amp;quot; cannot appear in a conditional clause. Table 1 gives examples of limitations on expressions. In Table 1, &amp;quot;masu&amp;quot; expresses an auxiliary verb, which indicates the level of politeness, and &amp;quot;darou&amp;quot; expresses an auxiliary verb, which indicates the speaker's supposition.</Paragraph>
      <Paragraph position="3"> The checking and correcting method is explained here, using the conditional clause &amp;quot;((anata wa ryohkin wo shiharau masu) baai) \[you TOPIC fee OBJECT pay POLITE CONDITION\].&amp;quot; First, the process checks on limitations for conditional clauses by referring to Table 1, so it understands that neither &amp;quot;wa&amp;quot; nor &amp;quot;masu&amp;quot; can appear in a conditional clause(X baai). Second, the process analyzes the clause, so it understands that the case of &amp;quot;anata\[you\]&amp;quot; is &amp;quot;ga&amp;quot;, and &amp;quot;masu&amp;quot; can be deleted. Finally, the process gets the right conditional clause &amp;quot;((anata ga ryohkin wo shiharau) baai)&amp;quot;</Paragraph>
    </Section>
    <Section position="2" start_page="196" end_page="196" type="sub_section">
      <SectionTitle>
3.2 Adjustment
</SectionTitle>
      <Paragraph position="0"> A sentence that is only grammatically appropriate, is not as natural as a colloquial sentence. The adjustment module refines the sentence by changing, 1 \[wl ... w,\] is the list of corresponding English words. Uppercase shows the meaning of a function word.</Paragraph>
      <Paragraph position="1"> Table h limitations on Japanese clauses example clause topic polite supposition &amp;quot;wa&amp;quot; &amp;quot;masu&amp;quot; &amp;quot;darou&amp;quot;  agent recipient example for okuru\[send\] -- hearer o-okuri-suru hearer speaker okut-tekudasaru hearer -- o-okurini-naru adding, or deleting words. This module handles honorific expressions and redundant personal pronouns, which are important for generating natural Japanese sentences.</Paragraph>
      <Paragraph position="2"> Personal pronouns are usually redundant in Japanese conversations, because honorific expressions and modality expressions limit the agent of the action. The procedure is as follow: 1) Change the verb into the appropriate form based on the agent or the recipient, 2) Delete the redundant personal pronouns based on the verb form, or the modal, 3) Generate a final output by adjusting the morphological feature.</Paragraph>
      <Paragraph position="3"> This method is explained here, using the sentence &amp;quot;((anata wa watashi ni youshi wo okuru masu) ka) \[you TOPIC I OBJECT form OBJECT send POLITE INTERROGATIVE\].&amp;quot; First, the process changes the verb &amp;quot;okuru&amp;quot; into &amp;quot;okut-tekudasaru&amp;quot; by referring to Table 2. Second, it deletes the redundant pronouns &amp;quot;anata wa&amp;quot; and &amp;quot;watashi ni&amp;quot; . Finally, it generates the sentence &amp;quot;youshi wo okuttekudasai masu ka \[form OBJECT send-RESPECT</Paragraph>
    </Section>
  </Section>
  <Section position="6" start_page="196" end_page="196" type="metho">
    <SectionTitle>
POLITE INTERROGATIVE\].&amp;quot;
4 Evaluation
</SectionTitle>
    <Paragraph position="0"> The prototype system was evaluated by using model conversations between an applicant and a secretary about conference registration. The model conversations consist of 607 sentences, and cover basic expressions. The system provided an average translation time of 2 seconds for sentences with an average length of 10 words, and produced a translation result for all of the sentences. 480 of the results were as natural as colloquial sentences and giving a success rate of 79%.</Paragraph>
  </Section>
class="xml-element"></Paper>
Download Original XML