File Information

File: 05-lr/acl_arc_1_sum/cleansed_text/xml_by_section/metho/04/p04-3001_metho.xml

Size: 6,893 bytes

Last Modified: 2025-10-06 14:09:01

<?xml version="1.0" standalone="yes"?>
<Paper uid="P04-3001">
  <Title>TransType2 - An Innovative Computer-Assisted Translation System</Title>
  <Section position="3" start_page="1" end_page="1" type="metho">
    <SectionTitle>
3 System Architecture
</SectionTitle>
    <Paragraph position="0"> The TT2 system consists of two major subsystems that interact closely: user interface (UI), written in Java, provides the typing and pointing modalities; a second UI supplements those with speech for operating the prototype via short commands uttered by the user . The user interface also produces a trace of all useractions that can later be replayed by a special program or analyzed in order to evaluate the effectiveness of TransType2 both in terms of number of keystrokes needed for typing a translation and the various patterns of use.</Paragraph>
    <Paragraph position="1"> prediction engine (PE), written in C/C++, of which there are multiple realizations available, several per language pair and specific domain (either technical documentation, EC official documents or Hansards). The translation engines developed by research partners are: RALI (French-English) is a maximum-entropy minimum-divergence translation model (Foster 2000) that proposes multiple completions for the next few words.</Paragraph>
    <Paragraph position="3"> based on finite-state techniques (Cubel et al. 2003) and suggest a single completion of a whole sentence.</Paragraph>
  </Section>
  <Section position="4" start_page="1" end_page="2" type="metho">
    <SectionTitle>
RWTH (French-English, Spanish-English,
</SectionTitle>
    <Paragraph position="0"> German-English) are statistical based (Och et al.</Paragraph>
    <Paragraph position="1"> 2003) and suggest a single completion of a whole sentence.</Paragraph>
    <Paragraph position="2"> The main communications between the UI and  the PE are the following: 1. To initialize the PE, the UI calls a generic create method API function with the appropriate parameters required by each PE and checks its successful completion.</Paragraph>
    <Paragraph position="3"> 2. Once the user has selected the file he/she wants to work with, the UI produces a list of text segments (sentences) and displays them in the source text pane of the interface.</Paragraph>
    <Paragraph position="4"> 3. The selection of a source sentence is communicated to the PE by the UI. The sentence becomes the source text context prediction for the PE until the user selects another sentence. 4. The UI communicates to the PE every single modification of the target text: insertion/removal of a new character (letter, digit, punctuation sign or white space) and cursor movements within the target text. The UI communicates left-right onecharacter-at-a-time movements in the target text area. However, the PE does not take into account the text to the right of the cursor for making its predictions.</Paragraph>
    <Paragraph position="5"> 5. In response to the request, the PE initiates the search for completions that are eventually returned to the UI for their display.</Paragraph>
    <Paragraph position="6"> 6. As part of the general exit procedure, the UI calls  a generic destroy method API function with the appropriate parameters required by each PE and checks its successful completion.</Paragraph>
    <Paragraph position="7"> All communication exchanges between the UI and the PE are initiated by the UI, while the PE is in charge of responding by doing some actual work. This is particularly the case in 5 (producing a list of completions), while the others are more of an informative nature (cases 3 and 4) or can hardly considered communication exchanges at all: cases 1, 2 (loading a text file and producing a list of sentences) and 6 (termination).</Paragraph>
    <Paragraph position="8"> Prediction engines and the speech recognizers are developed and tested under an operating platform (Linux) different than the one chosen for user testing (MS Windows). This duality implies that prediction engines and speech recognizers, while developed under Linux, should be able to run under Windows. The users (i.e. the two translation bureaus) voiced early in the project that TT2 system should run at least under Windows, although preferably it should also run under Linux. TT2 runs currently on both platforms, the dissemination and awareness of the TT2 prototype are broader, and go further than the initial objectives proposed inside the IST project.</Paragraph>
    <Paragraph position="9"> Given that developers of the prediction engines and speech recognizers were in favor of using C/C++ as their principal programming language, two practical alternatives were discussed: * Write code without operating platform dependencies and according to standards, that would allow compilers for both platforms to build functionally equivalent binary versions.</Paragraph>
    <Paragraph position="10"> * Employ tools that lessen to a certain extent the requirement of written C/C++ platform independent code, while allowing the porting of code from the Linux to the Windows platform.</Paragraph>
    <Paragraph position="11"> This was the preferred option and the three PE's actually make use of one of such tool: Cygwin  .</Paragraph>
    <Paragraph position="12"> Cygwin provides a C/C++ compiler for the Windows platform and a library (cygwin1.dll) that gives support to Linux/Unix operating system services under the Windows environment.</Paragraph>
    <Paragraph position="13"> The partners responsible for developing the user interface have opted for JAVA as the programming language because of its graphical user capabilities, in particular its text components, which are fully configurable and compatible with external C/C++ programs. This option solves the portability problem, since the resulting code will run under any JAVA-enabled operating system.</Paragraph>
    <Paragraph position="14">  Generally speaking, running the TT2 system demands a high-end personal computer or workstation in order to be able to provide translation completions in real-time and also to be able to incorporate multi-modal user input.</Paragraph>
    <Paragraph position="15"> The minimum user equipment is a high-end personal computer running under Windows with a minimum of 1GB of RAM; however, 2 GB of RAM and Windows XP Professional operating system is preferable. If a Linux operating system is used, the kernel version must be 2.4.20 or higher. It is also required to have installed the Java 2 Runtime Environment, preferably version 1.3.1_09. To produce the PE, cygwin1.dll version 1.5.5-1 is required.</Paragraph>
    <Paragraph position="16"> The interface requirements of both scenarios include standard keyboard and mouse equipment; video display capable of resolutions of 1024x768 pixels or higher and voice input hardware (microphone, a headset preferably, and sound card) if the optional speech recognition module is used.</Paragraph>
  </Section>
class="xml-element"></Paper>
Download Original XML