Top Banner
Faculty of Applied Modern Languages COURSE PAPER "DIFFICULTIES IN MACHINE AND AUTOMATIC TRANSLATION" Written by:
120

Course Paper on Machine Translation

Oct 26, 2014

Download

Documents

rebejatatiana

Research on machine translation for students!
Welcome message from author
This document is posted to help you gain knowledge. Please leave a comment to let me know what you think about it! Share it to your friends and learn new things together.
Transcript
Page 1: Course Paper on Machine Translation

Faculty of Applied Modern Languages

COURSE PAPER

"DIFFICULTIES IN MACHINE AND AUTOMATIC TRANSLATION"

Written by:

Scientific advisor:

Page 2: Course Paper on Machine Translation

Chişinău 2005

ContentIntroduction………………………………………………………………………3I. HISTORY OF MACHINE TRANSLATIONI.1 General introduction…………………………………………………………5I.2 Before the computer………………………………………………………...11I.3 The first beginnings (1946-1949)…………………………………………...15I.4 Weaver's memorandum (1949)…………………………………………….19I.5 From Weaver to the first MT conference (1950-1952)……………………20

I.5.1 First MT studies………………………………………………….…...21I.5.2 The decade of high expectation and disillusion, 1954-1966……...…23I.5.3 The ALPAC report and its consequences, 1966-1980………………24I.5.4 The 1980s……………………………………………………………....25I.5.5 The early and the late 1990s……………………………………….....26

II. COMMON ERRORS IN MACHINE TRANSLATIONII.1 The quality of translation……………………………………………….…29II.2 Mechanical

dictionaries……………………………………………………29II.3 Polysemy and

semantics……………………………………………………31II.4 Morphological

analysis………………………………………………….…35II.5 Syntactic

analysis…………………………………………………………...36II.6 Formal syntax and transformational grammar…………….

……………39II.7 Syntactic ambiguity and discourse relations…………………...

…………41II.8 Sentences and texts………………………………………………...

……….44 II.9 Transfer and synthesis………………………………………………..

……44II.10 System designed and strategies………………………………………45II.11 Respective and influences………………………………………….…48

III. DIFFICULTIES IN MACHINE TRANSLATIONIII.1 Difficulties in

translation………………………………………………...…52III.2 Machine translation

ambiguity………………………………………….…57III.3 Problems of machine

translation…………………………………………..60III.4 Cognitive

processes………………………………………………………….62

Page 3: Course Paper on Machine Translation

General conclusion……………………………………………………………...65

Appendix 1………………………………………………………………………...i

Appendix 2………………………………………………………………………..v

Appendix 3……………………………………………………………………….ix

Bibliography

Introduction

People who need documents translated often ask themselves whether they

could use a computer to do the job. When a computer translates an entire document

automatically and then presents it to a human, the process is called machine

translation. When a human composes a translation, perhaps calling on a computer for

assistance in specific tasks such as looking up specialized words and expressions in a

dictionary, the process is called human translation.

There is a gray area between human and machine translation, in which the

computer may retrieve whole sentences of previously translated text and make minor

adjustments as needed. However, even in this gray area, each sentence was originally

the result of either human translation or machine translation. "Machine translation" is

possible only for the case when a computer performs both the initial translations of

the sentences and subsequent manipulations. All else, I will call just "translator

tools".

This paper begins with a concise history of machine and computer-assisted

translation, followed by a brief analysis of the types of error and difficulties appeared

in the process of Machine Translation. It then describes the technology available to

translators in this first decade of the twenty-first century and examines the negative

and positive aspects of machine translation and of the main tools used in computer-

assisted translation: electronic dictionaries, glossaries, terminology databases,

concordances, on-line bilingual texts and translation memories here I selected the

results concerning what the evaluators regarded as translation errors caused by the

Page 4: Course Paper on Machine Translation

input and the errors caused by malfunctioning of the Machine Translation system.

The analysis of the data is carried out by systematizing the classification and

comments of the evaluators regarding the translation errors, the translation difficulties

and quantifying the results. My aim is to determine which characteristics of the text

are attributable to the writer's intention to use language that differed from formal

norms, and which characteristics are attributable to other factors, focusing on

language contact, a significant aspect of the area under study. Another important aim

of mine is to introduce some pieces of advice, which are helpful and necessary for a

good and productive translation. This paper presents the results of a study of the

problems of automatic language translation. Indeed, the future value of research on

automatic translation might well hinge more on its contributions to a fundamental

understanding of all levels of language structure and of the nature of automatic

information processing than on any resulting machine-produced translations.

My paper investigations have been confined to translation errors and difficulties.

There has been much speculation about general methods of translation, whether

directly between pairs of languages or via some natural or artificial immediate

language. It is known that we are yet too ill equipped for a frontal assault and more

promising studies of individual language pairs. But I do believe that soon we will

have such productive and qualitative technical equipment for a translation that will

not need a pre- or post-editing.

For investigating the machine translation process I have used the descriptive

and contextual methods.

Another objective of my paper is to present to the reader and to investigate

the classification of errors made in the process of automatic translation. In my paper I

introduced some drafts, which show the process of automatic and human translation

and some schedules with an algorithm for automatic translation.

Still, translation is a very difficult thing requiring much feeling and

understanding of cultural aspects, which is not available in a computer.

Page 5: Course Paper on Machine Translation

I. HISTORY OF MACHINE TRANSLATION 1.1 General introduction.

Machine translation is the application of computers to the translation of texts

from one natural language into another. There have been many different reasons for

attempting it. The principal reason is a severely practical one: scientists,

technologists, engineers, economists, agriculturalists, administrators, industrialists,

businessmen, and many others have to read documents and have to communicate in

languages they do not know; and there are just not enough translators to cope with the

ever increasing volume of material which has to be translated. Machine translation

would ease the pressure. Secondly, many researchers have been motivated by

idealism: the promotion of international cooperation and peace, the removal of

language barriers, the transmission of technical, agricultural and medical information

to the poor and developing countries of the world. Thirdly, by contrast, some

sponsors of machine translation activity have seen its importance in military and

intelligence contexts: to help them find out what the ‘enemy’ knows. Fourthly, there

are ‘pure research’ reasons: to study the basic mechanisms of language and mind, to

exploit the power of the computer and to find its limitations. Finally, there are simple

commercial and economic motives: to sell a successful product, or to maintain a high

standard of living in a competitive world.

At certain periods in the nearly forty years of the history of machine

translation, some of these motives have been more prominent than others. In the

United States during the 1950’s and 1960’s fear of Soviet technological prowess

Page 6: Course Paper on Machine Translation

(particularly after the launch of the first sputnik in 1957) stimulated much

governmental and military support for Russian-English translation. In the 1970’s the

multilingual problems of the European Communities encouraged research on

translation systems to deal with administrative, economic and technical

documentation within the languages of the communities. At the same time,

commercial interests began to gather pace. In the 1980’s the Japanese ‘fifth

generation’ project, in which machine translation plays an important role, has been

launched to establish for Japan a major position in the future world economy.

Throughout, however, there have always been researchers motivated by idealism and

by scientific curiosity, and there have been sponsors willing to support basic research.

Machine translation was one of the first non-numerical applications of

computers. For more than a decade until the mid 1960’s it was an area of intensive

research activity and the focus of much public attention; but early expectations were

not fulfilled, promises of imminent commercially viable systems came to nothing,

and the problems and linguistic complexities became increasingly apparent and

seemed to be ever more intractable. After a widely publicized report compiled for the

major US sponsors, the ‘notorious’ ALPAC report, machine translation was generally

considered to have been a ‘failure’, and no longer worthy of serious scientific

consideration.

Critics and skeptics have been fond of repeating alleged mistranslations,

howlers that no human translator would perpetrate, in order to ridicule the whole

enterprise. The most popular example has been a story involving the translation of

two idioms from English into Russian and then backs again from Russian into

English: Out of sight, out of mind, and the spirit is willing but the flesh is weak.

According to some accounts the first came back as “invisible insanity” and the

second was as “The whiskey is all right but the meat has gone bad”; according to

others, however, the versions were “Invisible and insane” and “The vodka is good

but the meat is rotten”; and yet others have given “invisible lunatics” and “the

ghost is willing but the meat is feeble”. There have been various other permutations

and variants; such variety is typical of hearsay, and indeed, some accounts give the

languages as German and English, and others assert they were Chinese and English.

Page 7: Course Paper on Machine Translation

Nevertheless, the supposed translations are repeated to the present day as genuine

examples of the ‘literal-mindedness’ of machine translation.

It would seem that a likely source was an article by John A. Kouwenhoven

‘The trouble with translation’ in Harper's Magazine for August 1962: Our own

attempts to communicate with the Russians in their language may be no more

successful. Thanks to Robert E. Alexander, the architect, we can pass along this

cheering bit of news. According to Colonel Vernon Walters, President Eisenhower's

official interpreter, some electronic engineers invented an automatic translating

machine into which they fed 1,500 words of Basic English and their Russian

equivalents, claiming that it would translate instantly without the risk of human error.

In the first test they asked it to translate the simple phrase: “Out of sight, out of

mind”. Gears spun, lights blinked, and the machine typed out in Russian “Invisible

Idiot”. On the theory that the machine would make a better showing with a less

epigrammatic passage, they fed it the scriptural saying: “The spirit is willing, but the

flesh is weak”. The machine instantly translated it, and came up with “The liquor is

holding out all right, but the meat has spoiled”.

It is a good story, but its superficial plausibility is damaged by the lack of any

evidence of a US system at the time which could translate from English into Russian

– for obvious reasons the Americans wanted to translate from Russian into English –

and by the discovery that both examples were familiar apocrypha of translation

before there were any machine translation systems in operation. For example, in

April 1956, E. H. Ullrich was reported as saying: "Perhaps the popular Press is the

most attractive outlet for mechanical translations, because it does not really matter

whether these are right or wrong and amusing versions such as ‘the ghost wills but

the meat is feeble’ might make mechanical translation into a daily feature as

indispensable as the cross-word puzzle".

From the mid-1960’s research on machine translation continued at a reduced

level, largely ignored and forgotten not only by the general public but even by

linguists and computer scientists. In recent years, however, the situation has changed.

There are now operational systems in a number of large translation bureaus and

agencies; computers are producing readable translations for administrators, scientists,

Page 8: Course Paper on Machine Translation

and technicians at ever increasing volumes; translation systems are being marketed

on a commercial basis for microcomputers; many translators are now becoming

familiar with machine translation systems and with machine aids; and there is

growing scientific interest in machine translation within the Artificial Intelligence

community in the United States, in Japan and elsewhere. Machine translation can no

longer be dismissed because it is a reality.

With distant memories of the ‘failure’ of machine translation in the 1950’s and

1960’s and supported by apocryphal translation horrors, there are still many who do

not believe that computers can translate. It is true that few systems would pass the

‘Turing test’ by producing translations that could never be distinguished from the

output of fluent human translators. This book contains numerous examples of

translations produced by computer programs: some are clearly unacceptable texts by

whatever criteria; others however are the equal of some human translations and

would not be readily identified as computer renditions. The question of how good

machine translation should be in order to qualify, as ‘true’ translation is a particularly

thorny one, and still not really resolved. What matters in most cases is whether the

translation serves the needs of the recipient: a rough translation (human or machine

produced) might be quite adequate on some occasions; on others only a ‘perfect’

finished version is acceptable. Judgments of quality are necessarily both subjective

and highly constrained by personal needs and attitudes. What are probably most

surprising to those unfamiliar with the complexities of machine translation are

examples of errors that no human translator, however inexperienced, would ever

make. A genuine, not apocryphal, howler from the Systran system is cited by

Wheeler & Lawson (1982): “la Cour de Justice considere la creation d'un sixieme

posted'avocat general” was rendered as “the Court of Justice is considering the

creation of a sixth general avocado station”. Such examples are reassuring; there is

no fear of being taken over by computers – and these fears are real among some

translators. This book will show that machine translation is not a threat, it is not an

insidious dehumanizing destructive monster, and it is not “Golem astride the Tower

of Babel”. Machine translation should be seen as a useful tool that can relieve

translators of the monotony of much technical translation and spare them from the

Page 9: Course Paper on Machine Translation

wasteful expenditure of much tedious effort on documents of ephemeral or marginal

interest. Translators can then be employed where their skills are most wanted, in the

translation of sensitive diplomatic and legal documents, and in the translation of

cultural and literary texts.

The term ‘machine translation’ has now established itself as the general

accepted name for any system that uses an electronic computer to transform a text in

one language into some kind of text in another natural language. The related term

‘machine-aided translation’ in order to designate the use of mechanized aids for

translation has likewise established itself, by and large, as the generally accepted

term. Researchers and writers have commonly used also the alternative terms

‘mechanical translation’ and ‘automatic translation’, but these are now more rarely

encountered. For many writers the phrase ‘mechanical translation’ suggests

translation done in an automaton-like manner by a human translator; and this has

been the primary reason for the dropping of this term. While in English-speaking

countries the use of ‘automatic translation’ has generally been much less common

than ‘machine translation’, this nomenclature is the only possibility for the French

and Russians. There is no straight equivalent for ‘machine translation’. In the earlier

periods there was often talk of ‘translating machines’, but since the realization that

computers do not have to be designed specifically to function as translators this usage

has died away. In recent years there has been increasing use of the terms ‘computer

translation’ and ‘computer-aided translation’ - terms which are certainly more

accurate than ‘machine translation’ and ‘machine-aided translation’ – but in this book

the traditional, long-established, and still most common term ‘machine translation’

will be used, abbreviated throughout in the customary way as MT.

A number of other common terms need also to be defined at the outset. Firstly,

it has now become accepted practice to refer to the language from which a text is

being translated as the ‘source language’ (SL), and the language into which the text is

being translated as the ‘target language’ (TL). Secondly, there are now commonly

accepted terms for the processes involved: ‘analysis’ procedures accept source

language texts as input and derive representations from which ‘synthesis’ procedures

produce or generate texts in the target language as output. These processes may

Page 10: Course Paper on Machine Translation

involve various aspects of language structure: ‘morphology’ is concerned with the

inflectional forms and derivational variants of words or lexical items, ‘syntax’ is

concerned with the ways in which words combine in sentence structures, and

‘semantics’ is concerned with meaning relationships among sentences and texts.

Other terms will be introduced and defined as they arise. Between fully automatic

translation on the one hand and human translation on the other there are a number of

intermediate possibilities where there are various kinds of collaboration between man

and machine. The intervention can take place before, during or after the machine

processes. There can be human preparation of input, or ‘pre-editing’ in the MT

jargon; there can be human revision of the output, or ‘post-editing’. There can be

collaboration during the translation processes, when a human assistant may be asked

by the computer to resolve problems which it cannot deal with. Finally, a translator

may do most of the work alone and call upon the machine to assist with problems of

terminology. We may then refer to: Machine Translation proper, MT with post

editing, MT with edited or restricted input, human-aided MT, machine-aided human

translation, and human translation with no machine aids. The dividing line between

some interactive MT systems and machine-aided translation is blurred on occasions,

but in most cases there is little dispute. It includes some details about the

development of mechanized aids for translating, i.e. primarily automatic dictionaries

of various kinds, it does not include aspects of natural language processing which are

not directly concerned with the translation problem. Hence, although nearly all

computational linguistics, natural language processing in Artificial Intelligence and

certain approaches to automatic indexing and abstracting have their origins in MT

research, these ‘offshoots’ of MT will not be covered. Obviously, the field has to be

restricted in some way. The consequence of this restriction is that methods of

potential relevance to MT problems will not be dealt with in any detail if they have

not in fact been applied in any MT system. Likewise research which may have been

seen at one time as of potential relevance to MT but which in fact did not lead to any

kind of MT system will be ignored by and large. Another area, which must be

excluded for obvious reasons, is the development of computer technology and

programming, except where these developments have direct bearing on particular

Page 11: Course Paper on Machine Translation

features of MT systems or MT research methodology. An attempt has been made to

be as comprehensive as possible and to be as balanced as possible in the evaluation of

the contributions of the many MT projects in the (40) forty years’ history of MT

research.

1.2 Before the computer

The use of mechanical devices to overcome language barriers was suggested

first in the XVII century. There were two stimulants: the demise of Latin as a

universal language for scientific communication, and the supposed inadequacy of

natural languages to express thought succinctly and unambiguously. The idea of

universal languages arose from a desire both to improve international communication

and to create a ‘rational’ or ‘logical’ means of scientific communication.

Suggestions for numerical codes to mediate among languages were common.

Leibniz’s proposals in the context of his monadic theory are perhaps the best known.

Descartes made another proposal in comments on the 16th proposition of his famous

correspondent Anonymous. In a letter to Pierre Mersenne on 20 November 1629

Descartes described a proposed universal language in the form of a cipher where the

lexical equivalents of all known languages would be given the same code number.

Descartes wrote: “Mettant en son dictionnaire un seul chiffre qui se rapporte a aymer,

amare, philein, et tous les synonymes, le livre qui sera ecrit avec ces caracteres

pourra etre interprete par tous ceux qui auront ce dictionnaire. At the height of

enthusiasm about machine translation in the early 1960's some writers saw these 17th

proposals as genuine forerunners of machine translation. Becher's book, for example,

was republished under the title Zur mechanischen Sprachubersetzung: ein

Programmierungversuch aus dem Jahre 1661 (Becher 1962), indicating the

conviction of its editor that Becher's ideas foreshadowed certain principles of

machine translation. Apart from an ingenious script, Becher's book is distinguished

from others of this kind only by the size of the dictionary: 10,000 Latin words were

provided with coding. Like others, however, Becher failed to tackle the real

difficulties of providing equivalent entries in other languages (Greek, Hebrew,

German, French, Slav, and Arabic were proposed) and the necessary means to cope

Page 12: Course Paper on Machine Translation

with syntactic differences. The vast work by John Wilkins, An Essay towards a Real

Character and a Philosophical Language (1668), was a more genuine attempt at a

universal language in that it sought to provide a logical or rational basis for the

establishment of inter-language equivalencies. Wilkins’ aim was “a regular

enumeration and description of all those things and notions, to which marks or names

ought to be assigned according to their respective natures”, i.e. a codification which

embodied a universal classification of concepts and entities, a genuine interlingua.

All these writers recognized the problems of genuine differences between languages

that could not be captured completely in dictionaries, however ‘logically’

constructed. Many of them like Kircher advised their fellows to write in a simple

style and avoid rhetorical flourishes. Suggestions for mechanical dictionaries on

numerical bases continued to be made throughout the following centuries until the

middle of the present century. Couturat and Leau in their Histoire de la langue

universelle (1903) list numerous examples, including one by W. Rieger entitled

Zifferngrammatik, welche mit Hilfe der Worterbucher ein mechanisches

Uebersetzen aus einer Sprache in alle anderen ermoglicht (Code-grammar, which

with the help of dictionaries enables the mechanical translation from one language

into all others); a title which links the present mechanical age to the 17th century.

As the reference to Couturat and Leau implies, all these apparent precursors of

MT should be regarded more accurately as contributions to the ideal of a ‘universal

language’ and to the development of international auxiliary languages, of which the

best known is now Esperanto. Both concepts have in fact inspired many of those

engaged in machine translation. None of these proposals involved the construction of

machines; all required the human translator to use the tools provided in a

‘mechanical’ fashion, i.e. for man to simulate a machine. It was not until the

invention of mechanical calculators in the XIX and XX centuries that an automatic

device could be envisaged which could perform some translating processes. In fact,

the first explicit proposals for ‘translating machines’ did not appear until 1933, when

two patents were issued independently in France and Russia. In both cases, the

patents were for mechanical dictionaries.

Page 13: Course Paper on Machine Translation

A French engineer of Armenian extraction, Georges Artsrouni was issued a

patent on 22nd July 1933 for a translation machine which he called a “Mechanical

Brain”. The invention consisted of a mechanical device worked by electric motor for

recording and retrieving information on a broad band of paper which passed behind a

keyboard. The storage device was capable of several thousand characters, and was

envisaged by its inventor in use for railway timetables, bank accounts, commercial

records of all sorts, and in particular as a mechanical dictionary. Each line of the

broad tape would contain the entry word (SL word) and equivalents in several other

languages (TL equivalents); corresponding to each entry were coded perforations on

a second band, either paper or metal, which functioned as the selector mechanism. A

prototype machine was exhibited and demonstrated in 1937; the French railway

administration and the post and telegraph services showed considerable interest and

only the start of the Second World War in 1940 prevented installation of Artsrouni’s

invention. More important in retrospect was the patent issued in Moscow on 5

September 1933 to Petr Petrovich Smirnov-Troyanskii for the construction of a

“machine for the selection and printing of words while translating from one

language into another or into several others simultaneously”. Troyanskii

envisaged three stages in the translation process; the machine was involved only in

the second stage, performing as an automated dictionary. In the first stage a human

editor knowing only the source language was to analyze the input text into a

particular ‘logical’ form: all inflected words were to be replaced by their base forms

(e.g. the nominative form of a noun, the infinitive form of a verb) and ascribed their

syntactic functions in the sentence. For this process Troyanskii had devised his own

‘logical analysis symbols’. In the second stage the machine was designed to

transform sequences of base forms and ‘logical symbols’ of source texts into

sequences of base forms and symbols of target languages. In the third stage an editor

knowing only the target language was to convert this sequence into the normal forms

of his own language. Troyanskii envisaged both bilingual translation and multilingual

translation. Although the machine was assigned the task only of automating the

dictionary, it is interesting to note that Troyanskii believed that “the process of

logical analysis could itself be mechanized, by means of a machine specially

Page 14: Course Paper on Machine Translation

constructed for the purpose” (quoted by Panov 1960a). It was this vision of the next

steps beyond a simple mechanical dictionary that marks Troyanskii's proposal as a

genuine precursor of machine translation.

In the 1933 patent, the technical implementation proposed was a purely

mechanical device, a table over which passed a tape listing in vertical columns

equivalent words from various languages. But, by 1939 he had added an improved

‘memory’ device operating with photo-elements (Delavenay 1960; Mounin 1964),

and by May 1941 it appears that an experimental machine was operational.

Troyanskii in fact went further towards the electronic computer; in 1948 he had a

project for an electro-mechanical machine. Similar to the Harvard Mark I machine

developed between 1938 and 1942, and which is regarded as a forerunner of the

ENIAC computer. Troyanskii was clearly ahead of his time; Soviet scientists and

linguists failed to respond to his proposal when he sought their support in 1939 and

later “the Institute of Automation and Telemechanics of the Academy of Sciences

was equally unforthcoming in 1944”. In retrospect, there seems to be no doubt that

Troyanskii would have been the father of machine translation if the electronic digital

calculator had been available and the necessary computer facilities had been ready.

History, however, has reserved for Troyanskii the fate of being an unrecognized

precursor; his proposal was neglected in Russia and his ideas had no direct influence

on later developments; it is only in hindsight that his vision has been recognised.

1.3 The first beginnings (1946-1949)

The electronic digital computer was a creation of the Second World War: the

ENIAC machine at the Moore School of Electrical Engineering in the University of

Pennsylvania was built to calculate ballistic firing tables; the Colossus machine at

Bletchley Park in England was built to decipher German military communications.

Immediately after the war, projects to develop the new calculating machines were

established at numerous centers in the United States and Great Britain. The first

applications were naturally in the fields of mathematics and physics, but soon the

enormous wider potential of the “electronic brain” were realized and nonnumeric

applications began to be contemplated.

Page 15: Course Paper on Machine Translation

The first suggestion that electronic computers could be used to translate from one

language into another seems to have been made during conversations in New York

between Andrew D. Booth and Warren Weaver.

Warren Weaver was at this time vice president of the Rockefeller Foundation.

During the war Weaver had served on a scientific mission to investigate Britain's

weapons development, and at the Rockefeller Foundation he was closely involved in

the sponsorship of computer research and development. Booth had become interested

in automatic digital calculation while working at the British Rubber Producers’

Association in Welwyn Garden City, and had started to build a machine for

crystallographic calculations. In 1945 he was appointed a Nuffield Fellow in the

Physics Department at Birkbeck College in the University of London under Professor

J. D. Bernal, where he constructed a relay calculator during 1945 and 1946 and

initiated plans for computational facilities in the University of London. As a

consequence of this work and the efforts of Bernal he obtained funds to visit the

United States in 1946 under the auspices of the Rockefeller Foundation. There he

visited all the laboratories engaged in computer research and development, at

Princeton, MIT, Harvard, and Pennsylvania.

The discussions that Booth had with Warren Weaver were entirely on the

subject of coming over to look into the question of acquiring the techniques for

building a machine for the University of London based on American experience.

Then, Booth submitted a report on computer development with particular reference to

x-ray crystallography, and he was offered a Rockefeller fellowship to enable him to

work at an institution of his own choice in the United States the following year.

Booth selected the von Neumann group at the Institute for Advanced Study,

Princeton University.

According to Booth (1985): “The discussion then was entirely on the question

of the Rockefeller Foundation financing a computer for the University of London,

and Weaver pointed out that there was very little hope that the Americans would fund

a British computer to do number crunching, although they might be interested if we

had any additional ideas for using the machine in a nonnumeric context. In the mid

1940's he had already thought about non-numerical applications from conversations

Page 16: Course Paper on Machine Translation

with A. M. Turing. One of that was in fact translation, although at that time he had

thought only of using the machine as a dictionary. Weaver suggested treating

translation as a cryptography problem. Weaver had in fact already on 4th March

1947, just before this meeting with Booth, written to Norbert Wiener of the

Massachusetts Institute of Technology, one of the pioneers in the mathematical

theory of communication, about the possibility of Machine Translation. Once,

Weaver wrote that recognizing fully, even though necessarily vaguely, the semantic

difficulties because of multiple meanings. He has wondered if it was unthinkable to

design a computer which would translate. Even if it would translate only scientific

material, and even if it did produce an inelegant result, it would seem worthy. Also

knowing nothing official about, but having guessed and inferred considerable about,

powerful new mechanized methods in cryptography, one naturally wonders if the

problem of translation could conceivably be treated as a problem in cryptography.

It is evident that the first serious discussions and investigations of the

possibilities of machine translation took place during 1947, beginning with Weaver's

letter to Wiener and his meeting with Booth in early March. However, at a later date

in 1955 when writing the 'Historical introduction' to the MT collection he edited with

Locke, Booth recollected the first mention of MT as having occurred during his 1946

visit. This has been generally accepted as the ‘birth’ date of MT; however, in other

later publications Booth gives the date 1947, and he has now confirmed the March

1947 meeting as the one when the MT discussion with Weaver occurred. Booth was

the first to suggest this possible use of mechanical translation; he also may

legitimately be regarded as the pioneer of what is now known as Artificial

Intelligence. In an essay written during September 1947, Booth mentions a number of

possible ways in which the new computers could demonstrate their ‘intelligence’:

1. Various games, e.g. chess, naught and crosses, bridge, poker;

2. The learning of languages;

3. Translation of languages;

4. Cryptography;

5. Mathematics

Page 17: Course Paper on Machine Translation

Evidently, Weaver and Turing were thinking along similar lines independently

and probably, others too. As there were no facilities available at Birkbeck College,

Booth began construction of a small computer at the laboratories of the British

Rubber Producers’ Research Association in Welwyn Garden City near London. The

machine was operational by 12th May 1948 and a demonstration was given on 25th

May to Warren Weaver and Gerard Pomerat, also of the Rockefeller Foundation. On

this occasion Weaver met Richard H. Richens, with whom Booth had been

collaborating in experiments on mechanical dictionaries. Richens had first met Booth

on the 11th November 1947. His interest in mechanical translation had arisen

independently out of experiments with punched cards for storing information at the

Commonwealth Bureau of Plant Breeding and Genetics, where he was Assistant

Director.

The idea of using punched cards for automatic translation arose as a spin-off, fuelled

by my realization as editor of an abstract journal ''Plant Breeding Abstract'' that

linguists conversant with the grammar of a foreign language and ignorant of the

subject matter provided much worse translations than scientists conversant with the

subject matter but hazy about the grammar. Richens is to be credited with the first

suggestion of the automatic grammatical analysis of word-endings. He proposed the

segmenting words into their stems and endings, both to reduce the size of dictionaries

and to introduce grammatical information into a dictionary translation system. For

example, in the case of the Latin verb amat a search was made for the longest

matching stem, i.e. ‘am-’, and for the ending ‘-at’. The stem provides the English

translation love and the ending gives the grammatical information ‘3rd person

singular’. In this way grammatical annotations augment a strict word-by-word

dictionary ‘translation’. The validity of the approach was tested by hand and by using

punched card machinery on a wide variety of languages; the texts were taken from

abstracts in plant genetics. From the French text: Il n’est pas etonn*ant de

constat*er que les hormone*s de croissance ag*issent sur certain*es espece*s,

alors qu'elles sont in*oper*antes sur d’autre*s, si l’on song*e a la grand*e

specificite de ces substance*s. (Where the stars (*) indicate automatic

segmentations).

Page 18: Course Paper on Machine Translation

The English translation: v not is not/step astonish v of establish v that/which? v

hormone m of growth act m on certain m species m, then that/which? v not operate m

on of other m if v one ream/consider z to v great v specificity of hose substance m.

Where v indicates French word not translated, m "multiple, plural r dual", z

"unspecific", and slashes alternative translations). These tentative experiments by

Booth and Richens were known to very few. Weaver's memorandum, which was

literally the first suggestion experts had seen that the new electronic computers could

be used as translating machines, launched machine translation as a scientific

enterprise in the United States and subsequently elsewhere. Its historic impact is

unquestionable; in his memorandum Weaver dates the origin of his speculations

about MT to his wartime experience with electronic computers and to stories of

startling achievements in cryptanalysis using computers.

1.4 Weaver's memorandum (1949)

Weaver’s memorandum concentrated more on the general strategies and long-

term objectives of MT than on the more technical problems Booth and Richens had

been tackling. Because of its historic importance it is worth enumerating in some

detail the issues and problems raised by Weaver. He raised four points: the problem

of multiple meaning, the logical basis of language, the application of communication

theory and cryptographic techniques, and the possibilities of language universals. The

problem of deciding which specific meaning an ambiguous word may have in a

particular text was, he suggested, solvable in principle if a sufficient amount of the

immediate context is taken into account. The practical question of how many contexts

are necessary could be answered by a statistical study of different types of texts on a

variety of subject matters. Weaver explicitly rejected the idea of actually storing in

dictionaries long sequences of words for this purpose, but did suggests that “some

reasonable way could be found of using the micro context to settle the difficult cases

of ambiguity. He expressed optimism about finding logical aspects in languages.

Weaver believed that the translation problem could be largely solved by “statistical

semantic studies”. For Weaver the “most promising approach of all” was the

Page 19: Course Paper on Machine Translation

investigation of language invariants or universals. He linked this again with logical

structures of language and with probabilistic uniformities. The analogy he suggested

was “of individuals living in a series of tall closed towers, all erected over a common

foundation. Perhaps the way is to descend, from each language, down to the common

base of human communication - the real but as yet undiscovered universal language”.

1.5 From Weaver to the first MT conference (1950-1952)

Weaver’s memorandum brought to the attention of a wide circle the

possibilities of a new and exciting application of the computers whose potentialities

were being discovered and proclaimed with enthusiasm and optimism at this time.

But, it did more. It indicated potentially fruitful lines of research in statistical

analyses of language, on the logical bases of language, and on semantic universals of

language. In addition, it pointed to some actual, even if tentative, achievements in the

work of Booth and Richens. It was, however, received with considerable skepticism

by many linguists who rejected it for its naivety in linguistic matters and for its

unfounded assumptions on the logicality of language, and they were naturally

skeptical about the possibility of formalizing language and translation processes.

The press had also noticed the memorandum. Booth’s APEXC computer program

was described as an “electronic translator”, at which an operator “could select which

of a dozen or more languages he desired to translate. As fast as he could type the

words, say, in French, the equivalent in Hungarian or Russian would issue on the

tape.”

1.5.1 First MT studies.

Weaver’s own favored approach, the application of cryptanalytic techniques,

was immediately recognised as mistaken. Confusion between the activities of

deciphering and translation arise whenever the same person does both. Obviously, no

translating is involved when an English-speaking recipient deciphers an English

message into English. Likewise, the decipherment of the highly complex Enigma

code used by Germany in the Second World War, with its immensely complex

Page 20: Course Paper on Machine Translation

sequences of manipulations and transpositions, was not translation; it was only after

the German texts had been deciphered that they were translated. The Colossus

computers at Bletchley Park were applied to cracking the cipher, not to translating the

German text into English. In practice, the cryptanalyst generally knows what the

language is of the texts to be deciphered and often what their content is likely to be

and the circumstances in which the message was transmitted. All this helps him to

guess which letters and words are likely to be most frequent in the text. In the case

cited by Weaver, the decipherment was based on “the frequencies of the letters, pairs

of letters, etc. in English”; fortunately they were much the same in Turkish and the

original could be interpreted. Though the cryptanalytic approach was mistaken, there

were sufficient stimulating ideas in Weaver’s paper to launch MT as a serious line of

research in the United States. During the next two years, individuals and groups

began MT studies at a number of locations, the Massachusetts Institute of

Technology (MIT), the University of Washington in Seattle, the University of

California at Los Angeles (UCLA), the National Bureau of Standards (NBS) also in

Los Angeles and the RAND Corporation nearby at Santa Monica. On 10th January

1950, Erwin Reifler circulated privately the first of a series of studies on MT. Reifler

was a Sinologist of German origin, head of the Department of Far Eastern and Slavic

Languages and Literature at the University of Washington in Seattle. Recognizing the

problem of multiple meanings as an obstacle to word-for-word translation of the kind

attempted by Booth and Richens, Reifler introduced the concepts of ‘pre-editor’ and

‘post editor’. The human ‘pre-editor’ would prepare the text for input to the computer

and the ‘post editor’ would resolve residual problems and tidy up the style of the

translation. One suggestion was that the pre-editor should indicate the grammatical

category of each word in the source language (SL) text by adding symbols or diacritic

marks, e.g. to distinguish between the noun convict and the verb convict. The post-

editor’s role was to select the correct translation from the possibilities found by the

computer dictionary and to rearrange the word order to suit the target language. As

we shall see, the concepts of pre-editor and post-editor recur in one form or another

throughout the development of MT research. Following Weaver’s suggestion for

statistical studies of micro context for resolving problems of multiple meaning,

Page 21: Course Paper on Machine Translation

Abraham Kaplan at the RAND Corporation investigated polysemy in mathematics

texts. A group of test subjects were presented with a set of words, each with a number

of possible meanings, and asked to select the most applicable sense. Kaplan limited

the test to nouns, verbs and adjectives on the assumption that “these are the major

carriers of the content of any discourse, and probably more markedly exhibit

ambiguities”. Each word was presented first in isolation, then together with preceding

and following word, and finally the whole sentence. It was found that the “most

practical context is one word on each side, increased to two if one of the context

words is a particle”, i.e. an article, preposition or conjunction. Despite its limitations

and deficiencies and the tentativeness of the conclusions, this study encouraged hopes

that problems of ambiguity were tractable and that statistical analyses could

contribute useful linguistic data for MT. In the latter half of 1950, a survey was

conducted by W. F. Loomis on behalf of Weaver to find out all those who were

interested in MT and what research was underway. The survey revealed a surprising

amount of activity already apart from Booth, Richens and Reifler; two groups had

been set up in California. One was at the RAND Corporation in Santa Monica under

J. D. Williams, and Kaplan's paper was to be the first of a series of MT studies. Harry

D. Huskey of the National Bureau of Standards in Los Angeles had formed the other,

with the intention of using the SWAC for MT research. The group included Victor A.

Oswald of the German Department at UCLA and William E. Bull of the UCLA

Spanish Department, and was soon joined by Kenneth E. Harper of the UCLA Slavic

Languages Department. In support of its work, the group received some funds from

the Rockefeller Foundation in July 1951. It is clear that word-for-word translation of

a language like German would produce obviously unsatisfactory results, Oswald and

Fletcher proposed a detailed grammatical coding of German sentences indicating

syntactic functions of nouns and verb forms in clauses and enabling the identification

of ‘noun blocks’ and ‘verb blocks’. On the basis of the codes, certain sequences

were identifiable as candidates for rearrangement when the output was to be in

English. Oswald and Fletcher concluded that syntax “does not constitute, as had been

thought by some, a barrier to mechanical translations”; they stressed the problems of

solving the “lexicographic difficulties” of Machine Translation.

Page 22: Course Paper on Machine Translation

1.5.2 The decade of high expectation and disillusion, 1954-1966

The earliest systems consisted primarily of large bilingual dictionaries where

entries for words of the source language gave one or more equivalents in the target

language, and some rules for producing the correct word order in the output. It was

soon recognised that specific dictionary-driven rules for syntactic ordering were too

complex and increasingly ad hoc, and the need for more systematic methods of

syntactic analysis became evident. A number of projects were inspired by

contemporary developments in linguistics, particularly in models of formal grammar,

and they seemed to offer the prospect of greatly improved translation. Optimism

remained at a high level for the first decade of research, with many predictions of

imminent "breakthroughs". However, disillusion grew as researchers encountered

"semantic barriers" for which they saw no straightforward solutions. There were

some operational systems – the Mark II system installed at the USAF Foreign

Technology Division, and the Georgetown University system at the US Atomic

Energy Authority and at Euratom in Italy – but the quality of output was

disappointing. By 1964, the US government sponsors had become increasingly

concerned at the lack of progress; they set up the Automatic Language Processing

Advisory Committee (ALPAC), which concluded in a famous 1966 report that MT

was slower, less accurate and twice as expensive as human translation and that "there

is no immediate or predictable prospect of useful machine translation." It saw no need

for further investment in MT research; and instead it recommended the development

of machine aids for translators, such as automatic dictionaries, and the continued

support of basic research in computational linguistics.

1.5.3 The ALPAC report and its consequences, 1966-1980

Although widely condemned as biased and shortsighted, the ALPAC report

brought a virtual end to MT research in the United States for over a decade and it had

Page 23: Course Paper on Machine Translation

great impact elsewhere in the Soviet Union and in Europe. However, research did

continue in Canada, in France and in Germany. Within a few years the Systran

system was installed for use by the USAF (1970), and shortly afterwards by the

Commission of the European Communities for translating its rapidly growing

volumes of documentation (1976). In the same year, another successful operational

system appeared in Canada, the Meteo system for translating weather reports,

developed at Montreal University. In the 1960s in the US and the Soviet Union MT

activity had concentrated on Russian-English and English-Russian translation of

scientific and technical documents for a relatively small number of potential users,

who would accept the crude unrevised output for the sake of rapid access to

information. From the mid-1970s onwards the demand for MT came from quite

different sources with different needs and different languages. The administrative and

commercial demands of multilingual communities and multinational trade stimulated

the demand for translation in Europe, Canada and Japan beyond the capacity of the

traditional translation services. The demand was now for cost-effective machine-

aided translation systems that could deal with commercial and technical

documentation in the principal languages of international commerce.

1.5.4 The 1980s.

The 1980s witnessed the emergence of a wide variety of MT system types, and

from a widening number of countries. First there were a number of mainframe

systems, whose use continues to the present day. Apart from Systran, now operating

in many pairs of languages, there was Logos; the internally developed systems at the

Pan American Health Organization (Spanish-English and English-Spanish); the Metal

system (German-English); and major systems for English-Japanese and Japanese-

English translation from Japanese computer companies. The wide availability of

microcomputers and of text-processing software created a market for cheaper MT

systems, exploited in North America and Europe by companies such as ALPS,

Weidner, Linguistic Products, and Globalink, and by many Japanese companies, e.g.

Page 24: Course Paper on Machine Translation

Sharp, NEC, Oki, Mitsubishi, Sanyo. Other microcomputer-based systems appeared

from China, Taiwan, Korea, Eastern Europe, the Soviet Union, etc.

Throughout the 1980s research on more advanced methods and techniques

continued. For most of the decade, the dominant strategy was that of ‘indirect’

translation via intermediary representations, sometimes interlingual in nature,

involving semantic as well as morphological and syntactic analysis and sometimes

non-linguistic ‘knowledge bases’. The most notable projects of the period were the

GETA-Ariane, SUSY, Mu, DLT, Rosetta, the knowledge-based project at Carnegie-

Mellon University and two international multilingual projects: Eurotra, supported by

the European Communities, and the Japanese CICC project with participants in

China, Indonesia and Thailand.

1.5.5 The early and the late 1990s

The end of the decade was a major turning point. Firstly, a group from IBM

published the results of experiments on a system based purely on statistical methods.

Secondly, certain Japanese groups began to use methods based on corpora of

translation examples, i.e. using the approach now called ‘example-based’ translation.

In both approaches the distinctive feature was that no syntactic or semantic rules are

used in the analysis of texts or in the selection of lexical equivalents; both approaches

differed from earlier ‘rule-based’ methods in the exploitation of large text corpora. A

third innovation was the start of research on speech translation, involving the

integration of speech recognition, speech synthesis, and translation modules – the

latter mixing rule-based and corpus-based approaches. The major projects are at ATR

(Nara, Japan), the collaborative JANUS project, and in Germany the government-

funded Verb Mobil project. However, traditional rule-based projects have continued,

e.g. the Catalyst project at Carnegie-Mellon University, the project at the University

of Maryland, and the ARPA-funded research at three US universities. Another feature

of the early 1990s was the changing focus of MT activity from ‘pure’ research to

practical applications, to the development of translator workstations for professional

Page 25: Course Paper on Machine Translation

translators, to work on controlled language and domain-restricted systems, and to the

application of translation components in multilingual information systems.

These trends have continued into the later 1990s. In particular, the use of MT

and translation aids by large corporations has grown rapidly – a particularly

impressive increase is seen in the area of software localization. There has been a huge

growth in sales of MT software for personal computers and even more significantly,

the growing availability of MT from on-line networked services. The demand has

been met not just by new systems but also by ‘downsized’ and improved versions of

previous mainframe systems. While in these applications, the need may be for

reasonably good quality translation (particularly if the results are intended for

publication), there has been even more rapid growth of automatic translation for

direct Internet applications (electronic mail, Web pages, etc.), where the need is for

fast real-time response with less importance attached to quality. With these

developments, MT software is becoming a mass-market product, as familiar as word

processing and desktop publishing.

Conclusion

It has long been a subject of discussion whether machine translation and

computer-assisted translation could convert translators into mere editors, making

them less important than the computer programs. The fear of this happening has led

to a certain rejection of the new technologies on the part of translators, not only

because of a possible loss of work and professional prestige, but also because of

concern about a decline in the quality of production. Some translators totally reject

machine translation because they associate it with the point of view that translation is

merely one more marketable product based on a calculation of investment versus

profits. They define translation as an art that possesses its own aesthetic criteria that

have nothing to do with profit and loss, but are rather related to creativity and the

power of the imagination. This applies mostly, however, to specific kinds of

translation, such as that of literary texts, where polysemy, connotation and style play

a crucial role. It is clear that computers could not even begin to replace human

translators with such texts. In fact translators should recognize and learn to exploit

Page 26: Course Paper on Machine Translation

the potential of the new technologies to help them to be more rigorous, consistent and

productive without feeling threatened. Translators need to accept the new

technologies and learn how to use them to their maximum potential as a means to

increased productivity and quality improvement.

II. COMMON ERRORS IN MACHINE TRANSLATION

Translation involves the production of a text in one language that was inspired by an

existing text in another language, such that the two texts are in some sense 'the same.

The relationship between proper translation and paraphrase involves decoding the

message from an existing text and re-encoding the message in a new text. This results

in two texts different in form with the same message. Translation parallels this except

that the source and target texts are in different languages.

2.1 The quality of translation

Paraphrases vary in quality as a function of their accuracy, i.e. the degree to

which the message they convey, when decoded, matches the message of the original

text. This is true whether they are translations or not. Since translations are often

performed in formal and public circumstances, the question of accuracy is raised with

a greater percentage of translations than with monolingual paraphrases. But accuracy

is important in all cases, and a complaint regarding accuracy should be taken

seriously. Literary translations adhere to a higher standard of quality, a standard that

considers the form of the text as well as the message transmitted. The target text in a

good translation is adjusted in form so that it corresponds to the source text. That is,

the translation of a sonnet should be a sonnet, and the translation of a joke should be

funny. Thus a translation may be criticized on both form and semantic content.

2. 2: Mechanical dictionaries

Page 27: Course Paper on Machine Translation

The creation of an automatic dictionary is the first and most obvious task of an

MT system. Mechanical dictionaries were the central concern of all the earliest

Machine Translation researchers and they are still crucial for the efficient operation of

present Machine Translation systems. Like Artsrouni and Troyanskii, many early

researchers tended to see the translation process almost exclusively in terms of

consulting dictionaries for finding Target Language (TL) words equivalent to Source

Language (SL) words. The resulting ‘dictionary translations’ presented the TL output

in the same word sequence as the SL input, i.e. ‘word-for-word’ translations. They

knew that this would not produce good translations; they expected the results to be

very poor and in need of considerable editing. Before any ‘word-for-word’ translations

had been seen, Reifler suggested a pre-editor and a post-editor, and the unintelligibility

of the results of Richens and Booth’s attempts confirmed the need for, at the least, post

editing.

Nevertheless, the ability of many readers to make some sense of these dictionary

translations encouraged MT researchers to believe that with suitable modifications the

‘word-for-word’ approach could in the end produce reasonable output. As we have

seen, Yngve considered that they were “surprisingly good” and worth taking as first

approximations to be worked on. The mechanizations of dictionary procedures posed

problems of a technical nature. Research on Machine Translation began at a time when

computers were limited in their storage capacities and slow in access times. There was

much discussion of storage devices and mechanisms for improving access times.

Booth (1955) and Stout (1954), for example, assessed the relative merits of paper tape,

punched cards and magnetic tape as external storage means and the various

possibilities for internal ‘memory’ storage, cathode-ray-tube dielectric stores, vacuum

tubes, magnetic drums, photographic drums, etc. Since the external storage could only

be searched serially, the most efficient method of dictionary lookup was to sort all the

words of the SL text into alphabetical order and to match them one by one against the

dictionary entries. Once found, entries could often be stored internally where faster

access was possible. Various proposals were made for efficient searching of internal

stores, including the sequencing of items by frequency, the ‘binary cut’ method first

put forward by Booth (1955a), and the letter-tree approach of Lamb. A popular

Page 28: Course Paper on Machine Translation

method for reducing dictionary size was the division of words into stems and endings.

In languages like German and Russian it was obviously wasteful to include every

inflected form of nouns and verbs. The familiar regularities of noun and verb

paradigms encouraged researchers to investigate methods of morphological analysis to

identify stems and endings. However, there are so many peculiarities and irregularities

in the morphology of languages that procedures turned out to be more complex than

expected; as a result, when larger storage mechanisms with fast access times became

available many MT researchers went back to the older system of storing full forms in

dictionaries. Obviously, dictionaries cannot always include all the words occurring in

SL texts. A problem for all MT systems is to establish acceptable methods for dealing

with missing words; basically, there are two approaches, either to attempt some kind

of analysis and translation, or to print out the original unloaded SL form. In both cases,

there is a further problem with the rest of the sentence; whether to attempt an

incomplete translation or to give up and produce no translation. In experimental

Machine Translation systems it is obviously reasonable to admit failure, but in

operational systems it is desirable, on the whole, to produce some kind of translation.

2. 3: Polysemy and semantics.

The most obvious deficiency of any word-for-word translation, whether

mechanized or not, is that the order of words in the resulting TL text is more often

wrong than correct. As we have seen, it was clear to Oswald and Fletcher that

translation of German texts into English demanded some kind of structural analysis of

the German sentences. At the simplest level, such analysis may take into account

morphological features, such as the endings of nouns, adjectives and verbs, or basic

syntactic sequences, such as noun-adjective and subject-verb relations. As we shall

see, it is possible to use this kind of information to devise procedures for

rearrangement in basically ‘word-for-word’ systems. But, in order to go beyond the

inherent limitations of the word-for-word approach, the analysis of syntactic structures

must involve the identification of phrase and clause relationships. Methods of

syntactic analysis will be the subject of the next section. The second obvious problem

is that there are rarely one-to-one correspondences in the vocabularies of natural

Page 29: Course Paper on Machine Translation

languages. In most cases, a particular SL word may correspond to a number of

different TL words, so that either the MT system prints out all the possibilities or it

attempts to select the one that is most appropriate for the specific text in question. The

first option was adopted by many systems, as we shall see, often as an intermediate

stopgap; the problem of selecting the right TL equivalent was left to the post-editor.

Attempts to deal with the problem took a number of approaches.

The difficulty occurs usually because the SL word has what Weaver and many after

him called ‘multiple meanings’. Linguists distinguish between homonyms and

polysemes; homonyms are words like bank that have two or more distinct and

unrelated meanings (‘geological feature’ or ‘financial institution’); polysemes are

words like face that reflect different shades of meaning according to context. They

distinguish also between homophones (words which sound the same but have

different meanings) such as pear, pair and pare, and homographs (words which are

spelled the same but have different meanings) such as tear (‘crying’ versus ‘ripping’).

Fortunately, the homophone problem is irrelevant since MT deals only with written

texts. For practical purposes it is also immaterial whether the SL word is a homograph

or a polyseme: the problem for MT is the same, the relevant meaning for the context

must be identified and the appropriate TL form must be selected. Consequently, it is

now common in MT research to refer to methods of ‘homograph resolution’, whether

the words concerned are strictly homographs or not. Sometimes the TL vocabulary

makes finer sense distinctions than the SL. There are familiar examples in translating

from English into French or German the verb know may be conveyed by savoir or

connaître in French and by wissen or kennen in German; likewise the English river

may be either rivière or fleuve in French and either Fluss or Strom in German. In

neither case can we say that the English words have more than one meaning; it is just

that French and German make distinctions, which English does not. Nevertheless, in

the context of a MT system the problem of selecting the correct TL form is much the

same as when the SL form is a genuine homograph or polyseme. MT systems do,

however, differ according to whether this type of SL-TL difference is tackled at the

same stage as SL homograph resolution or not. The difficulties are further

compounded in languages like English where many words may function as nouns,

Page 30: Course Paper on Machine Translation

verbs or adjectives without any formal distinctions; e.g. control can be a verb or noun,

green can be an adjective or a noun. The fact that there can be stress differences, e.g.

between the verb permit and the noun permit, is of no assistance. For practical

purposes these forms are also treated as homographs and much the same procedures

for ‘homograph resolution’ are applied. Various methods for tackling such SL-TL

lexical differences have been proposed. One has already been mentioned, the

identification of grammatical category either by morphological clues or by syntactic

analysis. For example, the endings ‘-ed’ and ‘-ing’ generally indicate participial forms

of English verbs (although they may be functioning as adjectives). Similarly, if in a

two-word sequence the first is definitely an adjective the second is probably a noun.

Therefore, homographs that happen to belong to different syntactic categories may

sometimes be distinguished in this way. Another method is to reduce the incidence of

homography in the MT dictionaries. The concept of the ‘micro-glossary’ was proposed

not only to keep the size of dictionaries reasonably small but also to minimize

problems of ‘multiple meanings’. It was maintained, for example, that the Russian вид

was to be translated usually as species in biological contexts and not as view, shape or

aspect. A micro-glossary for Russian-English translation in biology could, therefore,

include just one of the English equivalents. In many cases the entry has to be the

equivalent, which is most often correct. In physics, for example, Russian измененйе

is usually equated with change; although in some contexts other translations may be

better, the one, which fits best most frequently, should be selected. The suggestion by

Weaver was to examine the immediate context of a word. As we have seen, Kaplan

concluded that a five-word sequence was in general sufficient ‘micro context’ for

disambiguation, i.e. for identifying the particular meaning of a polyseme. There are

two ways in which immediate context can be implemented: one by expanding

dictionary entries to include sequences of two or more words, i.e. phrases, the other by

testing for the occurrence of specific words. For example, if the word образование is

modified by кристаллическое then it is to be translated formation (rather than

education); either the dictionary includes the whole phrase or the analysis procedure

tests for the particular adjective. The dictionary solution obviously requires storage

facilities of sufficient capacity, and it is also more appropriate when phrases are

Page 31: Course Paper on Machine Translation

‘idiomatic’, i.e. when the meaning (translation) of the phrase as a whole cannot be

deduced (or constructed) from its individual words. Apart from familiar idioms such

as hold one's tongue, not move a finger, red herring and blue blood, it could include

verbal phrases such as make away with, draw forth, look up and pass off, and noun

phrases such as full speed, upper class and brute force. A more fundamental use of

contextual information is the search for semantic features that are common to or

prominent in the sentence or text as a whole, and to use this information to decide on

the most fitting translation for SL words. This method involves the investigation of

semantic ‘invariants’ or semantic regularities in vocabulary and texts, and necessarily

goes far beyond the examination of lexical equivalents between languages. It involves,

for example, the investigation of synonymy and paraphrase, of semantic ‘universals’

or ‘primitive’ elements (e.g. features such as ‘human’, ‘animate’, ‘liquid’, etc.), and of

semantic relations within sentences and texts (e.g. agent-action, cause-effect, etc.)

Finally, the problem of polysemy may simply be avoided completely by insisting that

texts input to a MT system be written in a regularized and normalized fashion. In other

words, writers are encouraged not to be ambiguous, or rather not to include words and

phrases, which the MT system in use has difficulty in disambiguating. The obverse of

this is to ‘solve’ polysemy by using a highly restricted form of TL as output, a kind of

‘pidgin’ language with its own idiosyncratic vocabulary usages. As we have seen,

Dodd made the first suggestion of this approach; the groups at the University of

Washington and at Cambridge were particularly interested in MT pidgin and methods

of improving output of this kind. In theory, any of these methods can be used in any

MT system; in practice, particular MT systems have emphasized one or two

approaches, they have concentrated on exploiting their full potentialities and have

generally neglected the alternatives. Concentration on the contextual and micro-

glossary approaches was characteristic of the MT groups at Rand and Michigan.

Concentration on the dictionary and lexicographic approaches was characteristic of the

groups at Harvard, at the University of Washington and at IBM. Concentration on text

semantics was pursued most strongly by the Milan group with its ‘co-relational

analysis’ approach and by the Cambridge group with its ‘thesaurus’ approach.

Page 32: Course Paper on Machine Translation

2. 4: Morphological analysis

In order to perform any kind of syntactic analysis the grammatical categories

(noun, verb, adjective, adverb, etc.) of the words of sentences must be determined. The

first step of analysis in any MT system is, however, the identification of the words in

the SL text. This is relatively easy in English and most European languages, since

words are separated by spaces in written text, but it is not for example in languages

such as Chinese and Japanese where there are no external markers of word boundaries.

Obviously, dictionary entries could indicate the grammatical categories (‘word class’

or ‘part of speech’) of all SL words. However, it was clearly unnecessary to include

every inflected form of a noun or a verb, particularly in languages such as Russian and

German. The familiar regularities of noun and verb paradigms encouraged researchers

to investigate methods of morphological analysis that would identify stems and

endings. To give an English example, the words: analyzes, analyzed, and analyzing

might all be recognized as having the same stem analyze and the common endings -s, -

ed, -ing. At the same time, identification of endings was a first step towards the

determination of grammatical categories, e.g. to continue the example: -s indicates a

plural noun form or a third person singular present verb form, -ed indicates a past verb

form, and - ing a present participle or adjectival form, etc. As these examples

demonstrate, however, many (perhaps most) endings are ambiguous, even in Russian,

and the final establishment of the grammatical category of particular words in text

takes place during syntactic analysis. Morphological analysis deals necessarily with

regular paradigms; irregular forms, such as the conjugation of verbs such as be and

have, and the plural forms of nouns such as geese and analyses, are generally dealt

with by inclusion of the irregularities in full forms in the dictionary.

2. 5: Syntactic analysis.

The first step beyond the basic word-by-word approach is the inclusion of a few

rearrangement rules, such as the inversion of ‘noun-adjective’ to ‘adjective-noun’,

e.g. in French- English translation. In many early MT systems rearrangement rules

were often initiated by codes attached to specific dictionary entries. Examples are to

be found in the 1954 Georgetown-IBM experiment, and in the experiment by Panov

Page 33: Course Paper on Machine Translation

and his colleagues shortly afterwards in the Soviet Union. When there were

differences of syntactic structure more complex than inversion, the solution was often

the inclusion of phrases in the dictionary, i.e. rather like idiomatic expressions. This

approach was expanded and refined as the ‘lexicographic’ approach of the University

of Washington. Rearrangement rules may take into account fairly long sequences of

grammatical categories, but they do not imply any analysis of syntactic structure, e.g.

the identification of a noun phrase. The next step beyond the basic word-for-word

approach is therefore the establishment of syntagmas, such as noun phrases (nouns and

modifiers, compound nouns, etc., verbal complexes (e.g. auxiliaries and modals in

conjunction with infinitives or participle forms), and coordinate structures. This level

of analysis is to be seen in the later ‘Georgetown system’. Complete syntactic analysis

involves the identification of relationships among phrases and clauses within

sentences.

Syntactic analysis aims to identify three basic types of information about sentence

structure:

1) The sequence of grammatical elements, e.g. sequences of word classes:

Art(icle) + N(oun) + V(erb) + Prep(osition)..., or of functional elements: subject +

predicate. These are linear relations.

2) The grouping of grammatical elements, e.g. nominal phrases consisting of

nouns, articles, adjectives and other modifiers, prepositional phrases consisting of

prepositions and nominal phrases, etc. up to the sentence level. These are constituency

relations.

3) The recognition of dependency relations, e.g. the head noun determines the

form of its dependent adjectives in inflected languages such as French, German and

Russian. These are hierarchical (or dominance) relations. Included among the basic

objectives of any method of syntactic analysis must be at least the resolution of

homographs (by identification of grammatical categories, e.g. whether watch is a noun

or a verb), the identification of sequences or structures which can be handled as units

in SL-TL transfer, e.g. nouns and their associated adjectives.

Various models of syntactic structure and methods of parsing have been adopted in

MT systems and are described in more detail in connection with particular MT

Page 34: Course Paper on Machine Translation

projects. At this point, the main approaches will be outlined, illustrated in the most

part by analyses (whole or partial) of the sentence The gold watch and chain were

sold by the jeweler to a man with a red beard. This is a passive sentence (the

grammatical subject is the object of the verb), containing a homograph (watch); an

ambiguous coordinate structure (are both the watch and the chain modified by gold?)

and three prepositional phrases each of which could in theory modify the verb or their

preceding noun phrase. An example of an analysis program (parsing program) to

identify sequential (linear) information was the Predictive Syntactic Analyzer

developed at the National Bureau of Standards and at Harvard University. The premise

was that on the basis of an identified grammatical category (article, adjective, noun,

etc.) the following category or sequences of categories could be anticipated with an

empirically determinable measure of probability. The system had the following

characteristics: under the general control of a push-down store (i.e. last in first out) a

sentence was parsed one word at a time left to right, the action taken for each word

being determined by a set of predictions associated with the grammatical category to

which the word had been assigned. At the beginning of the analysis certain sentence

types were predicted in terms of sequences of grammatical categories. Examination of

each word was in two stages: first to test whether its category ‘fulfilled’ one of the

predictions, starting from the most probable one, then either to alter existing

predictions or to add further predictions. Formally, the system was an implementation

of a finite state grammar. The analysis of a sentence was completed if a terminal state

has been reached and all categories have been accounted for. Initially, only the single

most probable path through the series of predictions was taken during parsing, but in

later models all possible predictions were pursued. The method did not in principle

need to recognize phrase structures or dependency relations, although these could be

derived from the identification of specific category sequences.

1. Finite state grammar

The second approach, analysis of dependency relations, is based on the identification

of ‘governors’, e.g. the ‘head’ noun in a noun phrase, and their dependants or

‘modifiers’, e.g. adjectives. The governor of the sentence as a whole is generally taken

to be the finite verb since this specifies the number and nature of dependent nouns. A

Page 35: Course Paper on Machine Translation

verb such as buy, for example, can have four dependants (purchaser, object purchased,

price, seller) – a concept referred to as ‘valency’: a transitive verb such as see has a

valency of two, an intransitive such as go has a valency of one, etc. The gold watch

was sold by the jeweler to a man with a red beard

2. Dependency structure analysis

The parsing of dependency structure can operate either top-down (identification first

of governors and then dependants) or bottom-up (determination of governors by a

process of substitution). The top-down approach was most common, and can be

illustrated by Garvin’s fulcrum parser in a series of passes the algorithm identified first

the key elements of the sentence, e.g. main finite verb, subject and object nouns,

prepositional phrases, then the relationships between sentence components and finally

the structure of the sentence as a whole. The third approach, that of phrase structure

analysis, provides labels for constituent groups in sentences: noun phrase (NP), verb

phrase (VP), prepositional phrase (PP), etc. The phrase structure approach is

associated most closely in the early period of MT research with the MIT project.

Parsing can be either bottom-up or top-down. In the former, structures are built up in a

series of analyses from immediate constituents, e.g. first noun phrases, then

prepositional structures, then verb relationships and finally the sentence structure as a

whole. In top-down parsing, the algorithm seeks the fulfillment of expected

constituents NP, VP, etc. by appropriate sets and sequences of grammatical categories.

The bottom-up parsing strategy was the most common approach in the early MT

system, but at MIT some investigation was made into the top-down strategy (‘analysis

by syntheses). In systems since the mid-1960 this strategy is now probably more

common. The gold watch was sold by the jeweler to a man with a red beard

3. Phrase structure analysis

It may be noted that categorial grammar developed by Bar-Hillel which was one of the

first attempts at formal syntax, is a version of constituency grammar. In a categorial

grammar, there are just two fundamental categories, sentence s and nominal n; the

other grammatical categories (verb, adjective, adverb, etc.) are defined in terms of

their potentiality to combine with one another or with one of the fundamental

categories in constituent structures. Thus a transitive verb is defined as n\s because it

Page 36: Course Paper on Machine Translation

combines with a nominal (to its left) to form sentences; and an adjective is defined as

n/n because in combination with a nominal n to its right it forms a (higher order)

nominal n. In other words, the category symbols themselves define how they are to

combine with other categories. Combination operates by two simple 'cancellation'

rules: x/y, y > x, and y, y\x > x.

2. 6 Formal syntax and transformational grammar

Research in MT helped to stimulate much interest in formal linguistics. An early

result of this mathematization of syntax and linguistic theory was the demonstration

that all phrase structure and dependency grammars are formally (i.e. mathematically)

equivalent and that since they can be implemented on pushdown automata, they are

equivalent also to the so-called finite state grammars (Gross & Lentin 1967). All these

grammars belong to the class of ‘context-free’ grammars. A context-free grammar

consists of a set of rewriting rules (or production rules) of the form A > a, where A

belongs to a set of ‘non-terminal’ symbols and a is a string of non-terminal or terminal

symbols. Non-terminal symbols are grammatical categories (S, NP, VP, N, Adj, etc.)

and terminal symbols are lexical items of the language. Context-free grammars are

important not only as the basis for formal grammars of natural languages but as the

basis for computer programming, since the standard algorithmic methods used in

compilers rely on finding only context-free structures in programming languages.

However, Noam Chomsky demonstrated the inherent inadequacies of finite state

grammars, phrase structure grammars and the formally equivalent dependency

grammars for the representation and description of the syntax of natural languages.

Context-free grammars are unable, for example, to relate different structures having

the same functional relationships, e.g. where discontinuous constituents are involved:

He looked up the address and He looked the address up; or where there are

differences of voice, e.g. the active: The jeweler sold the watch to the man yesterday

and the passive: Yesterday the man was sold the watch by the jeweler. Chomsky

proposed a transformational-generative model which derived ‘surface’ phrase

structures from ‘deep’ phrase structures by transformational rules. Thus a passive

construction in a ‘surface’ representation is related to an underlying active

Page 37: Course Paper on Machine Translation

construction in a ‘deep’ representation, where the surface’ subject noun appears as the

‘deep’ logical object. Deep structures are generated from an initial symbol S by

‘context-sensitive’ rewriting rules. An essential feature of the Chomskyan model is

that syntactic structures are generated top-down from initial symbol S, to ‘deep’

structure tree and then by transformational rules to ‘surface’ structure trees. In the case

of a coordinate phrase such as gold watch and chain the base ‘deep’ structure would

make explicit the fact that both watch and chain are gold. To produce the elliptical

‘surface’ form a transformation rule would delete the repeated adjective. The model is

not intended to provide the basis for a recognition grammar (e.g. a parser), but only to

define mathematically the set of well-formed sentences, and to assign “a structural

description indicating how the sentence is understood by the ideal speaker-

hearer” (Chomsky 1965: 5). The implications of this approach became clearer when

researchers attempted to develop ‘transformational parsers’

1. ‘Deep’ structure analysis

The gold watch and gold chain the gold watch and chain

2. Transformational rule (loss of phrase structure relationship)

Chomsky’s notion of transformational rules derived formally from the work of Zellig

Harris (1957). Harris' concern was the development of a symbolism for representing

structural relationships. Grammatical categories were established primarily on the

basis of distributional analysis. Thus, the subject of a sentence can be a (single) noun

(The man...), a clause (His leaving home...), and a gerundive (The barking of dogs...),

and an infinitive clause (To go there...), etc. In order to function as subjects, clauses

have to undergo transformations from 'kernel' (atomic sentence-like) forms: e.g. He

left home > His leaving home, Dogs bark > the barking of dogs. For Harris,

transformations were a descriptive mechanism for relating surface structures, while in

Chomsky’s model; transformational rules derive surface structures from ‘deeper’

structures. By the mid-1960's an additional requirement of transformational rules was

that they should be ‘meaning-preserving’, i.e. from a ‘deep’ structure should be

generated semantically equivalent surface structures. Although Chomsky’s syntactic

theory has undoubtedly had most influence, the formalization of transformations by

Page 38: Course Paper on Machine Translation

Harris had considerable impact in MT research, particularly in the representation of

SL-TL structural transfer rules.

2. 7: Syntactic ambiguity and discourse relations.

Although the identification of grammatical categories and of sentence structures

is clearly important in linguistic analysis, there are inherent limitations in syntactic

analysis that were recognized before even efficient parsers had been developed. A

familiar example is the problem of multiple analyses of prepositional phrases.

Syntactic analysis alone cannot decide which relationship is correct in a particular

case. For example take the sentences: The coastguard observed the yacht in the

harbor with binoculars. The gold watch was sold by the jeweler to a man with a

beard. In the first case, it was the coastguard who had the binoculars; therefore the PP

with the binoculars modifies the verb. But in the second case, the PP with a beard

modifies the preceding noun man. Only semantic information can assist the analysis

by assigning semantic codes allowing binoculars as ‘instruments’ to be associated

with 'perceptual' verbs such as observe but prohibiting beards to be associated with

objects of verbs such as sell. Such solutions have been applied in many MT systems

since the mid 1960's (as the following descriptions of systems will show). However,

semantic features cannot deal with all problems of syntactical ambiguity. As Bar-

Hillel argued in 1960 (Bar-Hillel 1964), human translators frequently use background

knowledge to resolve syntactical ambiguities. His example was the phrase slow

neutrons and protons. Whether slow modifies protons as well as neutron, which can

be decided only with subject knowledge of the physics involved. Similarly, in the case

of the gold watch and chain our assumption that both objects are gold is based on past

experience. On the other hand, in the case of the phrase old men and women the

decision would probably rest on information conveyed in previous or following

sentences in the particular text being analyzed. The most frequent occasions on which

recourse is made to ‘real world’ knowledge involve the reference of pronouns.

Examples are the two sentence pairs: The men murdered the women. They were

caught three days later. The men murdered the women. They were buried three days

later. The correct attribution of the pronoun they to the men in the first pair and to the

Page 39: Course Paper on Machine Translation

women in the second depends entirely on our knowledge that only dead people are

buried, that murder implies death, that murder is a criminal act, and that criminals

ought to be apprehended. This knowledge is non-linguistic, but it has linguistic

implications in, for example, translation of these sentences into French where a choice

of ils or elles must be made. Of course, it is not only in cases of syntactic ambiguity

that we use ‘real world’ knowledge to help in understanding text. Homographs can, as

indicated earlier, be resolved by identification of grammatical categories, e.g. whether

watch is a noun or a verb. However, the resolution of some homographs require, as in

the physics example, knowledge of the objects referred to. There is, for example, a

third sense of watch in the sentence: The watch included two new recruits that night.

It can be distinguished from the other noun only by recognition that timepieces do not

usually include animate beings. It was from such instances that Bar-Hillel was to argue

in an influential paper (Bar-Hillel (1960) that fully automatic translation of a high

quality was never going to be feasible. In practice this type of problem can be lessened

if texts for translation are restricted to a more or less narrow scientific field, and so

dictionaries and grammars can concentrate on a specific ‘sub-language’ (and this was

the argument for ‘micro-glossaries’). Nevertheless, similar examples recur regularly,

and the argument that MT requires ‘language understanding’ based on encyclopedic

knowledge and complicated inference procedures has convinced many researchers that

the only way forward is the development of ‘interactive’ and Artificial Intelligence

approaches to MT.

In general, semantic analysis has developed, by and large, as an adjunct of

syntactic analysis in MT systems. (Exceptions are those MT systems with an explicitly

semantic orientation) In most MT systems semantic analysis goes no further than

necessary for the resolution of homographs. In such cases, all that is generally needed

is the assignment of such features as ‘human’, ‘animate’, ‘concrete’, ‘male’, etc. and

some simple feature matching procedures. For example, crook can only be animate in

The crook escaped from the police, because the verb escape demands an animate

subject noun. The ‘shepherd’s staff’ sense of crook is thus excluded. In many

systems semantic features have been assigned as 'selection restrictions' in an ad hoc

manner, as the demands of the analysis of a particular group of lexical items seem to

Page 40: Course Paper on Machine Translation

require them, and also somewhat too rigidly. There are difficulties, for example, if the

verb sell is defined as always having inanimate objects; the sentence. The men were

sold at a slave market would not be correctly parsed. One answer suggested has been

to make such 'selection restrictions' define not obligatory features but preferences.

True semantic analysis should include some decomposition of lexical items according

a set of semantic ‘primitives’ or putative ‘universals’. Only by such means is it

possible to derive common semantic representations for a pair of sentences such as

The teacher paid no attention to the pupil and The pupil was ignored by the teacher.

In general, the majority of MT systems have avoided or held back from the intricacies

and complexities and no doubt pitfalls of this kind of semantics. It is found therefore

only in those MT groups that have investigated interlinguas, and in some of those

recent groups with an interest in AI methods.

2. 8: Sentences and texts

The difficulties with pronominal reference described above stem also from the

exclusive concentration of syntax-based analysis on sentences. The need for text-based

analysis can be illustrated by the following two German sentences: In der Strasse

sahen wir einen Polizist, der einem Mann nachlief. Dem Polizist folgte ein grosser

Hund.

Translation into English sentence by sentence would normally retain the active verb

forms producing: In the street we saw a policeman running after a man. A large dog

followed the policeman. Text cohesion would be improved if the second sentence

were passivized as: The policeman was followed by a large dog. This inversion

requires that a MT system adheres as far as possible to the information structure of the

original, i.e. in this case retains the ‘policeman’ as the head (or topic) of the sentence.

The problems of topicalisation and text cohesion are of course far more complex than

this example. Scarcely any MT projects have even considered how they might be

tackled.

2. 9: Transfer and synthesis.

Page 41: Course Paper on Machine Translation

The production of output text in the target language (TL) is based on the

information provided from dictionaries and from the results of analysis. In general the

synthesis of TL sentences is less complex than the analysis of SL input. The process

involves nearly always the derivation of correct morphological forms for TL words

(unless dictionaries contain only full TL forms). Thus, for example, TL synthesis must

produce the right forms of verbs, e.g. for English simple past forms it is not a matter of

just adding -ed as in picked (from pick), since sometimes endings must be deleted or

altered as in lived (not: liveed) and tried (not: tryed), etc. Irregular forms are generally

handled by the dictionary (e.g. went would be coded directly as the past form of go). If

analysis has included the establishment of syntactic structure (e.g. a phrase structure)

then synthesis must convert this structure into an appropriate TL structure and produce

a linear representation, i.e. it must invert the analysis process in some way. However,

it should be stressed that inversion does not imply that the rules devised for the

analysis of structures for a particular language (as SL) can be simply reversed to

obtain rules for synthesis of that language (as TL). At some point in many systems (the

exceptions being interlingual systems, cf. next section), the syntactic structures of SL

texts are transformed into TL structures. Whether such transformations apply to only

short segments (as in word-for-word systems) or to whole sentences, the process

involves the specification of transformation rules. For example, a rule for changing a

German construction with final past participle (Er hat das Buch gestern gelesen) into

an English construction with a simple past form (He read the book yesterday).

Clearly, such transformation rules have much in common with the transformation rules

that Harris devised for relating structures within the same language.

2. 10: System designs and strategies

In broad terms, there have been three types of overall strategy adopted in MT

systems. SL Analysis and Synthesis TL text; text SL-TL dictionaries and grammars

1. ‘Direct translation’ system

The first approach is the ‘direct translation’ approach. Systems are designed in all

details specifically for one particular pair of languages. The basic assumption is that

the vocabulary and syntax of SL texts need not be analyzed any more than strictly

Page 42: Course Paper on Machine Translation

necessary for the resolution of ambiguities, the correct identification of appropriate TL

expressions and the specification of TL word order. Thus if the sequence of SL words

is sufficiently close to an acceptable sequence of TL words, then there is no need to

identify the syntactic structure of the SL text. The majority of MT systems of the

1950’s and 1960’s were based on this approach. They differed in the amount of

analysis and/or restructuring incorporated. There was none at all in the straight

‘dictionary translation’ experiment of Richens and Booth; there was just a minimum of

local restructuring in the ‘word-for-word’ systems of the University of Washington

and IBM; there was partial analysis of SL structure in the Georgetown system; and

there was full sentence analysis in the systems at Ramo-Wooldridge, Harvard, and

Wayne State University primary characteristic of ‘direct translation’ systems of the

earlier period was that no clear distinctions were made between stages of SL analysis

and TL synthesis (cf. particularly the account of the Georgetown system below). In

more recent (post-1970) examples of ‘direct’ systems there is a greater degree of

‘modular’ structure in the systems.

SL Analysis Interlingual Synthesis TL text representation text SL TL dictionaries SL-

TL dictionaries and grammars dictionary and grammars

2. ‘Interlingual’ system

The second basic MT strategy is the ‘interlingual’ approach, which assumes that it is

possible to convert SL texts into semantico-syntactic representations common to more

than one language. From such ‘interlingual’ representations texts would be generated

into other languages. In such systems translation from SL to TL is in two distinct and

independent stages: in the first stage SL texts are fully analyzed into interlingual

representations, and in the second stage interlingual forms are the sources for

producing (synthesizing) TL texts. Procedures for SL analysis are intended to be SL-

specific and not devised for any particular TL in the system; likewise, TL synthesis is

intended to be TL-specific. Interlingual systems differ in their conceptions of an

interlingual language: a ‘logical’ artificial language, or a ‘natural’ auxiliary language

such as Esperanto; a set of semantic primitives common to all languages, or a

‘universal’ vocabulary, etc. Interlingual MT projects have also differed according to

the emphasis on lexical (semantic) aspects and on syntactic aspects. Some

Page 43: Course Paper on Machine Translation

concentrated on the construction of interlingual lexica (e.g. the Cambridge and the

Leningrad groups); others have concentrated on interlingual ‘syntax’ (e.g. the

Grenoble and Texas groups). SL Analysis Transfer Synthesis TL text SL repr TL repr

text SL SL-TL TL dictionaries dictionary dictionaries and grammars and grammars

Transfer rules

3. ‘Transfer’ system

The third approach to overall MT strategy is the ‘transfer’ approach. Rather than

operating in two stages through a single interlingual representation, there are three

stages involving underlying representations for both SL and TL texts; i.e. the first

stage converts SL texts into SL ‘transfer’ representations, the second converts these

into TL ‘transfer’ representations, and the third produces from these the final TL text

forms. Whereas the interlingual approach necessarily requires complete resolution of

all ambiguities and anomalies of SL texts so that translation should be possible into

any other language, in the ‘transfer’ approach only those ambiguities inherent in the

language in question are tackled. Differences between languages of the

know-savoir/connaitre type would be handled during transfer. In English analysis,

know is treated as unambiguous, there is no need to determine which kind of

‘knowing’ is involved. Whereas the ‘interlingual’ approach would require such

analysis, the ‘transfer’ approach does not; problems of mismatch between SL and TL

lexical ranges are resolved in the transfer component. Systems differ according to the

‘depth’ of analysis and the abstractness of SL and TL transfer representations. In the

earliest systems analysis went no further than ‘surface’ syntactic structures, with

therefore structural transfer taking place at this depth of abstraction. Later (post-1970)

transfer systems have taken analysis to ‘deep’ semantico-syntactic structures (of

various kinds), with correspondingly more abstract transfer representations and

transfer rules.

The basic difference between these two ‘indirect’ approaches and the (generally

earlier) ‘direct’ approach lies in the configuration of dictionary and grammar data. In

‘direct’ systems the main component is a single SL-TL bilingual dictionary

incorporating not only information on lexical equivalents but also all data necessary

for morphological and syntactic analysis, transfer and synthesis. In ‘indirect’ systems,

Page 44: Course Paper on Machine Translation

this information is dispersed among separate SL and TL dictionaries, separate SL and

TL grammars, and either the interlingua vocabulary and syntax, or the SL-TL

‘transfer’ dictionary (of lexical equivalences) and a ‘grammar’ of SL-TL structure

transfer rules.

2. 11: Perspectives and influences.

While the classification of MT systems in terms of basic strategy is a convenient

descriptive device and will be employed in the grouping of system descriptions in later

chapters, it has not been the most prominent perspective for MT researchers,

particularly in the 1950’s and 1960’s. For this period, the most important distinctions

were between the engineering and the ‘perfectionist’ approaches, between the

empiricist and other methodologies, and between the syntax orientation and various

lexical and word-centered approaches. The most immediate point of dispute was

between those groups who agreed with Dostert and Booth on the importance of

developing operational systems as quickly as possible (ch.2.4.3) and those who argued

for more fundamental research before such attempts. The engineering approach held

basically that all systems can be improved and that the poor quality early word-for-

word systems represent a good starting point. There were differences between what

Garvin (1967) dubbed the ‘brute force’ approach, which assumed that the basic need

was larger storage capacity (e.g. the IBM solution, ch.4.2), and the engineering

approach which believed that algorithmic improvements based on reliable methods of

(linguistic) analysis could lead to better quality. The ‘perfectionists’ included all those

groups which concentrated on basic linguistic research with ‘high quality’ systems as

the objective. The latter differed considerably in both theories and methods. Disputes

recurred frequently between the ‘perfectionists’ and the ‘engineers’ until the mid-

1960. On questions of methodology the main point of difference concerned the

‘empiricist’ approach, exemplified by the RAND group. The approach emphasized the

need to base procedures on actual linguistic data; it was distrustful of existing

grammars and dictionaries; it believed it was necessary to establish from scratch the

data required and to use the computer as an aid for gathering data. The approach

stressed statistical and distributional analyses of texts, and a ‘cyclic’ method of system

Page 45: Course Paper on Machine Translation

development: i.e. routines devised for one corpus were tested on another, improved,

tested on a third corpus, improved again, and so forth. The empirical approach was in

fact fully in accord with the dominant linguistic methodology of the 1940’s and 1950’s

in the United States, the descriptivist and structuralist ‘tradition’ associated

particularly with Leonard Bloomfield (1933). The descriptivists adopted the

behaviorist and positivistic method that insisted that only interpersonally observed

phenomena should be considered ‘scientific’ data, and which rejected introspections

and intuitions. They distrusted theorizing, stressed data collection, and concentrated on

methods of discovery and analysis. Traditional grammars were suspect: Charles Fries

(1952), for example, undertook a distributional analysis of telephone conversations

that resulted in new grammatical categories for English. Most descriptivists worked,

however, on phonetics and phonology. Only in the mid-1950 did some descriptivist

such as Zellig Harris start work on syntax. It was therefore not surprising that the

empiricists regarded their research within MT as extending the range of descriptive

linguistics.

The ‘empiricist’ emphasis on probabilistic and statistical methods, however, has

perhaps a different origin. It is likely to be the considerable influence of the statistical

theory of communication associated with Claude Shannon, i.e. ‘information theory’,

and to which Warren Weaver made a substantial contribution. The theory had great

impact on the anti-metaphysical inclinations of most American linguists, since it

seemed to provide a basis for developing mechanical methods for ‘discovering’

grammars. It may be noted that when Yngve first presented his ideas on ‘syntactic

transfer’ Yngve (1957), he related his tripartite model to the information-theoretic

triple of sender, channel and receiver.

A third area of difference in early MT groups was the question of what should be

taken as the central unit of language. The majority assumed the centrality of the

sentence; their approach was sentence-oriented (as was and still is, in essence, that of

most linguists and logicians), and so there was an emphasis on syntactic relations and

problems. A minority upheld the centrality of the word. They emphasized lexical and

semantic relations and problems. They included the ‘lexicographic’ approach of

Reifler and King, the ‘thesaural’ approach of the Cambridge group, the ‘word-

Page 46: Course Paper on Machine Translation

centered’ theories of Lamb at Berkeley, and the dictionary-centered aspects of

Mel’chuk’s ‘meaning-text’ approach. It should be stressed that these are ‘differences’

only of general orientation; the ‘syntax-oriented’ groups did not neglect lexical and

semantic issues, and the ‘lexis oriented’ groups did not by any means neglect syntax.

Indeed, in the case of Lamb and Mel’chuk it is very much an open question whether

their models can be said to be oriented one way or the other.

In the descriptions above of various aspects of MT system design and methods

of analysis, it may well have been implied, at a number of points, that language

systems are intrinsically multileveled; that is to say, that linguistic description is

necessarily couched in terms of phonetics, phonology, morphology (word formation),

syntax, and semantics; and furthermore, that analysis proceeds through each of these

levels in turn: first morphological analysis, then syntactic analysis, then semantic

analysis. (Lamb and Mel’chuk in fact developed the most extensive ‘stratification list’

models within the MT context.) Although undoubtedly a ‘stratal’ view of language

systems is dominant in linguistics and has been since the time of Saussure, the founder

of modern (structuralism) linguistics, it has not been the conception of some MT

project teams. Indeed, many (particularly in the earliest period) would have rejected

such a stratal view of language both for being too rigid and for not conforming to

reality. For them, all aspects of language (lexical, semantic, and structural) interact

inextricably in all linguistic phenomena. There is no doubt that among the most

linguistics-oriented MT groups there has been sometimes an excessive rigidity in the

application of the stratal approach to analysis (e.g. in parsing systems); and it has led

to failures of various kinds. Nevertheless, the basic validity of the approach has not

been disproved, and most modern (linguistics-oriented) MT systems retain this basic

conception.

Conclusion:

Computer programs are producing translations - not perfect translations, for

that is an ideal to which no human translator can aspire. Machine Translation is not

primarily an area of abstract intellectual inquiry but the application of computer and

Page 47: Course Paper on Machine Translation

language sciences to the development of systems answering practical needs. When

scientist started looking at the Machine Translation component, they didn’t really

know how to go about evaluating its performance. Not having much past research to

go by, they began by checking the translations. Once researchers realized this

approach wouldn’t work, they translated the documents themselves and checked the

human translations against the Machine Translation versions. This allowed them to

compile a list of the most common types of errors that occurred during automatic

translation process which are word order, context, pronoun, dictionary, missing word,

extra word, proper name and many more. Machine Translation aims primarily at

comprehension and not at the production of a perfect Target Text, it is important to

follow two basic rules in order to make the best use of programs. First, we need to

recognize that certain types of texts, such as poetry, for example, are not suitable for

Machine Translation. Second, it is essential to correct the Source Text, as even one

letter can radically change meaning.

Page 48: Course Paper on Machine Translation

III. DIFFICULTIES IN MACHINE TRANSLATION

Machine Translations are mostly due to various types of ambiguity, concerning

polysemy of words, phrase attachment, coordination, anaphoric reference, scope of

logical and modal operators, and so on. Unknown words and phrases are another

major source of difficulty. Translation accuracy is expected to drastically improve if

the input documents are marked up with appropriate tags which resolve such

ambiguities or supply missing information.

3.1 Difficulties in translation.

One difficulty in translation stems from the fact that most words have multiple

meanings. Whether a human or a computer does a translation, meaning cannot be

ignored. A word with sharply differing meanings has several different translations,

depending on how the word is being used. The word 'bank' is often given as an

example of a homograph, that is, a word entirely distinct from another that happens to

be spelled the same. But further investigation shows that historically the financial and

river meanings of 'bank' are related. They both come from the notion of a "raised

shelf or ridge of ground". The financial sense evolved from the moneychanger's table

or shelf, which was originally placed on a mound of dirt. Later the same word came

to represent the institution that takes care of money for people. The river meaning has

remained more closely tied to the original meaning of the word. Even though there is

an historical connection between the two meanings of 'bank,' we do not expect their

translation into another language to be the same, and it usually will not be the same.

Page 49: Course Paper on Machine Translation

This example further demonstrates the need to take account of meaning in translation.

A human will easily distinguish between the two uses of 'bank' and simply needs to

learn how each meaning is translated. To see their importance to translation, we do

not expect the two meanings of 'bank' to have the same translation in another

language. Each language follows its own path in the development of meanings of

words. As a result, we end up with a mismatch between languages, and a word in one

language can be translated several different ways, depending on the situation. With

the extreme examples given so far, a human will easily sense that multiple

translations are probably involved, even if a computer would have difficulty. What

causes trouble in translation for humans is that even subtle differences in meaning

may result in different translations. A human can learn the distinctions o meanings

through substantial effort. It is not clear how to tell a computer how to make them.

Being a native or near-native speaker involves more than just memorizing lots

of facts about words. It includes having an understanding of the culture that is mixed

with the language. It also includes an ability to deal with new situations

appropriately. No dictionary can contain all the solutions since the problem is always

changing as people use words in usual ways. These usual uses of words happen all

the time. Some only last for the life of a conversation or an editorial. Others catch on

and become part of the language. Some native speakers develop a tremendous skill in

dealing with the subtleties of translation. However, no computer is a native speaker of

a human language. All computers start out with their own language and are 'taught'

human language later on. They never truly know it the way a human native speaker

knows a language with its many levels and intricacies. Does this mean that if we

taught a computer a human language starting the instant it came off the assembly line,

it could learn it perfectly? Computers do not learn in the same way we do. We could

say that computers can't translate like humans because they do not learn like humans.

Then we still have to explain why computers don't learn like humans. What is

missing in a computer that is present in a human? Building on the examples given so

far, there are three types of difficulty in translation that are intended to provide some

further insight into what capabilities a computer would need in order to deal with

human language the way humans do.

Page 50: Course Paper on Machine Translation

Human Translation Process

1) Distinguishing between general vocabulary and specialized terms.

The first type of translation difficulty is the most easily resolved. It is the case

where a word can be either a word of general vocabulary or a specialized term.

Consider the word 'bus.' When this word is used as an item of general vocabulary, it

is understood by all native speakers of English to refer to a roadway vehicle for

transporting groups of people. However, it can also be used as an item of specialized

terminology. Specialized terminology is divided into areas of knowledge called

domains. In the domain of computers, the term 'bus' refers to a component of a

computer that has several slots into which cards can be placed .One card may control

a CD-ROM drive. Another may contain a fax/modem. If you turn off the power to

your desktop computer and open it up, you can probably see the 'bus' for yourself. As

always, there is a connection between the new meaning and the old. The new

meaning involves carrying cards while the old one involves carrying people. In this

case, the new meaning has not superseded the old one. They both continue to be used,

but it would be dangerous, as we have already shown with several examples, to

assume that both meanings will be translated the same way in another language. The

way to overcome this difficulty, either for a human or for a computer, is to recognize

whether we are using the word as an item of general vocabulary or as a specialized

term. Humans have an amazing ability to distinguish between general and specialized

uses of a word. Once it has been detected that a word is being used as a specialized

term in a particular domain, and then it is often merely a matter of consulting a

Source Text

Analysis Meaning Synthesis

Target Text

Page 51: Course Paper on Machine Translation

terminology database for that domain to find the standard translation of that term in

that domain. It is common for a translator to spend a third of the time needed to

produce a translation on the task of finding translations for terms that do not yet

appear in the terminology database being used. Where computers shine is in

retrieving information about terms. They have a much better memory than humans.

But computers are very bad at deciding which the best translation to store in the

database is. This failing of computers confirms our claim that they are not native

speakers of any human language in that they are unable to deal appropriately with

new situations. When the source text is restricted to one particular domain, such as

computers, it has been quite effective to program a machine translation system to

consult first a terminology database corresponding to the domain of the source text

and only consult a general dictionary for words that are not used in that domain. Of

course, this approach does have pitfalls. Suppose a text describes a very sophisticated

public transportation vehicle that includes as standard equipment a computer. A text

that describes the use of this computer may contain the word 'bus' used sometimes as

general vocabulary and sometimes as a specialized term. A human translator would

normally have no trouble keeping the two uses of 'bus' straight, but a typical machine

translation system would be hopelessly confused. This first type of difficulty is the

task of distinguishing between a use of a word as a specialized term and its use as a

word of general vocabulary. One might think that if that distinction can be made, we

are home free and the computer can produce an acceptable translation.

Fully Automated Translation

2) Distinguishing between various meanings of a word of general vocabulary.

Text Formatting

Dictionary Search

Source Text

Target Text

Analysis Transfer Synthesis

Page 52: Course Paper on Machine Translation

The second type of difficulty is distinguishing between various uses of a word

of general vocabulary. It is essential to distinguish between various general uses of a

word in order to choose an appropriate translation. Already in 1960, an early machine

translation researcher named Bar-Hillel provided a now classic example of the

difficulty of machine translation. He gave the seemingly simple sentence "The box is

in the pen". He pointed out that to decide whether the sentence is talking about a

writing instrument pen or a child's play pen, it would be necessary for a computer to

know about the relative sizes of objects in the real world. Of course, this two-way

choice, as difficult as it is for a human, is a simplification of the problem, since 'pen'

can have other meanings, such as a short form for 'penitentiary' or another name for

a female swan. But restricting ourselves to just the writing instrument and play pen

meanings, only an unusual size of box or writing instrument would allow an

interpretation of 'pen' as other than an enclosure where a child plays. The related

sentence, "the pen is in the box", is more ambiguous. One would assume that the

pen is a writing instrument, unless the context is about unpacking a new playpen or

packing up all the furniture in a room. The point is that accurate translation requires

an understanding of the text, which includes an understanding of the situation and an

enormous variety of facts about the world in which we live. For example, even if one

can determine that, in a given situation, 'pen' is used as a writing instrument; the

translation into Spanish varies depending on the Spanish-speaking country.

3) Taking into account the total context, including the intended audience and

important details such as regionalisms.

The third type of difficulty is the need to be sensitive to total context,

including the intended audience of the translation. Meaning is not some abstract

object that is independent of people and culture. A serious example of insensitivity to

the total context and the audience is the translation of the English expression 'thank

you' which is problematical going into Japanese. There are several translations that

are not interchangeable and depend on factors such as whether the person being

thanked was obligated to perform the service and how much effort was involved. In

English, we make various distinctions, such as 'thanks a million' and 'what a

Page 53: Course Paper on Machine Translation

friend', but these distinctions are not stylized as in Japanese nor do they necessarily

have the same boundaries. A human can learn these distinctions through substantial

effort. It is not clear how to tell a computer how to make them. Languages are

certainly influenced by the culture they are part of. The variety of thanking words in

Japanese is a reflection of the stylized intricacy of the politeness in their culture as

observed by Westerners.

Machine Translation (MT) can be defined as a translation where the initiative

is with a computer system, either autonomously (Fully Automatic High Quality

Translation). Machine Aided Translation (MAT) is human translation supported by a

computer system. Support is available by lexical data, grammatical help, translation

memory, domain information and organizational support.

3.2 Machine translation ambiguity.

What makes machine translation so difficult? Part of the problem is that

language is highly ambiguous when looked at as individual words. For example,

consider the word "cut" without knowing what sentence the word came from. It could

have been any of the following sentences:

a) He told me to cut off a piece of cheese.

b) The child cut out a bad spot from the apple.

c) My son cut out early from school again.

d) The old man cut in line without knowing it.

e) The cut became infected because it was not bandaged.

f) Cut it out! You're driving me crazy.

If a computer is only allowed to the word "cut" and the rest of the sentence is

covered up, it is impossible to know which meaning of "cut" is intended. This may

not matter if everything stays in English, but when the sentence is translated into

another language, it is unlikely that the various meanings of "cut" will all be

translated the same way. This phenomenon is called "asymmetry". Illustrating an

Page 54: Course Paper on Machine Translation

asymmetry between English and French of the word "bank" the principal translation

of the French word banque (a financial institution) is the English word "bank." If

banque and "bank" were symmetrical then "bank" would always translate back into

French as banque. However, this is not the case. "Bank" can also translate into French

as rive, when it refers to the edge of a river. Now we may object that this is unfair

because the meaning of "bank" was allowed to shift. But a computer does not deal

with meaning, it deals with sequences of letters, and both meanings, the financial

institution one and the edge of a river one, consist of the same four letters, even

though they are different words in French. Thus English and French are

asymmetrical.

Early researchers in machine translation were already aware of the problem of

asymmetry between languages, but they seriously underestimated the difficulty of

overcoming it. They assumed that by giving the computer access to a few words of

context on either side of the word in question the computer could figure out which

meaning was intended and then translate it properly. But later some researchers had

realized that even if the entire sentence is available, it is still not always obvious how

to translate without using knowledge about the real world. A classic sentence that

illustrates this difficulty uses the word "pen", which can refer to either a writing

instrument or to an enclosure in which a child is placed to play so that it will not

crawl off into another room. The ambiguity must be resolved or the word "pen" will

probably be translated incorrectly.

- The pen was in the box.

This sentence will typically be interpreted by a human as referring to a writing

instrument inside a cardboard box, such as a gift box for a nice fountain pen or gold-

plated ballpoint pen, rather than a play pen in a big box. However, look what happens

if the sentence is rearranged as follows:

- The box was in the pen.

Page 55: Course Paper on Machine Translation

This sentence will typically be interpreted by a human as referring to a normal-size

cardboard box inside a child's play pen rather than as a tiny box inside a writing

instrument. A human uses knowledge about typical and relative sizes of objects in the

real world to interpret sentences. For a human, this process is nearly effortless and

usually unconscious. For a computer that does not have access to real-world

knowledge, this process is impossible. The situation is also taken into account.

Returning to the sentence about the pen in the box, there are texts, such as a

description of a family with small children moving their affairs to another apartment,

in which a human would interpret the pen as the child's play pen being put into a

large box to protect it while it is moved to a new location. And there are texts, such as

a spy story about ultra-miniature boxes of top-secret information, in which the

sentence about the box in the pen would be interpreted as referring to a writing

instrument containing a tiny box. The words in these sentences do not change, yet the

interpretation changes. Here even real-world knowledge is insufficient. Only some

sense of the flow of discourse and the current situation are needed.

3.3 Problems of Machine Translation

Translating may be defined as the process of transforming signs or

representations into other signs or representations. If the originals have some

significance, we generally require that there images also have the same significance,

Page 56: Course Paper on Machine Translation

or, more realistically, as nearly the same significance as we can get. Keeping

significance invariant is the central problem in translating between natural

languages.

Some typical factors contribute to the difficulty of Machine Translation which

is: words with multiple meanings, sentences with multiple grammatical structures,

uncertainty about what a pronoun refers and other problems of grammar. Two most

important misunderstandings make translation seem simpler than it is: first of all

translation is not primarily a linguistic operation: "The police refused the students a

premise because they feared violence". This sentence is to be translated in the French

but "police" is feminine; "they" will also have to be feminine; and secondly

translation is not an operation that preserves meaning. Different languages have

different usage, for example there are languages like French in which pronouns must

show numbers and gender, Japanese where pronouns are often omitted altogether,

Russian where there are no articles, Chinese where nouns do not differentiate singular

and plural nor verbs present and past. That is why, the most important problem in

Automatic and Machine translation is:

a) Polysemy: a word which has several similar meanings, sometimes the proper

translation is difficult to find even for a human translator. For example the word

"fair" might mean any of "beautiful", "light", "blond", "free from bias", etc.

b) Homonymy: are considered to be several independent words which share the

same linguistic corpse, they are difficult to be translated, because the translation of

homonyms often depends on context and semantics. For example, the word "reif"

might mean "ring", "bracelet", or "white frost". As well as the word "screen"

might mean "schirm", "leinwand", "raster", or "abschirmung".

c) Syntactical ambiguities: structure of sentences not only depends on types of

words but often also on semantics. "Flying planes can be dangerous" in this case

the sentence is ambiguous because words can be grouped into two ways: "(flying

planes) can be dangerous" and "(flying) (planes) can be dangerous".

d) Referential ambiguity: pronouns refer to certain words but it's often not obvious

to which, references might even cross sentence boundaries. Reference resolution

Page 57: Course Paper on Machine Translation

is considered to be one particular area of language, in which ambiguity is often

problematic for computers. For example, it would be useful if an MT system could

somehow distinguish between the various meaning of "it" in the three sentences:

The monkey ate banana because it was hungry.

The monkey ate banana because it was ripe.

The monkey ate banana because it was teatime.

Native speakers will normally identify the intended meaning of such language very

easily. Indeed, it would probably not even enter their heads that alternative

interpretations are possible. Non-native speakers of a language, when presented with

the text, will often have to narrow down the possible meaning in a more conscious

way. An MT system on the other hand will not only find it equally difficult to decide

between several sensible interpretations of a given sentence, but will also have no

way of distinguishing between sensible and absurd interpretations of a given sentence

or text.e) e) Fuzzy hedges: these are vague words and expressions which are very

difficult to be translated. For example, "in a sense", "irgendwie", "very", "in a sense

machine translation works nowadays"

f) Metaphors and symbols: on the underlying culture and history, they often cannot

be translated (Chinese sayings sometimes just do not make sense), in this very

situation just idiomatic dictionaries may be used to ease translation. For example

"Mit eiserner Miene feuerte er seinen treuesten Mitarbeiter" corresponding English

idiom is "with a stony expression".

g) New developments: all languages of the world are dynamic, always new words

are created, as well as proper names of new technologies. For example, "secure

shell", "telnet".

h) Synonyms: there are always several words with the same meaning and for

computer it is difficult to choose the right one because it depends on context, style

and semantics.

Page 58: Course Paper on Machine Translation

3.4 Cognitive processes

To understand the essential principles underlying machine translation it is

necessary to understand the functioning of the human brain. The first stage in human

translation is complete comprehension of the source language text. This

comprehension operates on several levels:

1. Semantic level: understanding words out of context, as in a dictionary.

2. Syntactic level: understanding words in a sentence.

3. Pragmatic level: understanding words in situations and context.

Furthermore, there are at least five types of knowledge used in the translation

process:

a) Knowledge of the source language, which allows us to understand the original

text.

b) Knowledge of the target language, which makes it possible to produce a

coherent text in that language.

c) Knowledge of equivalents between the source and target languages.

d) Knowledge of the subject field as well as general knowledge, both of which

aid comprehension of the source language text.

e) Knowledge of socio-cultural aspects that is, of the customs and conventions of

the source and target cultures.

Given the complexity of the phenomena that underlie the work of a human

translator, it would be absurd to claim that a machine could produce a target text of

the same quality as that of a human being. However, it is clear that even a human

translator is seldom capable of producing a polished translation at first attempt. In

reality the translation process comprises two stages: first, the production of a rough

text or preliminary version in the target language, in which most of the translation

problems are solved but which is far from being perfect; and second, the revision

stage, varying from merely re-reading the text while making minor adjustments to the

implementation of radical changes. It could therefore be said that machine translation

Page 59: Course Paper on Machine Translation

aims at performing the first stage of this process in an automatic way, so that the

human translator can then proceed directly to the second, carrying out the meticulous

and demanding task of revision. The problem is that the translator now faces a text

that has not been translated by a human brain but by a machine, which changes the

required approach because the errors are different. It becomes necessary to harmonize

the machine version with human thought processes, judgments and experiences.

Machine translation is thus both an aid and a trap for translators: an aid because it

completes the first stage of translation; a trap because it is not always easy for the

translator to keep the necessary critical distance from a text that, at least in a

rudimentary way, is already translated, so that mistakes may go undetected. In no

sense should a translation produced automatically be considered final, even if it

appears on the surface to be coherent and correct.

Conclusion:

From the given information we can find out that translation is not an easy

process. It had to be studied till we'll have a good or good enough translation. People

had to work hard to make the translation easier by using machines. Currently, Machine

Translation systems are already very helpful, but not perfect. There are linguistic

problems that cannot be satisfyingly solved by computers that can not think like

humans. Some of the difficulties in Machine Translation are mostly due to various

types of ambiguity, concerning polysemy of words, phrase/clause attachment,

coordination, anaphoric reference, scope of logical/modal operators, and so on. Maybe

in the future, further progress in Artificial Intelligence will help to solve the remaining

problems. Translation is a very difficult thing requiring much feeling and understanding

of cultural aspects, which is not available in a computer. Some grammatical structures

in a given language do not exist in another language, and that is why the translation

without interpretation, remains still an unsolved problem.

Page 60: Course Paper on Machine Translation

General conclusion

Machine translation and computer-assisted translation has long been a subject

of discussion. At the beginning of investigation some translators totally reject even

the idea of machine translation because they associate it with the point of view that

translation is merely one more marketable product based on a calculation of

investment versus profits. They define translation as an art that possesses its own

aesthetic criteria that have nothing to do with profit and loss, but are rather related to

creativity and the power of the imagination. This applies mostly, however, to specific

kinds of translation, such as that of literary texts, where polysemy, connotation and

style play a crucial role. It is clear that computers could not even begin to replace

human translators with such texts. Even with other kinds of texts, the analysis of the

roles and capabilities of Machine Translation which shows that it is not efficient and

accurate enough to eliminate the necessity for human translators. The first point to be

made is that MT is a translation method that focuses on the source language, while

human translation aims at comprehension of the target language. Machine

translations are therefore often inaccurate because they take the words from a

dictionary and follow the situational limitations set by the program designer. In fact,

translators should recognize and learn to exploit the potential of the new technologies

to help them to be more rigorous, consistent and productive without feeling

threatened.

Translation is a very difficult thing requiring much feeling and understanding

of cultural aspects, which is not available in a computer. The reason is that a

computer can't be able to think and interpret the environment (social and cultural

aspects) as a human being. Another reason is that some grammatical structures in the

Page 61: Course Paper on Machine Translation

source Language do not exist in the Target Language and computer doesn't know

which grammatical structure to follow. Translating with the help of the computer is

definitely not the same as working exclusively on paper and with paper products such

as conventional dictionaries, because computer tools provide us with a relationship to

the text which is much more flexible than a purely lineal reading. Furthermore, the

Internet with its universal access to information and instant communication between

users has created a physical and geographical freedom for translators that were

inconceivable in the past. Translators need to accept the new technologies and learn

how to use them to their maximum potential as a means to increased productivity and

quality improvement. Of course, the quality improvement of machine translation

(MT) is mainly the task of its developers. However, the users can also make some

efforts for reaching acceptable results because first of all the quality of machine

translation directly depends on the quality of the delivered source text.

Page 62: Course Paper on Machine Translation

Appendix 1

ENGLISH IDIOMS MACHINE TRANSLATION

RUSSIAN EQUIVALENT

As the crow fliesПо мере того как ворона летает

Прямиком

Ask for the moon/cryПопросите луна Желать невозможного

Blue-eyed boy Голубоглазый мальчик Любимчик Bow and scrape Приклонитесь и царапанье Пресмыкаться A country cousin Кузен страны Провинциал, бедный

родственник Carte blanche Карт-бланш Полная свобода действийClean as a whistle Уберите как свист Чистый, как стеклышкоCold comfort Слабое утешение Слабое утешениеCome to a sticky end Липкий конец приедьте в Попасть как кур в ощипComfortable in one’s skin Удобный в коже Чувствовать себя в своей

тарелке

Creature comfortsУдобства существа Все удобства

Crocodile tears Крокодиловы слезы Крокодиловы слезыCut to the chase Сокращение к

преследованиюПерейти к делу

Diamond in the rough (a) Алмаз в грубом Не(от)шлифованный Don’t lock the barn door after the horse is gone

Не захватите дверь сарая, после того, как лошадь ушла

После драке кулаками не машут

Down at heelВниз в пятке Потрепанный,

обшарпанныйDrag somebody’s name through the mire/mud

Тяните название сурьмы через болото/грязь

Запачкать ч-л имя

Eye for an eye, a tooth for a tooth

Глаз для глаз, зуб для зуб Зуб за зуб, око за око

Feeding frenzy (a) Подача безумия Погоня за сенсациейGet one’s back up рассердитесь Злиться

Page 63: Course Paper on Machine Translation

Get one’s brain in gear Получите мозг в механизме

Собраться с мыслями

Get one’s feet wet Получите влажные ноги ПопытатьсяGet sth down to a fine art Получите что-то вниз к

прекрасному искусству Довести до совершенства

Get the better of sb Возьмите верх над кем-то Взять верхGive the bum’s rush Дайте порыв задницы Выставить за дверьGo cold turkey Пойдите холодная индейка Завязывать Go on the wagon Продолжите фургон Завязать Go through thick and thin Пройдите толстый и

тонкийСъесть под соль

Gobbledygook Напыщенность речи Птичий языкGuinea pig Морская свинка Подопытный кролик Have (had) a good innings (имели) хорошую

возможность Прожить долгую и счастливую жизнь

Have one’s back to the wall Имейте оn's назад к стене Припереть к стенкеHen night Только для женщин ночь Девишник (an) Indian summer Индийское лето Бабье лето, золотая

осень, вторая молодостьKangaroo court Суд кенгуру Самозванный судKeep (oneself) to oneself Держите (себя) к себе Сторониться людей,

иэбегать обществаKeep a stiff upper lip Держат, жесткая верхняя

губаДержаться молодцом

Keep books Держат книги Вести учетKeep one’s shirt on Держат, рубашка на Оставаться спокойнымKeep peace Идут в ногу Не отставать, держать

темпKickback Вознаграждение Взятка Knock off one’s feet Сбивает с ног Ошеломить Knock one’s block off Сбивают преградить ИзбитьLast ditch effort Усилие последнего рубежа Отчаянная попыткаLet off steam Отпущенный пар Выпускать парNeck and neck Голова в голову Ноздря в ноздрюNo elbow room (1) Никакая комната локтя Негде повернуться,

отсутствие свободы действий

Nosey Parker Любопытный Parker Очень любопытный человек

Not for the world Не для мира Ни за что на светеNot to give on the time of day

Чтобы не дать одному время дня

Полностью избегать кого-либо

Not to touch something with a ten-foot pole

Чтобы не коснуться кое-чего с десятифутовым полюсом

И на версту не подходить

Nothing if not Ничем, если не Безусловно Nutty as a fruitcake Психованный как кекс с

цукатами и орехамиСовершенно спятивший

Of one’s own free will Собственной доброй воли По собственному почину

Page 64: Course Paper on Machine Translation

Off guard От охраны Неготовый к неожиданностям

Off the beaten track Удаленный Непривычный, оригинальный

Off-color Не цвет Безвкснный, неприемлемый

On a shoestring На шнурке На маленьком бюджетеOn and off На и прочь УрывкамиOn pins and needles На булавках и иглах На иголкахOn the block Под давлением На продажуOut of sorts Из видов Не в духеPeriod Период И точкаPock one’s nose in something

Тыкают нос кое в чем Совают свой нос во что-либо

Probe the ground Исследуют основание Прощупывать почвуPull a long face Тянут длинное лицо ПомрачнетьPull up stakes Напряжение, доли Сняться с местаPull in a good word for Вставляли хорошее слово

дляЗамолвить словечко

Put two and two together Соединяют два и два ты Сопоставить фактыRant and rave Напыщенную речь и рэйв Рвать и метатьRed herring Красная сельдь Для отвода глазScratch the surface Царапает, поверхность Снять сливкиSeparate the wheat from the chaff

Отделяют пшеницу от мякины

Отделять зерна от плевел

Sitting duck/target Сидящая утка/цель Легкая добычаSow one’s wild oats Сеет дикий овес Вести распущенный

образ жизниSow the seeds of doubt Сеют семена сомнения Сеять сомненияStand a good chance Выдерживают хороший

шансИметь шансы

Stay the course Пребывание курс Не сбиваться с курсаStraw that breaks the camel’s back

Солома, которая нарушает спину верблюда

Последняя капля

Sugar daddy Сахарный папа Спонсор, богатый любовник

Under one’s breath При дыхании ШепотомUp and about И о Быть на ногахUpper crust Верхняя корка Сливки обществаUpset the applecart Расстроенный планы Спутать кому-либо картыVariety store Склад разнообразия Магазин торгующий

всевозможными мелочами

Wear the pants in one’s family

Носит штаны в семье Быть хозяином в доме

Weasel word Слово ласки Двусмысленное словоWeight one’s word Весит слова Взвешиват свои словаWell-heeled Хорошо-преследуемый При деньгахWet one’s whistle Влажный свист Выпить рюмочку

Page 65: Course Paper on Machine Translation

What have you/what not Что имеет вас/какой не все Все, что угодноWhat’s the (big) idea? Что является (большой)

идеей?Что это ты задумал?

When hell freezes over Когда ад замерзает Когда рак свистнетWhen the chips are down Когда чипсы снижаются В решающий моментWhere the shoe pinches Где обувь зажимает Причина трудностей или

неудобствWhispering campaign Шептание кампании ШушуканьеWhistle in the dark Свист в темноте Не поддаваться страхуWhite sale Белая продажа Распродажа по низким

ценамWith/in a whole skin С/в целая кожа Без единой царапиныWords of one syllable Слова одного слога Сказанное понятным

языкомWork one’s fingers to the bone

Работа пальцы до крайности

Работать до кровавых мозолей

World is your oyster Мир-ваша устрица Все дороги открытыWorth one’s salt Ценность соль Быть хорошим

работникомYak-yak Як яка Болтовня ни о чемYou bet (your boots) Вы держите пари (ваши

ботинки и т.д.)Без сомнения, наверняка

Appendix 2 I.

Page 66: Course Paper on Machine Translation

Original Text in Romanian

Fiecare om se poate prevala de toate drepturile şi libertăţile proclamate în prezenta Declaraţie fără nici un fel de deosebire ca, de pildă, deosebirea de rasă, culoare, sex, limbă, religie, opinie politică sau orice altă opinie, de origine naţională sau socială, avere, naştere sau orice alte împrejurări. În afară de aceasta, nu se va face nici o deosebire după statutul politic, juridic sau internaţional al ţării sau al teritoriului de care ţine o persoană, fie că această ţară sau teritoriu sînt independente, sub tutelă, neautonome sau supuse vreunei alte limitări a suveranitătii. Human translation in English

Everyone is entitled to all the rights and freedoms set forth in this Declaration, without distinction of any kind, such as race, colour, sex, language, religion, political or other opinion, national or social origin, property, birth or other status. Furthermore, no distinction shall be made on the basis of the political, jurisdictional or international status of the country or territory to which a person belongs, whether it be independent, trust, non-self-governing or under any other limitation of sovereignty. Automatic Translation in English

Each arid man can prevala of all right the si the liberties proclamate at bring forward Declaratie except whatever discrepancy as the, of ensample, discrepancy of breed, color, sex, the borage, religion, judgement policy or whatever another judgement, of intranational origination or social, have, bear or any another circumstance. At out thereon, don't arid he will do either a discrepancy after the politic judicial status or intern? Ional of czar or of territory of which holds a man, is as this tare or territory am the independence, below ward, neautonome or obeied another vreunei former confinements sovereignty.

II.

Original Text in English

Everyone is entitled to all the rights and freedoms set forth in this Declaration, without distinction of any kind, such as race, colour, sex, language, religion, political or other opinion, national or social origin, property, birth or other status. Furthermore, no distinction shall be made on the basis of the political, jurisdictional or international status of the country or territory to which a person belongs, whether it be independent, trust, non-self-governing or under any other limitation of sovereignty. Human translation in Romanian

Fiecare om se poate prevala de toate drepturile şi libertăţile proclamate în prezenta Declaraţie fără nici un fel de deosebire ca, de pildă, deosebirea de rasă, culoare, sex, limbă, religie, opinie politică sau orice altă opinie, de origine naţională sau socială, avere, naştere sau orice alte împrejurări. În afară de aceasta, nu se va face nici o deosebire după statutul politic, juridic sau internaţional al ţării sau al teritoriului de care ţine o persoană, fie că această ţară sau teritoriu sînt independente, sub tutelă, neautonome sau supuse vreunei alte limitări a suveranitătii. Automatic Translation in Romanian

Oricine este îndreptatit sa toate drepturi si libertati expuse în aceasta Declaratie, fara deosebire de oricare amabila, ca de exemplu concurati, culoare, sex, limba, religie, politica sau parere alta, nationala sau originea sociala, proprietate, nastere sau starile alte. Nu mai mult, nici o deosebire nu va fi va face pe baza politice, de jurisdictie sau stari tarii internationale sau teritoriul la care persoana îi apartine, daca el este independent, încredeti-va în, non-de sine-de guvernamânt? sau sub oricare alta limitare de suveranitate.

III.

Original Text in Romanian

Fiecare om se poate prevala de toate drepturile şi libertăţile proclamate în prezenta Declaraţie

Page 67: Course Paper on Machine Translation

fără nici un fel de deosebire ca, de pildă, deosebirea de rasă, culoare, sex, limbă, religie, opinie politică sau orice altă opinie, de origine naţională sau socială, avere, naştere sau orice alte împrejurări. În afară de aceasta, nu se va face nici o deosebire după statutul politic, juridic sau internaţional al ţării sau al teritoriului de care ţine o persoană, fie că această ţară sau teritoriu sînt independente, sub tutelă, neautonome sau supuse vreunei alte limitări a suveranitătii. Human translation in English

Everyone is entitled to all the rights and freedoms set forth in this Declaration, without distinction of any kind, such as race, colour, sex, language, religion, political or other opinion, national or social origin, property, birth or other status. Furthermore, no distinction shall be made on the basis of the political, jurisdictional or international status of the country or territory to which a person belongs, whether it be independent, trust, non-self-governing or under any other limitation of sovereignty. Automatic Translation in English

Either jack herself maybe prevala from all drepturile and libertăţile proclamate in prezenta Statement but neither fel from variant ca , from pildă , variant from rasă , color , sex , tongue , religion opinie policy or any altă opinie , from origin national or socialist , estate , birth or any alte împrejurăriÃŽn out from aceasta , non herself va face neither variant after statue political , lawyer or international of ţării or of teritoriului from what ţine one person fie că această country or teritoriu sînt independente , below tutelă neautonome or supuse vreunei alte limit of suveranitătii.

IV.

Original Text in English

Everyone is entitled to all the rights and freedoms set forth in this Declaration, without distinction of any kind, such as race, color, sex, language, religion, political or other opinion, national or social origin, property, birth or other status. Furthermore, no distinction shall be made on the basis of the political, jurisdictional or international status of the country or territory to which a person belongs, whether it be independent, trust, non-self-governing or under any other limitation of sovereignty. Human translation in Romanian

Fiecare om se poate prevala de toate drepturile şi libertăţile proclamate în prezenta Declaraţie fără nici un fel de deosebire ca, de pildă, deosebirea de rasă, culoare, sex, limbă, religie, opinie politică sau orice altă opinie, de origine naţională sau socială, avere, naştere sau orice alte împrejurări. În afară de aceasta, nu se va face nici o deosebire după statutul politic, juridic sau internaţional al ţării sau al teritoriului de care ţine o persoană, fie că această ţară sau teritoriu sînt independente, sub tutelă, neautonome sau supuse vreunei alte limitări a suveranitătii. Automatic Translation in Romanian

Fiecare is entitled la spre tot arthot rights şi freedoms a ezat a aranja fortăreaţă înăuntru this Declaration , fără distinction de orice kind such as race , colonie , sex , limbaj , religie , politic sau alt opinion , naţional sau socialist origine , propriu , naştere sau alt statuie. Mult mai îndepărtat , nu distinction shall a fi made pe basis de la politic jurisdictional sau internaţional statuie de la ţară sau territory la spre care un persoană sub , dacă it a fi independent trust , nu - de sine - guvern sau jos orice alt limitation de sovereignty.

V.

Original Text in Romanian

Fiecare om se poate prevala de toate drepturile şi libertăţile proclamate în prezenta Declaraţie fără nici un fel de deosebire ca, de pildă, deosebirea de rasă, culoare, sex, limbă, religie, opinie politică sau orice altă opinie, de origine naţională sau socială, avere, naştere sau orice alte împrejurări. În afară de aceasta, nu se va face nici o deosebire după statutul politic, juridic sau internaţional al ţării sau al teritoriului de care ţine o persoană, fie că această ţară sau teritoriu sînt independente, sub tutelă, neautonome sau supuse vreunei alte limitări a suveranitătii.

Page 68: Course Paper on Machine Translation

Human translation in English

Everyone is entitled to all the rights and freedoms set forth in this Declaration, without distinction of any kind, such as race, colour, sex, language, religion, political or other opinion, national or social origin, property, birth or other status. Furthermore, no distinction shall be made on the basis of the political, jurisdictional or international status of the country or territory to which a person belongs, whether it be independent, trust, non-self-governing or under any other limitation of sovereignty. Automatic Translation in English

Each man perhaps had been predominating of all the proclaimed liberties and the rights in present declaration without no difference kind that, thus , the breed difference , colour, sex, language, religion, policy opinion or whatever other opinion , of social her national origin , money-bag, birth or whatever other circumstances . Excepting this , it will not do no difference after the status politically, judicial her international of country or of territory of which holds a person , fIE that this territory or country are independent , in trust , nEAUTONOME or obedient VREUNEI other limitations of sovereignty .

VI.

SOURCE TEXT: Le Monde Diplomatique, September 2002

Depuis le 11 septembre 2001, l'esprit guerrier qui souffle sur Washington semble avoir balayé ces scrupules. Désormais comme l'a dit le président George W. Bush, "qui n'est pas avec nous est avec les terroristes".

Systran   Reverso   Human translation

Since September 11, 2001, the warlike spirit which blows on Washington seems to have swept these scruples. From now on, like said it the president George W Bush, "which is not with us is with the terrorists". (37 words)

  Since September 11, 2001, the warlike spirit which blows on Washington seems to have swept (annihilated) these scruples. Henceforth, as said it the president George W. Bush, "which (who) is not with us is with the terrorists". (35 +2 words)

  Since 11 September 2001 the warmongering mood in Washington seems to have swept away such scruples. From that point, as President George Bush put it, "either you are with us or you are with the terrorists." (36 words)

The first point to be made is that Machine Translation is a translation method

that focuses on the source language, while human translation aims at comprehension

of the target language. Machine translations are therefore often inaccurate because

they take the words from a dictionary and follow the situational limitations set by the

program designer. Various types of errors can be seen in the above translations.

Page 69: Course Paper on Machine Translation

Appendix 3.

Practical Tips for Pre-Editing

1. Always run the draft for translation through grammar-checking software, which can catch overly complex construction, compound verbs and obscure phrasing (which they often flag as being in the passive voice).

2. Use a word processor or use that function in a MT program.

Page 70: Course Paper on Machine Translation

3. Use a thesaurus to simplify uncommon usages.

4. Stick to a logical sequences of events, without flashbacks.

5. Spell out abbreviations when they’re first used, with the abbreviations put in all-caps in brackets.

6. Avoid idiomatic, slang and regional or national expressions.

7. Don’t use complex compound structures.

8. Be precise. Avoid fuzzy language.

9. Don’t make the comprehension of the text dependent on formatting like italics or indents.

10. Try to use the ISO Format for dates.

11. Be careful with contracts, where language may have a precise but obscure legal meaning.

12. Translate back and forth (back to the original language) to see where the translation goes astray, and reword.

Tips for Preparing Your Document for TranslationTranslating English materials into other languages has its share of pitfalls, many of which can be avoided. At Simultaneous Translation, we look for the following primary difficulties at the beginning of a translation to prevent problems and ensure consistency and clarity in the target language:

Maintain consistency of terminology Strive for clarity and use simple, direct sentences with basic grammatical construction International users generally prefer straightforward, factual wording Provide a list of all terms which should remain in English (for example proper names,

product names and titles) to alert the translator .

Making Machine Translation Work

While expecting MT to fully translate the complexities of language remains an unrealistic standard, MT can help people get access to a vast amount of information and extract the essence of the meaning. In going beyond that to create their own original messages, people have three alternatives if they want to get acceptable results without help from a human translator:

1. Make adjustments before sending a message. While avoiding extreme “controlled language” approaches, people can learn to speak carefully and add visual hints such as graphics, if the desire to communicate is strong. See Appendices C and D for tips. It’s also a good idea to translate results back into the original language and if something’s completely off base, reword the original.68

Page 71: Course Paper on Machine Translation

2. Check the translation in progress. Some Web programs allow users to list words, such as proper names, that they do NOT want translated. Communication with the person receiving the translation can also become interactive; but people have to be willing to send back-translated sentences that are unclear for clarification, and ask questions. That involves delays that interfere with direct communication and can also means getting over conventions where people want to avoid anything that might imply criticism of the sender. Programs like Translator that allow people to add their own expressions to standard dictionaries can help, and will likely become more widely available.69

3. All parties to the communication adjust their expectations and tolerance. “There are many millions of people around the world, particularly younger people under 30, using the new technology . . . who have no problem at all in accepting the raw English output of the better MT systems as being acceptable as the fractured Americanized English that they use as a common language when they get together with foreign contemporaries on line, or face-to-face in our increasingly global society,” asserts Haynes. He even expects a computerized equivalent of Pidgin English to develop.

ADVICE

Be Concise

Remember that machine translation is a computer process that prefers common words and phrases Start with simple, clear and formal sentences and phrases Keep sentences short, limiting them to 15-20 words for best results If a sentence contains multiple ideas/thoughts, break them into one sentence per idea/thought Avoid unnecessarily complex words and sentences

Write clearly and formally

Word your documents in such a way as to avoid idioms, clichés, colloquial expressions and slang Consider the literal meaning of words and try to express this instead

Avoid Ambiguity

Try not to use words that have more than one meaning for example: - Use "movie" instead of "film" - Use "painting" or "photograph" instead of "picture" Words ending in "ing" can sometimes be ambiguous, such as "rowing", which can be a noun or a verb. Where possible, choose an alternative

Always check spelling and grammar

Incorrect spelling or grammar leads to translation errors, for example, if a word is spelt incorrectly, the translator will not be able to identify the word.

Page 72: Course Paper on Machine Translation

Include appropriate accents

Always use the correct accent marks in your text.

Be aware of Punctuation Pitfalls

Avoid the use of complicated punctuation marks such as parentheses and hyphens. Avoid abbreviations or if you need to use them, keep them consistent. Use articles in front of listed items, for example:- Instead of: the judge and jury - Use: the judge and the jury

Do not leave words out Some words can be implied in everyday use, such as "that, which, who," etc. and are often omitted when writing text - try not to do this as they may be required in the target language.

Bibliography

[1] "Language and Machines. Computers in Translations and Linguistics";

Washington, 1966.

[2] Oettinger, Anthony G.;" Automatic Language Translation. Lexical and Technical

Aspects with Particular Reference to Russian". London, 1960.

Page 73: Course Paper on Machine Translation

[3] Lorscher Wolfgang, "Translation Performance, Translation Process and

Translation Strategies". Tubingen, 1991.

[4] Heaton J.B., Turton N.D., "Longman dictionary of common Errors". Haslow,

1994.

[5] An evolution of Machine – Aided Translation Activities at F.T.D., Contract AF

33(657) – 13616, Case 66556, May 1, 1965, p. G-10.

[6] Бархудов, Л.С., "Язык и Перевод, Вопросы Общей и Частной Теорий

Перевода". Международные Оснащения, 1975.

[7]. www . nbrigham . org / brigham _ machinetranslation . html

[8]. www . deltatranslator . com / tips . htm

[9] http://archiv.tu-chemnitz.de/pub/2001/0043/data/presentation-html

[10] Hutchins, W.J.: Machine Translation: Past, Present, Future. Ellis Horwood Limited, 1986.

[11] Hutchins, W.J.: Research Methods and System Designs in Machine Translation: A Ten-Year Review 1984-1994. In: Proceedings of "Machine Translation Ten Years On". Cranfield, U.K. 1994.

[12] Hutchins, W.J.: Machine Translation: Past, Present, Future. Ellis Horwood Limited, 1986.

[13] Hutchins, W.J.: Machine Translation. History, Current Status, and Possible Future Developments. Lecture at Tzigov Summer School on Applied Linguistics 1995.

[14] Leontieva, N. and Shaljapina, Z.: Current state of MT. In: E. Popov (ed.), Handbook of Artificial Intelligence, Vol. 1, Moscow, "Radio i swjaz", 1990, pp. 216-247.

[15] Sowa, J.F. Conceptual Structures: Information Processing in Mind and Machine. Reading, MA: Addison-Wesley.

[16] http://www.idioms.ru/?q=forum75

[17] Claudia Gdaniec. 1999. Using MT for the Purpose ofInformation Assimilation from the Web. In Workshopon Problems and Potential of English-GermanMT systems.TMI, Chester, UK.

[18] John White. 1995. Approaches to Black-box MachineTranslation Evaluation. In Proceedings of the MTSummit 1995. Luxembourg.

Page 74: Course Paper on Machine Translation

[19] ARNOLD, Doug, BALKAN, Lorna et al. Machine Translation: An Introductory Guide. URL:. MTbook/HTML/.

[20] HUTCHINS, W. John. Translation. Technology and the Translator. URL:

[21] HUTCHINS, W. John. Computer-based Translation Systems and Tools. URL: .

[22] KAY, Martin. History of machine Translation. URL: .

[23] http: //www.foreignword.com/tools/Transnow.htm

[24] www.onlinetranslator.com/?lang=eng

[25] Olivia Craciunescu, Constanza Gerding-Salas, Susan Stringer-O'Keeffe "Machine Translation and Computer-Assisted Translation: a New Way of Translating?"

[26] TRUJILLO, A. Translation engines: techniques for machine translation. London: Springer, 1999. URL: .

[27] William J. Sullivan, "Ideology and Patterns in Translation Error"

[28] Bassnett, S. (1991). Translation Studies. London: Routledge

[29] www.thefreedictionary.com

[30] www.ttt.org/technology.html

[31] www.-306.ibm.com/softwarel

[32] http://archive.tu-chemnitz.de/pub/2001/0043/data/resentation-html/

[33] http://est6.freetranslation.com