Electronic Dictionaries, Glossaries and Terminology Databases
Consulting electronic or digital dictionaries on the computer does not at first appear radically different from using paper dictionaries. However, the advantages soon become clear. It takes far less time to type in a word on the computer and receive an answer than to look through a paper dictionary; there is immediate access to related data through links; and it is possible to use several dictionaries simultaneously by working with multiple documents.
Electronic dictionaries are available in several forms: as software that can be installed in the computer; as CD-ROMs and, most importantly, through the Internet. The search engine Google, for example, gives us access to a huge variety of monolingual and bilingual dictionaries in many languages, although it is sometimes necessary to become on-line subscribers, as with the Oxford English Dictionary. On-line dictionaries organize material for us from their corpus because they are not simply a collection of words in isolation. For example, we can ask for all words related to one key word, or for all words that come from a particular language. That is to say, they allow immediate cross-access to information.
For help with specific terminology there is a wide range of dictionaries, glossaries and databases on the Internet.
b) Analyze the results and answer the questions:
1) Which way is more convenient? Why do you think so?
2) Which way is more accurate? Why?
3) What corrections and editing did you have to make?
c) Decide whether the following sentences are true (T) or false (F):
1 Some machine translation (MT) systems produce good translations.
2 It is difficult to compare different MT systems.
3 The easiest way to evaluate any machine translation of a given text is to compare it to a human translation of the same text.
d) Read the following text and check your answers to the previous task:
Lost in machine translation
You can go out right now and buy a machine translation system for anything between £100 and £100,000. But how do you know if it’s going to be any good? The big problem with MT systems is that they don't actually translate: they merely help translators to translate. Yes, if you get something like Metal (very expensive) or GTS (quite cheap) to work on your latest brochure, they will churn[20] out something in French or whatever, but it will be pretty laughable stuff.
All machine-translated texts have to be extensively post-edited (and often pre-edited) by experienced translators. To offer a useful saving, the machine must make the time the translator spends significantly less than he or she would have taken by hand.
Inevitably, the MT manufacturers’ glossies talk blithely[21] of ‘a 100 percent increase in throughput’ but skepticism remains. Potential users want to make their own evaluation, and that can tie up key members of the corporate language centre for months.
A few weeks ago, translators, system developers, academics, and others from Europe, the US, Canada, China, and Japan met for the first time in a Swiss hotel to mull[22] over MT matters. A surprisingly large number of European governmental and corporate organizations are conducting expensive and elaborate evaluations of MT, but they may not produce, ‘buy or don’t buy’ results.
Take error analysis, a fancy name for counting the various types of errors the MT system produces. You might spend five months working out a suitable scoring scheme – is correct gender agreement more important than correct number? - and totting[23] up figures for a suitably large sample of text, but what do those figures mean? If one system produces vastly more errors than another, it is obviously inferior. But suppose they produce different types of error in the same overall numbers: which type of error is worse? Some errors are bound to cost translators more effort to correct, but it requires a lot more work to find out which.
It isn't just users who have trouble with evaluation. Elliott Macklovitch, of Canada, described an evaluation of a large commercial MT system, in which he analysed the error performance of a series of software updates only to find - as the system's suspicious development manager had feared – that not only had there been no significant improvement, but the latest release was worse.
And bugs are still common. Using a 'test suite' of sentences designed to see linguistic weaknesses, researches in Stuttgart found that although one large system could cope happily with various complex verb-translation problems in a relative clause, it fell apart when trying to do exactly the same thing in a main clause. Developers are looking for bigger, better test suites to help to keep such bugs under control.
Good human translators produce good translations; all MT systems produce bad translations. But just what is a good translation? One traditional assessment technique involves a bunch of people scoring translations on various scales for intelligibility ('Does this translation into English make sense as a piece of English?'); accuracy (‘Does this piece of English give the same information as the French original?'); style, and so on. However, such assessment is expensive, and designing the scales is something of a black art.
Properly designed and integrated MT systems really ought to enhance the translator's life, but few take this on trust. Of course, they do things differently in Japan. While Europeans are dabbling their toes and most Americans deal only in English, the Japanese have gone in at the deep End. The Tokyo area already sports two or three independent MT training schools where, as the eminent Professor Nagao casually noted in his presentation, activities are functioning with the efficiency of the Toyota production line. We're lucky they're only doing it in Japanese.
E) Each of the sentences below (except one) summarizes an individual paragraph of the text. Order the sentences so that they form a summary of the text. One of the sentences contains information, which is not in the text. Which one?
1 The developers of MT systems have also had problems evaluating their systems.
2 Many European organizations are evaluating MT, but the results may not be conclusive.
3Assessing machine translations as good or bad is very difficult because such judgments cannot be made scientifically.
4 It is time-consuming for potential users to test the MT manufacturers’ claims that their products double productivity.
5 Better tests are needed to monitor linguistic weaknesses in MT Systems.
6 All machine translations need to be edited by a human translator.
7 A reliable MT system is unlikely to be available this century.
8 The price of MT systems varies greatly and none actually translates.
9 The Japanese have a few independent MT training schools, which are said to be very efficient.
10 Analysing the errors made by MT systems is inconclusive because it may only show that different systems produce similar numbers of error types.
f) Match each of the following verbs from the text with the expression similar meaning:
1) churn out 2) tie up 3) mull over 4) tot up 5) cope with 6) fall apart | a) add up b) think carefully about c) manage successfully d) produce large amounts of e) fail f) occupy the time of |
g) Using the paragraph reference given, find words or phrases in the text which have a similar meaning to:
1 ridiculous (para.l)
2 colour brochures (para.3)
3 casually (para.3)
4 sure to (para. 5)
5 group (para. 8)
6 mysterious ability (para.3)
7 experimenting in a small way (para.9)
8 invested heavily (para.9)
h) Look at these sentences. Discuss why a machine might find them difficult to translate:
I bought a set of six chairs. The sun set at 9 p.m.
He set a book on the table. We set off for London in the morning.
She had her hair set for the party. The VCR is on the television set.