Want to read Slashdot from your mobile device? Point it at m.slashdot.org and keep reading!

 



Forgot your password?
typodupeerror
×
Technology

Romancing The Rosetta Stone 486

Roland Piquepaille writes "Not only this news release from the University of Southern California has a fantastic title, it also has a great content. This story is about one of their scientists, Franz Josef Och, whose software ranks very high among translation systems. "Give me enough parallel data, and you can have a translation system for any two languages in a matter of hours," said Dr. Och, paraphrasing Archimedes. His approach relies on two concepts, gathering huge amounts of data, and applying statistical models to this data. It completely ignores grammar rules and dictionaries. "Och's method uses matched bilingual texts, the computer-encoded equivalents of the famous Rosetta Stone inscriptions. Or, rather, gigabytes and gigabytes of Rosetta Stones." Read my summary for more details."
This discussion has been archived. No new comments can be posted.

Romancing The Rosetta Stone

Comments Filter:
  • Article text (Score:4, Informative)

    by Anonymous Coward on Monday July 28, 2003 @10:51AM (#6551246)
    Romancing the Rosetta Stone

    'Give me enough parallel data, and you can have a translation system in hours'

    University of Southern California computer scientist Franz Josef Och echoed one of the most famous boasts in the history of engineering after his software scored highest among 23 Arabic- and Chinese-to-English translatio systems, commercial and experimental, tested in in recently concluded Department of Commerce trials.

    "Give me a place to stand on, and I will move the world," said the great Greek scientist Archimedes, after providing a mathematical explanation for the lever.

    "Give me enough parallel data, and you can have a translation system for any two languages in a matter of hours," said Dr. Och, a computer scientist in the USC School of Engineering's Information Sciences Institute.

    Och spoke after the 2003 Benchmark Tests for machine translation carried out in May and June of this year by the U.S. Commerce Department's National Institute of Standards and Technology.

    Och's translations proved best in the 2003 head-to-head tests against 7 Arabic systems (5 research and 2 commercial-off-the-shelf products) and 14 Chinese systems (9 research and 5 off-the-shelf). In the previous, 2002 evaluations they had proved similarly superior.

    The researcher discussed his methods at a NIST post-mortem workshop on the benchmarking held July 22-23 at Johns Hopkins University in Baltimore, Maryland.

    Och is a standout exponent of a newer method of using computers to translate one language into another that has become more successful in recent years as the ability of computers to handle large bodies of information has grown, and the volume of text and matched translations in digital form has exploded, on (for example) multilingual newspaper or government web sites.

    Och's method uses matched bilingual texts, the computer-encoded equivalents of the famous Rosetta Stone inscriptions. Or, rather, gigabytes and gigabytes of Rosetta Stones.

    "Our approach uses statistical models to find the most likely translation for a given input," Och explained

    "It is quite different from the older, symbolic approaches to machine translation used in most existing commercial systems, which try to encode the grammar and the lexicon of a foreign language in a computer program that analyzes the grammatical structure of the foreign text, and then produces English based on hard rules," he continued.

    "Instead of telling the computer how to translate, we let it figure it out by itself. First, we feed the system it with a parallel corpus, that is, a collection of texts in the foreign language and their translations into English.

    "The computer uses this information to tune the parameters of a statistical model of the translation process. During the translation of new text, the system tries to find the English sentence that is the most likely translation of the foreign input sentence, based on these statistical models."

    This method ignores, or rather rolls over, explicit grammatical rules and even traditional dictionary lists of vocabulary in favor of letting the computer itself find matchup patterns between a given Chinese or Arabic (or any other language) texts and English translations.

    Such abilities have grown, as computers have improved, by enabling them to move from using individual words as the basic unit to using groups of words -- phrases.

    Different human translators' versions of the same text will often vary considerably. Another key improvement has been the use of multiple English human translations to allow the computer to more freely and widely check its rendering by a scoring system.

    This not coincidentally allows researchers to quantitatively measure improvement in translation on a sensitive and useful scale.

    The original work along these lines dates back to the late 1980s and early 1990s and was done by Peter F. Brown and his colleagues at IBM's Watson Research Center.

    Much of the improvement and
  • Let me know (Score:5, Funny)

    by gazuga ( 128955 ) on Monday July 28, 2003 @10:51AM (#6551249) Homepage
    when it's in the form of a fish, and can fit in my ear...
  • Great summary (Score:3, Insightful)

    by spectasaurus ( 415658 ) on Monday July 28, 2003 @10:53AM (#6551266)
    You know, it's not really a summary when you just delete half the article.
  • DARPA (Score:2, Insightful)

    That reference to DARPA has me a little worried about the sort of uses this technology will be put to. I wonder, are the CIA trying to shore up holes in their translation abilities (particularly for Arabic/etc) by using software. What happens when you pair this technology up with the Echelon project? Are we going to see a dramatic rise in the ability of the government to spy on nationals and particularly foreign nationals now?
    • Re:DARPA (Score:5, Insightful)

      by Abcd1234 ( 188840 ) on Monday July 28, 2003 @11:04AM (#6551367) Homepage
      Oh please... so many conspiracy theories. You do realize that the *internet* was originally developed by DARPA, right? My point: DARPA does a lot of work... not all of it revolves around spying on or otherwise taking away the rights of American citizens.
      • Re:DARPA (Score:5, Insightful)

        by wwest4 ( 183559 ) on Monday July 28, 2003 @11:30AM (#6551571)
        well, not EVERY bottle of beer at the duff plant has a nose or hitler's head in it, but i'm glad the inspector is tasked to look at every single bottle.

        just because government abuse isn't guaranteed doesn't mean we shouldn't vigilantly examine the possibilities when we see them.

        it's all boils down to balancing powers of government and freedom of individuals, and this country (USA) was founded upon principles intended to favor the rights of individuals. i'll go out on a limb and make a value statement - that's the way to go. power to the people, man!
    • Re:DARPA (Score:2, Interesting)

      by kmac06 ( 608921 )
      Kneejerk /. response: its a government conspiracy to take away more of our rights.

      Kneejerk /. mod response: he's right.
    • Well I dont have to worry, I dont speak arabic OR chinese!
  • No really... what if it used a shared database and there were hundreds, or thousands, of the systems around the world... Seems like it could become a pretty sophisticated system. And maybe one day it will be available in the form of a small fish which you place in your ear?
  • Oh god... (Score:4, Funny)

    by gerf ( 532474 ) on Monday July 28, 2003 @10:55AM (#6551278) Journal
    The uber-geeks are going to have a field day with Klingon...
  • by Speare ( 84249 ) on Monday July 28, 2003 @10:56AM (#6551290) Homepage Journal

    "Give me enough" is a key element of the Law of Eventuality. Give me enough money, and I'll solve the Microsoft monopoly threat with a hostile takeover. Give me enough time and I'll clean up almost any unnatural disaster site by leveraging nature's own methods.

    Give me enough simulated neurons and enough truisms and I'll make a sentient machine.

    Eventually, with enough resources, anything is possible. Throwing more time and resources to a problem is rarely exciting science. Reducing the inconveniently large values of 'eventually' and 'enough' are the real problem.

    • Err... how is this interesting or insightful? It's barely related to the discussion! If what you're is referring to is the large corpus of paired texts they inject into the system, you've completely missed the point.

      The cool science here is in the advancements in their statistical model and new techniques they've developed for "scoring" translations in order to improve their output. In addition, they've also demonstrated the ability to statistically translate whole phrases effectively, rather than indiv
    • OT, I know, but how would a hostile takeover solve the "Microsoft monopoly threat"? Sounds like one giant replacing another...
    • by Pac ( 9516 ) <paulo...candido@@@gmail...com> on Monday July 28, 2003 @11:45AM (#6551702)
      ...and I will make pseudo-insightful comments based on the headline text without reading any of the source articles, until my karma is excellent?
  • by zptdooda ( 28851 ) <deanpjm&gmail,com> on Monday July 28, 2003 @10:59AM (#6551322) Journal
    That's an example from a few years' back of an attempt to translate "the spirit is willing but the flesh is weak" from English to Russian and back to English using a different translator.

    Can anyone try this on the new (or some other recent) algorithm?

    BTW here's Doc Och's most recent website:

    Franz Josef Och [isi.edu]

    • translated to russian using systran [systranbox.com] and back using babelfish [altavista.com] I got "spirit is willingly ready but flesh it is weak"
    • by rossz ( 67331 ) <ogreNO@SPAMgeekbiker.net> on Monday July 28, 2003 @11:13AM (#6551427) Journal
      That particular phrase translated badly because they used a word-for-word translation program. You simply can't do that, especially when dealing with euphenisms. This new system is the only possible way that could properly translate text.

      My wife is a professional translator and has absolutely no respect for machine translatations.
    • by quantum bit ( 225091 ) on Monday July 28, 2003 @11:14AM (#6551444) Journal
      You know, that actually does sound like something that would be a Russian aphorism...
    • Heh, given this is a not-uncommon phrase in the English language, it very well may be in their English-to-target-language corpus, meaning it could end up being a straight lookup-and-translate operation. Which is, of course, one of the advantages of a system like this (you can translation common idioms without having to analyze the text itself).
    • Let's see what google has to say:

      English: The spirit is willing but the flesh is weak.

      German: Der Geist ist bereit, aber das Fleisch ist schwach.
      back: The spirit is ready, but the flesh is weak.

      French: L'esprit est disposé mais la chair est faible.
      back: The spirit is laid out but the flesh is weak.

      Italian: Lo spirito è disposto ma la carne è debole.
      back: The spirit is arranged but the meat is weak person.

      Portugese: O espírito é disposto mas a carne é fraca.
      back
    • by JJ ( 29711 ) on Monday July 28, 2003 @05:31PM (#6554347) Homepage Journal
      This actually is a myth. That particular text and translation was taken as anecdotal in a 1964 report. I did a masters thesis on MT at the University of Chicago and my advisor (once a major figure in MT) refused to approve my thesis until I got that statement correct.
  • by tuxlove ( 316502 ) on Monday July 28, 2003 @10:59AM (#6551328)
    I believe that using a statistical approach like this is a step in the right direction. Manually building sets of rules, dictionaries, etc., is a waste of time and hard to do. And manuall-built systems become stale as languages evolve, unless a lot of continuing work is done.

    For me the holy grail is when I can converse with a computer meaningfully. I believe a similar approach will be required for the computer to "understand" language, and to be able to formulate a coherent and appropriate response.

  • Isn't the Doc [mit.edu] supposed to be in the next Spiderman movie?

  • Universal translator anyone?

    Er, aging geek embarrasing self again, mutter...
  • by Alton_Brown ( 577453 ) on Monday July 28, 2003 @11:02AM (#6551348)
    From the article: his software scored highest among 23 Arabic- and Chinese-to-English translatio systems

    Oops - guess we need some more parallel data (or a few more gigs of rosetta stones).
  • by Zog The Undeniable ( 632031 ) on Monday July 28, 2003 @11:03AM (#6551357)
    "Most the bay only of news of the college of southern extremity California it knows an all big contents all there is this emission annular subject, it also there is a RolandPiquepaille and it writes. The Franz taxes where his software height one lyel with lines up between the translation system quite phu the Och and this history are the summary thing their scientist. The Och "it gave the data which is parallel is sufficient in me, it spread out," inside questioning the hour 2 specialties the language which it does not do of the multi Archimedes which is the possibility which there will be a hazard translation system the doctor repulsively it talked. It approach collects the sheep which data is enormous, apply the statistical model in this data a foundation in 2 concepts which it puts. It is complete and the wool of rule lu the dictionary of grammar "the m3ethode of the Och the duplex language original and the Rosetta which agree one equivalent with computer password of noble and wise pebble epitaph adopts. Or, rather, the gigaoctets and pebble gigaoctets of the Rosetta." Detail fact compared to read the hazard my synopsis.

    English --> French --> English --> Korean --> English. Of course, it helps that the first sentence is munged anyway ;-)

  • Integration (Score:3, Interesting)

    by slusich ( 684826 ) <slusich@gmRASPail.com minus berry> on Monday July 28, 2003 @11:03AM (#6551358)
    Sounds like a brilliant idea. Hopefully this is something that could eventually be compacted enough to fit into consumer electronics. It would be great to be able to watch TV from every country without any language barrier!
    • Re:Integration (Score:4, Interesting)

      by ahfoo ( 223186 ) on Monday July 28, 2003 @11:39AM (#6551653) Journal
      Not to sound arrogant, but I find actually learning another language by watching foreign TV with subtitles in the original language to be even more interesting than watching the dubbed or English subtitled version. It involves commitment to get to the point where you can understand the basics, but there are rewards to making a commitment to learn something new.
      I like the idea of translating sentence by sentence as opposed to grammatically and word for word. I'm sure this guy is right that at some point this will produce reasonably acurate translations in many cases, but multiple languages are one of our greatest treasures.
      I have read that the single most important factor in preventing senile dementia is the difference between those who continue to create novel memories throughout their lives and those who stick to what they have already learned. Learning multiple languages is a wonderful thing and once you get well into it, it is a lot of fun. It certainly increases your options for punning and rhyming and you end up with lots of aliases.
  • Dialects? (Score:3, Interesting)

    by dethl ( 626353 ) on Monday July 28, 2003 @11:06AM (#6551381)
    How can this system compensate for the different dialects of all of the different languages?
  • Well, so? (Score:4, Funny)

    by k98sven ( 324383 ) on Monday July 28, 2003 @11:09AM (#6551408) Journal
    What is the novelty of this?

    It's hardly news that you can always find correlations in two sufficiently large sets of data.

    Reminds me of the Steve Martin joke:

    "Chicks go for the intellectual types. I figured the best way to impress 'em was to read a lot of books. But hey, do you know how many books there are? Why, there must be, hundreds of them. But I was already a pretty smart guy. I didn't waste my time reading all those books. Heck no.
    I read, the dictionary. Hey--I figure it's got all the other books in it."
  • Another IT masters thinks he can invent a perfect translation system, simply based on 0s and 1s.

    I have said it before, on /. and elsewhere, machine translation does not work.

    A good translation is based on several non-quantifiable parameters:
    1. Context.
    2. Grammar.
    3. Vocabulary.
    4. Nuance.

    Example:

    "My controller has failed. He is going to be replaced" can mean:

    • My HDD controller is dead. I need to replace it, so that my computer can access its hard disks (For the slashdot crowd).
    • The financial controller of m
    • by radish ( 98371 ) on Monday July 28, 2003 @11:43AM (#6551682) Homepage
      You're right, traditional machine translation is difficult, primarily due to context. However, you're also right that the example you gave is a bad one - in english it only has one meaning (the second one you give). A HDD controller would never have an assigned gender. Of course in German for example, it would (not sure which though - neuter?).

      However you're missing what I think is the most important point. If an example is so ambiguous as to confuse an "ideal" machine, it would confuse us too. What you're really saying is "it is possible to write sentences with ambiguous meaning in most languages" - which is of course true. That doesn't however make it impossible to create a machine which is at least as good as a human at translating (and wouldn't that be good enough?). When you read something you interpret it according to a set of learned rules. Obviously there's the basic syntax and vocab, but then you add context like the other clauses in the prose, the identity of the author, the subject matter. We're a long way off getting those concepts into a machine reader, but I would be very hesitant to say we'll never get there.

      Besides, the artical is about taking a different approach to the problem - one which should be quite happy with ambiguity. They're looking at essentially pattern matching, so provided your sample data sets include enough info to describe the ambiguity it should have a decent enough chance of working it out.
  • by TwistedGreen ( 80055 ) on Monday July 28, 2003 @11:11AM (#6551423)
    "One of the great advantages of the statistical approach," Och explained, "is that most of the work goes into components that are language-independent. As long as you give me enough parallel data to train the system on, you can have a new system in a matter of days, if not hours."

    This statistical method is probably the best approach to computerized translation. It seems to approximate how the human mind will translate a give sentence most efficiently. Language can get awfully complex, and individual words often have, at best, an ambiguous meaning when interpreted alone. One must take into account the context of that word to specify and refine its meaning. This obviously leads to a huge number of permutations to represent a huge variety of thoughts, but the relative size of this number is diminishing as computers become more powerful.

    Therefore, instead of playing with messy grammars and sentence structures, we can simply have a catalogue of thoughts as represented by words, and correlate that catalogue with a different set of words to facilitate translation. This software would operate on a deeper level than it would if it operated with the words and symbols themselves. It would utilize a map of the deep structures of language, instead of a map of the less-meaningful words and grammars.

    I really like this method, and while it may seem like a brute-force hack applied to translation, the simple fact that languages do not contain elegant patterns must be accepted. It also appears to be a most efficient method, as the simple comparisons involved would bring the speed of translation into realtime.
    • by Jerf ( 17166 ) on Monday July 28, 2003 @03:53PM (#6553647) Journal
      This software would operate on a deeper level than it would if it operated with the words and symbols themselves. It would utilize a map of the deep structures of language, instead of a map of the less-meaningful words and grammars.

      Actually, as a result, it operates on a shallower level. In fact, it's almost like you wrote this comment for an article in a parallel universe where statistical translation was the norm, and somebody was just now proposing symbolic translation, so much so that it's almost spooky.

      This translation technique is so shallow it doesn't even particularly care what languages it works with. In a way, it can't really be said to be "translating" in the traditional sense; it's just correlating phrases with no clue what they are.

      Traditional symbolic translation is better described by what you said:

      Therefore, instead of playing with messy grammars and sentence structures, we can simply have a catalogue of thoughts as represented by words, and correlate that catalogue with a different set of words to facilitate translation.

      Word(/phrase) -> symbol -> word(/phrase) is traditional tranlation. This is word -> word translation.

      It's working better because we've had little or no success creating the middle part of the symbolic translation; matching the symbology used in our head has proven impossible to date. This works better by skipping that step, which introduces horrible distortions by forcing the words to fit into an incredibly poor symbology (compared to what we're actually using).

      However, in theory, traditional translation should still have a brighter future; this is a hack around our ignorance, perhaps even a good one, but eventually we will want to extract the symbols.

      (Incidentally, it's also why this same technique can't be used to match words -> symbols; we don't know how to represent the symbols yet! This kind of technique could eventually potentially be hybridized with something else to attack that problem, but simple, direct application can't result in the complicated relationships between symbols that exist, and we'd want a computer to "understand" those relations before we'd say it was truly translating or understanding English.)

      Anyways, just flip your comments around 180 degrees and you're pretty close.
  • University of Southern California computer scientist Franz Josef Och echoed one of the most famous boasts in the history of engineering after his software scored highest among 23 Arabic- and Chinese-to-English translatio systems, commercial and experimental, tested in in recently concluded Department of Commerce trials.

    Maybe what Dr. Och should do next is write some software to double-check the work of whoever translates his press releases from the original Latin. The translator seems to have missed a fe
  • Copyright issues (Score:2, Insightful)

    by PhilHibbs ( 4537 )
    I wonder if the resultant translation engine could be considered a derivative work of the texts that populated it. This system is standing on the shoulders of all the translation efforts that went in to it. I think it's a great idea, but in the current IP climate, could well be shot down in flames. How much dual-language text is available in the PD or on open content licence?
  • by jd ( 1658 ) <imipakNO@SPAMyahoo.com> on Monday July 28, 2003 @11:15AM (#6551447) Homepage Journal
    The Rosetta stone encoded three languages, not two, where two were known in advance. Indeed, there have been many three-way translations of treaties found, now.


    The use of three languages is critical. Grammar isn't consistant, and words have multiple meanings. By using two known languages, you can eliminate many of the errors thus introduced, because the chances of some error fitting both known languages in the same way is much smaller.


    If you double the number of known languages, you more than quarter the number of errors, because although errors can occur in either or both, they're unlikely to be the same error. Once more information exists, you can re-scan the same text and fill in the blanks.


    Me, personally - I'd require four languages, three of which were known. The number of texts required would be considerably smaller and the number of residual errors would be practically non-existant.


    They chose two languages for the obvious reason: It's simple. It's easy to find a student who knows two languages. At least, easier than finding one who knows four.


    However, the price of simplicity is bad science. The volume of information they require makes their system little better than an infinite number of very smart monkeys with text editors and a grep function. That they're being paid signficant money on such stuff is a joke.


    If they offered me the same money (and one of those Linux NetworX clusters) I could have a superior system in a month, although (as stated above) it would require more than one known language.

    • Not to mention.. (Score:4, Interesting)

      by k98sven ( 324383 ) on Monday July 28, 2003 @11:47AM (#6551727) Journal
      The Rosetta stone itself did not do much in the way of our knowledge of the egyptian language.
      What it did do, was provide insight into their method of writing.
      It was the latter discovery of the the relation between Coptic and Egyptian that revealed most of the actual language.

      (IIRC)
      • Re:Not to mention.. (Score:4, Interesting)

        by LenE ( 29922 ) on Monday July 28, 2003 @01:44PM (#6552765) Homepage

        For those who don't know, Coptic is Egyptian written in Greek, or at least the Greek alphabet. It would be similar to transcribing a language that uses glyphs for words by recording them with the phonemes and alphabet of another language.

        A more modern example is what happened with the slavic Croatian language. The original speakers had a glyph based alphabet called Glagolitic [www.hr], through the middle ages. This would be as foreign as Egyptian hieroglyphs to people today, and could stand in nicely for an alien text in any sci-fi movie.

        Through falling under different feudal states (Venice, Austro-Hungary) the language was cast under both the Cyrillic and Roman alphabets. Today Croatian uses an accented Roman alphabet (like French), but each letter has only one pronunciation, like Russian.

        -- Len

    • by Abcd1234 ( 188840 ) on Monday July 28, 2003 @11:51AM (#6551764) Homepage
      If they offered me the same money (and one of those Linux NetworX clusters) I could have a superior system in a month, although (as stated above) it would require more than one known language.

      LOL! If this problem was so friggin' easy, why are these researchers the first to demonstrate a working system using this technique (which blows away all existing systems, BTW)? Hell, if it's as easy as you say, this whole "translating text" thing must be a breeze. I wonder why so much money is spent every year on R&D in this area? Hell, why didn't they just hire you to whip up a system in a month?

      Why? Because it ain't that easy and you have no idea what you're talking about. Given these are world-class researchers, I'm sure they've considered the multiple-translation route, and subsequently rejected it for very good reasons (likely far more complex than your simplistic "it's easier" excuse). Moreover, the really hard work in this area is the statistical modelling necessary to generate a working system, something which would, I suspect, be far more complex if a multiple-translation route were taken. But, hey, that's just some number crunching, right? What's so hard about that?
    • by William Tanksley ( 1752 ) on Monday July 28, 2003 @11:59AM (#6551864)
      If you double the number of known languages, you more than quarter the number of errors

      Your post is reasonable and interesting (using three-way parallelism would give better translations), but you're missing something important here.

      First, none of these languages are "known" to this interpreter program. The program reads parallel texts, and when you feed it a text without a parallel, it generates the parallel for you. In other words, it can translate either way. So you don't have two known languages and one unknown; all you have is three text corpuses. (Well, in this case you have two, but you know what I mean.)

      Second, yes; three would be FAR better than two; but two is also useful, and in more situations. You don't always have a Rosetta stone.

      They're doing well here. Yes, there's an obvious next step to take; but no, the existance of a "next step" doesn't destroy the usefulness of this step.

      -Billy
    • RTFA. The method described in the article is a purely statistical method, NOT a semantic one; it has zero "knowledge" of grammar, syntax, or meaning. So having more than one "known" language to start with would not help in the slightest, because the advantages that you describe are only applicable to semantic methods.

      I agree though that the analogy to the Rosetta Stone is a poor one.
  • Ranking System (Score:3, Interesting)

    by freeze128 ( 544774 ) on Monday July 28, 2003 @11:18AM (#6551480)
    Even existing translation programs could benefit from a ranking system. Wouldn't it be helpful if you could tell just how confident the translator is about a certain phrase or word? That way, you could rephrase your sentence before you foolishly ask someone to "taste" you....
  • by meshko ( 413657 ) on Monday July 28, 2003 @11:20AM (#6551500) Homepage
    I understand that this is a cool idea for building automatic translators, but is it practical? Basically what they are doing is taking a well-researched domain of languages and trying to make something new and cool in it by completely ignoring the domain knowledge. My intuition tells me that "always use as much domain knowledge as posssible" is an engineering axiom.
  • by Flwyd ( 607088 ) on Monday July 28, 2003 @11:21AM (#6551512) Homepage
    As press releases tend to do, this leaves much to be desired for folks who are familiar with the discipline. As I read it, it seems to imply that the main driver is phrase-matching. What does it do with phrases it hasn't seen before? The problem is solved by throwing lots of data at it -- how much data is needed for a reasonable system? How well does it generalize to text outside the domains of the training data?

    Incidentally, had my brother been a girl, he was in serious danger of being named Rosetta Stone.

    -- Trevor Stone, aka Flwyd
  • Wordrank (Score:2, Interesting)

    by chronos2266 ( 514349 )
    I always thought it would be interesting if google applied its page rank algorithm to provide a translation service. Like poll the top 5 translation service sites for a translated sentence and then based on what each of them return, generate a 'average' or best possible result for that sentence.
  • by Potpatriot ( 684373 ) on Monday July 28, 2003 @11:28AM (#6551564)
    How about piping in various algorirhtms encoded in Pascal and C into the thing and seeing what it does to convert arbitrary sources. Where Can I get the soource? Pawel
  • by MobyDisk ( 75490 ) on Monday July 28, 2003 @11:30AM (#6551579) Homepage
    So, can I train this program with a bunch of requirements documents, and a bunch of implementations, and have it learn how to code? :-) If so, I think I am obsolete. *poof*
  • by The Raven ( 30575 ) * on Monday July 28, 2003 @11:40AM (#6551659) Homepage
    I wonder how this would fare putting two computer languages side by side? I mean... take a few thousand programs, coded using the same algorithms but different computer languages... would his language translation software translate between them? Would it be able to differentiate between languages that manually allocate memory and those that use garbage collection? How about between procedural langauages like C, and more esoteric and oddly structured languages like LISP?

    An interesting challenge, eh?

    Would there be any benefit to this?
  • Scientific Papers (Score:4, Informative)

    by acoustiq ( 543261 ) <{ten.emohtfos} {ta} {qitsuoca}> on Monday July 28, 2003 @11:50AM (#6551751) Homepage
    Being an undergrad hoping to do research in this area in the next few years, I've already read a few of Och's papers and others in the field. Some of the best that I remember are: Kevin Knight prepared an excellent (if now somewhat outdated) introduction to statistical machine translation that you can see in HTML [jhu.edu] or RTF [jhu.edu] (the formatting was corrupted when the RTF was converted to HTML - I recommend the RTF).
  • by gemseele ( 172754 ) on Monday July 28, 2003 @11:53AM (#6551786)
    Time for inflamatory reasoning. The statistical approach will beat out the grammar and rule based ones, at least for English, is for the simple reason:

    English is not a language

    Or rather, it resembles one but is more not than is, IMO. It is a large collection of idiomatic expressions that changes quite rapidly (and not only in colloquial forms, just look at what the political-correctness movement has done to phraseology). You know the story... more exceptions than rules, things that are legitimate to say language-wise are considered incorrect anyways, and vice versa, etc. etc.

    That's not to say it doesn't have advantages; it's relatively easy to learn the basics of communication since it's weakly conjugated, has genderless articles, fairly simple uncased sentence structure. But, it is monstrous to master and I suspect most native speakers aren't true masters (not to mention the orthographical nightmare; is English the only language with spelling bee contests?)

    The reason it's the new lingua franca (or should it be lingua angla now?) is techno-socio-political as is always the case. Stop harping on Americans for being largely mono-lingual. "Why didn't the Romans learn the local languages when they controlled Europe? Because they didn't have to." If every state spoke a different language, which would be more akin to Europe, then there would be need.
    • English is not a language... [because it]... is a large collection of idiomatic expressions that changes quite rapidly

      Fair enough, English changes rapidly alright, but how would you define a language? A set of logical syntactic and semantic rules that haven't changed for the past few thousand years? I can think of only two languages like that, Latin and Sanskrit.

      Nope, I can't agree with your assertion; language is much more than mere (unchanging) grammar. In many multi-cultural places, it is a strong

    • by Jeremi ( 14640 ) on Monday July 28, 2003 @02:20PM (#6553055) Homepage
      English is not a language. Or rather, it resembles one but is more not than is, IMO. It is a large collection of idiomatic expressions that changes quite rapidly


      You are actually arguing that English is not a dead language. Every language that is actually in use by large numbers of people is as you describe.

  • How's that news? (Score:3, Interesting)

    by Yurka ( 468420 ) on Monday July 28, 2003 @01:14PM (#6552532) Homepage
    This has already been done some years ago in Canada, where the translation system was fed the complete text of parliamentary debates for umpteen years (required by law to be translated by humans into French, if originally in English, and vice versa). I don't know how it fares when presented with a sample of parliament-speak (I concede, this is not a fair approximation of human language), but it fails miserably on a simple rhyme. Read your Hofstadter, guys.
  • by Bodrius ( 191265 ) on Monday July 28, 2003 @03:49PM (#6553617) Homepage
    Interesting method.

    It seems to me this is more similar to natural learning of a language (usually at a young age) by exposure and immersion, as opposed to scholar learning of a language in classrooms, etcetera.

    It shouldn't be surprising that in humans, the first method also works best at acquiring fluency in multiple languages. As a matter of fact, it's the only method through which we come to understand our FIRST language, which is in almost every case the one we command the best.

    I think most people get, by consuming huge amounts of information, a feeling of "what sounds right" and "what sounds wrong" that is more effective for them at predicting the unwritten rules and exceptions, both in translations and in original sentence-creation, than memorizing a set of grammar rules which, in the end, are just codifications of the current state of the language.

    I don't think the success of the approach means the symbolic methods are pointless for this endeavor, any more than the formal study of languages and their grammars is for human translators.

    Professional writers and translators do study such rules to dramatically improve their command of the different languages, and do get much better results.

    But it seems to me they are more successful going from "statistical matching with massive real-use data" to "optimized grammar rules matching the data" than going backward, from "scholastic grammar rules" to "consumption of massive data to acquire exceptions, and correct and complement the rules".

    What would be interesting, I think, is if one can study the state of the system after it's performing well and extract/deduct grammar rules, algorithmically.

    It would be interesting to see the results of a program doing that, collecting (and correcting) the grammar using the data, and using the grammar rules when no match in the dictionaries is found to, say, apply a greater weight to the gramatically-correct choice among the alternatives.

    If the results were good with this approach, one could consider decreasing the size of the database as the grammar gains stability. Use that memory for other processes, other languages, or new sample data that could not be examined before.

Every nonzero finite dimensional inner product space has an orthonormal basis. It makes sense, when you don't think about it.

Working...