You could use forced translation. It won't be perfect, but it might
help you catch some of these errors.

In forced translation, you give the source sentence and a proposed
translation, and Moses attempts to translate the source sentence into
the proposed translation. If it succeeds, it will give you the
associated score.

Now, forced translation will fail even for some valid translations.
But one would hope that it would fail more often for invalid
translations.

Cheers,
Lane


On Mon, Sep 26, 2011 at 6:50 AM, Julian Myerscough
<[email protected]> wrote:
> Thanks to everyone for your replies.
>
> I am not sure if the intent of my original question was clear. What I
> would like to do is check newly translated (by a human) sentences
> against a set of "approved" human translations with a view to finding
> errors or inconsistencies (more emphasis on errors). For example, in the
> following human translated English -> Dutch translation, "clockwise" is
> mis-translated as "anti-clockwise".
>
> close firmly by turning the knob clockwise ;
> sluit het instrument stevig door de knop tegen de klok in te draaien ;
>
> This is the kind of error I want to check for, but ideally I also want
> to check for antonyms, inconsistent usage of verbs, inconsistent usage
> of nouns etc.
>
> I could imagine this might possible by "mining" the phrase-table file
> and to find the minimum probability of each part of a sentence
> translation (antonyms might have a negative probability for example).
> ...but wondering if there's an easier or better way to do it.
>
> Looking forward to your thoughts
>
> Julian
>
>
>
> Victor Chahuneau wrote:
>> What you are looking for is the perplexity of the language model (from with 
>> you can derive the probability of the sentence).
>> Depending on the toolkit you are using for language modelling, there are 
>> options to score a sequence of sentences or a single sentence.
>> e.g. SRILM (ngram)
>>  -ppl:                    text file to compute perplexity from
>> IRSTLM (compile-lm)
>> --score|-s [yes|no]  (computes log-prob scores from standard input)
>> --eval|-e text-file (computes perplexity of text-file and returns)
>>
>>
>> Le 23 sept. 2011 à 10:27, Julian Myerscough a écrit :
>>
>>
>>> Hi folks,
>>>
>>> I would like to check newly translated sentences against an existing
>>> language model to see how "likely" (or more importantly "unlikely") they
>>> are (how well they fit into the existing model).
>>>
>>> Could you give me any pointers?
>>>
>>> Cheers
>>>
>>> Julian
>>>
>>> _______________________________________________
>>> Moses-support mailing list
>>> [email protected]
>>> http://mailman.mit.edu/mailman/listinfo/moses-support
>>>
>>
>>
> _______________________________________________
> Moses-support mailing list
> [email protected]
> http://mailman.mit.edu/mailman/listinfo/moses-support
>



-- 
When a place gets crowded enough to require ID's, social collapse is not
far away.  It is time to go elsewhere.  The best thing about space travel
is that it made it possible to go elsewhere.
                -- R.A. Heinlein, "Time Enough For Love"

_______________________________________________
Moses-support mailing list
[email protected]
http://mailman.mit.edu/mailman/listinfo/moses-support

Reply via email to