hi Aboelhamd,

Your proposal looks good, I found these resource may be will be benefit.



<https://arxiv.org/pdf/1601.00710>
Multi-source *neural translation* <https://arxiv.org/abs/1601.00710>
https://arxiv.org/abs/1601.00710


<https://arxiv.org/pdf/1708.05943>
*Neural machine translation *with extended context
<https://arxiv.org/abs/1708.05943>
https://arxiv.org/abs/1708.05943

Handling homographs in *neural machine translation*
<https://arxiv.org/abs/1708.06510>https://arxiv.org/abs/1708.06510



Sevilay

On Sun, Apr 7, 2019 at 7:14 PM Aboelhamd Aly <aboelhamd.abotr...@gmail.com>
wrote:

> Hi all,
>
> I got a not solid yet idea as an alternative to yasmet and max entropy
> models.
> And it's by using neural networks to give us scores for the ambiguous
> rules.
> But I didn't yet set a formulation for the problem nor the structure of
> the inputs, output and even the goal.
> As I think there are many formulations that we can adopt.
>
> For example, the most straightforward structure, is to give the network
> all the possible combinations
> of a sentence translations and let it choose the best one, or give them
> weights.
> Hence, make the network learns which combinations to choose for a specific
> pair.
>
> Another example, is instead of building one network per pair,
> we build one network per ambiguous pattern as we did with max entropy
> models.
> So we give to the network the combinations for that pattern,
> and let it assign some weights for the ambiguous rules applied to that
> pattern.
>
> And for each structure there are many details and questions to yet answer.
>
> So with that said, I decided to look at some papers to see what others
> have done before
> to tackle some similar problems or the exact problem, and how some of them
> used machine learning
> or deep learning to solve these problems, and then try build on them.
>
> Some papers resolution was very specific to the pairs they developed, thus
> were not very important to our case. :
> 1) Resolving Structural Transfer Ambiguity inChinese-to-Korean Machine
> Translation
> <https://www.worldscientific.com/doi/10.1142/S0219427903000887>.(2003)
> 2) Arabic Machine Translation: A Developmental Perspective
> <http://www.ieee.ma/IJICT/IJICT-SI-Bouzoubaa-3.3/2%20-%20paper_farghaly.pdf>
> .(2010)
>
> Some other papers tried not to generate ambiguous rules or to minimize the
> ambiguity in transfer rules inference, and didn't provide any methods to
> resolve the ambiguity in our case. I thought that they may provide some
> help, but I think they are far from our topic :
> 1) Learning Transfer Rules for Machine Translation with Limited Data
> <http://www.cs.cmu.edu/~kathrin/ThesisSummary/ThesisSummary.pdf>.(2005)
> 2) Inferring Shallow-Transfer Machine Translation Rulesfrom Small
> Parallel Corpora <https://arxiv.org/pdf/1401.5700.pdf>.(2009)
>
> Now I am looking into some more recent papers like :
> 1) Rule Based Machine Translation Combined with Statistical Post Editor
> for Japanese to English Patent Translation
> <http://www.mt-archive.info/MTS-2007-Ehara.pdf>.(2007)
> 2) Machine translation model using inductive logic programming
> <https://scholar.cu.edu.eg/?q=shaalan/files/101.pdf>.(2009)
> 3) Machine Learning for Hybrid Machine Translation
> <https://www.aclweb.org/anthology/W12-3138.pdf>.(2012)
> 4) Study and Comparison of Rule-Based and Statistical Catalan-Spanish
> Machine Translation Systems
> <https://pdfs.semanticscholar.org/a731/0d0c15b22381c7b372e783d122a5324b005a.pdf?_ga=2.89511443.981790355.1554651923-676013054.1554651923>
> .(2012)
> 5) Latest trends in hybrid machine translation and its applications
> <https://www.sciencedirect.com/science/article/pii/S0885230814001077>
> .(2015)
> 6) Machine Translation: Phrase-Based, Rule-Based and NeuralApproaches
> with Linguistic Evaluation
> <http://www.dfki.de/~ansr01/docs/MacketanzEtAl2017_CIT.pdf>.(2017)
> 7) A Multitask-Based Neural Machine Translation Model with Part-of-Speech
> Tags Integration for Arabic Dialects
> <https://www.mdpi.com/2076-3417/8/12/2502/htm>.(2018)
>
> And I hope they give me some more insights and thoughts.
>
> --------------
>
> - So do you have recommendations to other papers that refer to the same
> problem ?
> - Also about the proposal, I modified it a little bit and share it through
> GSoC website as a draft,
>  so do you have any last feedback or thoughts about it, or do I just
> submit it as a final proposal ?
> - Last thing for the coding challenge ( integrating weighted transfer
> rules with apertium-transfer ),
>  I think it's finished, and I didn't get any feedback or response about
> it, also the pull-request is not merged yet with master.
>
>
> Thanks,
> Aboelhamd
>
>
> On Sat, Apr 6, 2019 at 5:23 AM Aboelhamd Aly <aboelhamd.abotr...@gmail.com>
> wrote:
>
>> Hi Sevilay, hi spectei,
>>
>> For sentence splitting, I think that we don't need to know neither syntax
>> nor sentence boundaries of the language.
>> Also I don't see any necessity for applying it in runtime, as in runtime
>> we only get the score of each pattern,
>> where there is no need for splitting. I also had one thought on using
>> beam-search here as I see it has no effect
>> and may be I am wrong. We can discuss in it after we close this thread.
>>
>> We will handle the whole text as one unit and will depend only on the
>> captured patterns.
>> Knowing that in the chunker terms, successive patterns that don't share a
>> transfer rule, are independent.
>> So by using the lexical form of the text, we match the words with
>> patterns, then match patterns with rules.
>> And hence we know which patterns are ambiguous and how much ambiguous
>> rules they match.
>>
>> For example if we have text with the following patterns and corresponding
>> rules numbers:
>> p1:2  p2:1  p3:6  p4:4  p5:3  p6:5  p7:1  p8:4  p9:4  p10:6  p11:8
>> p12:5  p13:5  p14:1  p15:3  p16:2
>>
>> If such text was handled by our old method with generating all the
>> combinations possible (multiplication of rules numbers),
>> we would have 82944000 possible combinations, which are not practical at
>> all to score, and take heavy computations and memory.
>> And if it is handled by our new method with applying all ambiguous rules
>> of one pattern while fixing the other patterns at LRLM rule
>> (addition of rules numbers), we will have just 60 combinations, and not
>> all of them different, giving drastically low number of combinations,
>> which may be not so representative.
>>
>> But if we apply the splitting idea , we will have something in the
>> middle, that will hopefully avoid the disadvantages of both methods
>> and benefit from advantages of both, too.
>> Let's proceed from the start of the text to the end of it, while
>> maintaining some threshold of say 24000 combinations.
>> p1 => 2  ,,  p1  p2 => 2  ,,  p1  p2  p3 => 12  ,,  p1  p2  p3  p4 => 48
>> ,,  p1  p2  p3  p4  p5 => 144  ,,
>> p1  p2  p3  p4  p5  p6 => 720  ,,  p1  p2  p3  p4  p5  p6  p7 => 720
>> p1  p2  p3  p4  p5  p6  p7 p8 => 2880  ,,  p1  p2  p3  p4  p5  p6  p7
>> p8  p9 => 11520
>>
>> And then we stop here, because taking the next pattern will exceed the
>> threshold.
>> Hence having our first split, we can now continue our work on it as usual.
>> But with more -non overwhelming- combinations which would capture more
>> semantics.
>> After that, we take the next split and so on.
>>
>> -----------
>>
>> I agree with you, that testing the current method with more than one pair
>> to know its accuracy is the priority,
>> and we currently working on it.
>>
>> -----------
>>
>> For an alternative for yasmet, I agree with spectei. Unfortunately, for
>> now I don't have a solid idea to discuss.
>> But in the few days, i will try to get one or more ideas to discuss.
>>
>>
>> On Fri, Apr 5, 2019 at 11:23 PM Francis Tyers <fty...@prompsit.com>
>> wrote:
>>
>>> El 2019-04-05 20:57, Sevilay Bayatlı escribió:
>>> > On Fri, 5 Apr 2019, 22:41 Francis Tyers, <fty...@prompsit.com> wrote:
>>> >
>>> >> El 2019-04-05 19:07, Sevilay Bayatlı escribió:
>>> >>> Hi Aboelhamd,
>>> >>>
>>> >>> There is some points in your proposal:
>>> >>>
>>> >>> First, I do not think "splitting sentence" is a good idea, each
>>> >>> language has different syntax, how could you know when you should
>>> >>> split the sentence.
>>> >>
>>> >> Apertium works on the concept of a stream of words, so in the
>>> >> runtime
>>> >> we can't really rely on robust sentence segmentation.
>>> >>
>>> >> We can often use it, e.g. for training, but if sentence boundary
>>> >> detection
>>> >> were to be included, it would need to be trained, as Sevilay hints
>>> >> at.
>>> >>
>>> >> Also, I'm not sure how much we would gain from that.
>>> >>
>>> >>> Second, "substitute yasmet with other method", I think the result
>>> >> will
>>> >>> not be more better if you substituted it with statistical method.
>>> >>>
>>> >>
>>> >> Substituting yasmet with a more up to date machine-learning method
>>> >> might be a worthwhile thing to do. What suggestions do you have?
>>> >>
>>> >> I think first we have to trying the exact method with more than 3
>>> >> language pairs and then decide  to substitute it or not, because
>>> >> what is the point of new method if dont achieve gain, then we can
>>> >> compare  the results of two methods and choose the best one. What do
>>> >> you think?
>>> >
>>>
>>> Yes, testing it with more language pairs is also a priority.
>>>
>>> Fran
>>>
>>>
>>> _______________________________________________
>>> Apertium-stuff mailing list
>>> Apertium-stuff@lists.sourceforge.net
>>> https://lists.sourceforge.net/lists/listinfo/apertium-stuff
>>>
>> _______________________________________________
> Apertium-stuff mailing list
> Apertium-stuff@lists.sourceforge.net
> https://lists.sourceforge.net/lists/listinfo/apertium-stuff
>
_______________________________________________
Apertium-stuff mailing list
Apertium-stuff@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/apertium-stuff

Reply via email to