+1, let's go for the release as soon as possible, Tommaso
Il giorno gio 22 set 2016 alle ore 18:29 Mattmann, Chris A (3980) < chris.a.mattm...@jpl.nasa.gov> ha scritto: > Sounds totally cool! Great work Matt > > ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ > Chris Mattmann, Ph.D. > Chief Architect, Instrument Software and Science Data Systems Section (398) > Manager, Open Source Projects Formulation and Development Office (8212) > NASA Jet Propulsion Laboratory Pasadena, CA 91109 USA > Office: 168-519, Mailstop: 168-527 > Email: chris.a.mattm...@nasa.gov > WWW: http://sunset.usc.edu/~mattmann/ > ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ > Director, Information Retrieval and Data Science Group (IRDS) > Adjunct Associate Professor, Computer Science Department > University of Southern California, Los Angeles, CA 90089 USA > WWW: http://irds.usc.edu/ > ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ > > > > On 9/22/16, 6:09 AM, "Matt Post" <p...@cs.jhu.edu> wrote: > > Hi folks, > > I have finished the comparison. Here you can find graphs for ar-en and > ru-en. The ground-up rewrite of Moses is > about 2x–3x faster than Joshua. > > http://imgur.com/a/FcIbW > > One implication (untested) is that we are likely as fast as or faster > than Moses. > > We could brainstorm things to do to close this gap. I'd be much > happier with 2x or even 1.5x than with 3x, and I bet we could narrow this > down. But I'd like to get the 6.1 release out of the way, first, so I'm > pushing this off to next month. Sound cool? > > matt > > > > On Sep 19, 2016, at 6:26 AM, Matt Post <p...@cs.jhu.edu> wrote: > > > > I can't believe I did this, but I mis-colored one of the hiero > lines, and the Numbers legend doesn't show the line type. If you reload the > dropbox file, it's fixed now. The difference is about 3x for both. Here's > the table. > > > > Threads > > Joshua > > Moses2 > > Joshua (hiero) > > Moses2 (hiero) > > Phrase rate > > Hiero rate > > 1 > > 178 > > 65 > > 2116 > > 1137 > > 2.74 > > 1.86 > > 2 > > 109 > > 42 > > 1014 > > 389 > > 2.60 > > 2.61 > > 4 > > 78 > > 29 > > 596 > > 213 > > 2.69 > > 2.80 > > 6 > > 72 > > 25 > > 473 > > 154 > > 2.88 > > 3.07 > > > > I'll put the models together and share them later today. This was on > a 6-core machine and I agree it'd be nice to test with something much > higher. > > > > matt > > > > > >> On Sep 19, 2016, at 5:33 AM, kellen sunderland < > kellen.sunderl...@gmail.com <mailto:kellen.sunderl...@gmail.com>> wrote: > >> > >> Do we just want to store these models somewhere temporarily? I've > got a OneDrive account and could share the models from there (as long as > they're below 500GBs or so). > >> > >> On Mon, Sep 19, 2016 at 11:32 AM, kellen sunderland < > kellen.sunderl...@gmail.com <mailto:kellen.sunderl...@gmail.com>> wrote: > >> Very nice results. I think getting to within 25% of a optimized > c++ decoder from a Java decoder is impressive. Great that Hieu has put in > the work to make moses2 so fast as well, that gives organizations two quite > nice decoding engines to choose from, both with reasonable performance. > >> > >> Matt: I had a question about the x axis here. Is that number of > threads? We should be scaling more or less linearly with the number of > threads, is that the case here? If you post the models somewhere I can > also do a quick benchmark on a machine with a few more cores. > >> > >> -Kellen > >> > >> > >> On Mon, Sep 19, 2016 at 10:53 AM, Tommaso Teofili < > tommaso.teof...@gmail.com <mailto:tommaso.teof...@gmail.com>> wrote: > >> Il giorno sab 17 set 2016 alle ore 15:23 Matt Post <p...@cs.jhu.edu > <mailto:p...@cs.jhu.edu>> ha > >> scritto: > >> > >>> I'll ask Hieu; I don't anticipate any problems. One potential > problem is > >>> that that models occupy about 15--20 GB; do you think Jenkins > would host > >>> this? > >>> > >> > >> I'm not sure, can such models be downloaded and pruned at runtime, > or do > >> they need to exist on the Jenkins machine ? > >> > >> > >>> > >>> (ru-en grammars still packing, results will probably not be in > until much > >>> later today) > >>> > >>> matt > >>> > >>> > >>>> On Sep 17, 2016, at 3:19 PM, Tommaso Teofili < > tommaso.teof...@gmail.com <mailto:tommaso.teof...@gmail.com>> > >>> wrote: > >>>> > >>>> Hi Matt, > >>>> > >>>> I think it'd be really valuable if we could be able to repeat the > same > >>>> tests (given parallel corpus is available) in the future, any > chance you > >>>> can share script / code to do that ? We may even consider adding a > >>> Jenkins > >>>> job dedicated to continuously monitor performances as we work on > Joshua > >>>> master branch. > >>>> > >>>> WDYT? > >>>> > >>>> Anyway thanks for sharing the very interesting comparisons. > >>>> Regards, > >>>> Tommaso > >>>> > >>>> Il giorno sab 17 set 2016 alle ore 12:29 Matt Post < > p...@cs.jhu.edu <mailto:p...@cs.jhu.edu>> ha > >>>> scritto: > >>>> > >>>>> Ugh, I think the mailing list deleted the attachment. Here is an > attempt > >>>>> around our censors: > >>>>> > >>>>> > https://www.dropbox.com/s/80up63reu4q809y/ar-en-joshua-moses2.png?dl=0 < > https://www.dropbox.com/s/80up63reu4q809y/ar-en-joshua-moses2.png?dl=0> > >>>>> > >>>>> > >>>>>> On Sep 17, 2016, at 12:21 PM, Matt Post <p...@cs.jhu.edu > <mailto:p...@cs.jhu.edu>> wrote: > >>>>>> > >>>>>> Hi everyone, > >>>>>> > >>>>>> One thing we did this week at MT Marathon was a speed > comparison of > >>>>> Joshua 6.1 (release candidate) with Moses2, which is a ground-up > >>> rewrite of > >>>>> Moses designed for speed (see the attached paper). Moses2 is > 4–6x faster > >>>>> than Moses phrase-based, and 100x (!) faster than Moses hiero. > >>>>>> > >>>>>> I tested using two moderate-to-large sized datasets that Hieu > Hoang > >>>>> (CC'd) provided me with: ar-en and ru-en. Timing results are > from 10,000 > >>>>> sentences in each corpus. The average ar-en sentence length is > 7.5, and > >>> for > >>>>> ru-en is 28. I only ran one test for each language, so there > could be > >>> some > >>>>> variance if I averaged, but I think the results look pretty > consistent. > >>> The > >>>>> timing is end-to-end (including model load times, which Moses2 > tends to > >>> be > >>>>> a bit faster at). > >>>>>> > >>>>>> Note also that Joshua does not have lexicalized distortion, > while > >>> Moses2 > >>>>> does. This means the BLEU scores are a bit lower for Joshua: > 62.85 > >>> versus > >>>>> 63.49. This shouldn't really affect runtime, however. > >>>>>> > >>>>>> I'm working on the ru-en, but here are the ar-en results: > >>>>>> > >>>>>> > >>>>>> > >>>>>> Some conclusions: > >>>>>> > >>>>>> - Hieu has done some bang-up work on the Moses2 rewrite! Joshua > is in > >>>>> general about 3x slower than Moses2 > >>>>>> > >>>>>> - We don't have a Moses comparison, but extrapolating from > Hieu's > >>> paper, > >>>>> it seems we might be as fast as or faster than Moses phrase-based > >>> decoding, > >>>>> and are a ton faster on Hiero. I'm going to send my models to > Hieu so he > >>>>> can test on his machine, and then we'll have a better feel for > this, > >>>>> including how it scales on a machine with many more processors. > >>>>>> > >>>>>> matt > >>>>>> > >>>>>> > >>>>> > >>>>> > >>> > >>> > >> > >> > > > > > > >