Re: modernmt

2017-07-02 Thread John Hewitt
I found the reference for that 1,000,000 number a bit too late -- according
to this more recent paper from Koehn, it's more like 15,000,000 tokens for
NMT to meet phrase-based MT, and they omit syntax-based.

https://arxiv.org/pdf/1706.03872.pdf

-John

On Sun, Jul 2, 2017 at 12:38 PM, John Hewitt <john...@seas.upenn.edu> wrote:

> I've talked with the ModernMT people; they're well aware that they're in a
> neural MT world, and they also know that there's a sizable market for
> non-neural MT solutions.
> To back this up -- Philipp Koehn gave a talk in March on comparing
> phrase-based, syntax-based, and neural MT in low-resource settings, that
> is, when the amount of bilingual text to train on is small.
>
> Neural MT needs (if I remember correctly) about 1,000,000 tokens of
> training data to outpace syntax-based MT.
> Many language pairs (and, for that matter, domains within a single
> language pair) do not meet that requirement, and in those cases
> syntax-based MT performs best.
>
> That being said, there are some cool opportunities to combine neural and
> syntax-based MT. I can't commit the work hours right now, necessarily, but
> I've worked with xnmt <https://github.com/neulab/xnmt>, an MIT-licensed
> neural MT library that is purpose-built to be highly modular. It may offer
> some good opportunities to make an ensemble system.
>
> On Sun, Jul 2, 2017 at 4:22 AM, Tommaso Teofili <tommaso.teof...@gmail.com
> > wrote:
>
>> I think it's interesting as it extends some features that also Joshua has,
>> it's open source and has good results compared with NMT.
>>
>> Tommaso
>>
>> Il giorno sab 1 lug 2017 alle ore 18:56 Suneel Marthi <
>> suneel.mar...@gmail.com> ha scritto:
>>
>> > Is this the latest/greatest paper around MT @tommaso ??
>> >
>> > On Sat, Jul 1, 2017 at 7:55 AM, Tommaso Teofili <
>> tommaso.teof...@gmail.com
>> > >
>> > wrote:
>> >
>> > > I accidentally found the paper about mmt [1]
>> > >
>> > > [1] :
>> > > https://ufal.mff.cuni.cz/eamt2017/user-project-product-
>> > > papers/papers/user/EAMT2017_paper_88.pdf
>> > >
>> > > Il giorno gio 1 dic 2016 alle ore 22:19 Mattmann, Chris A (3010) <
>> > > chris.a.mattm...@jpl.nasa.gov> ha scritto:
>> > >
>> > > > Guys I want to point you at the DARPA D3M program:
>> > > >
>> > > > http://www.darpa.mil/program/data-driven-discovery-of-models
>> > > >
>> > > > I’m part of the Government Team for the program. This will be a good
>> > > > connection
>> > > > to have b/c it’s focused on automatically doing model and code
>> building
>> > > > for ML based
>> > > > approaches.
>> > > >
>> > > >
>> > > > ++
>> > > > Chris Mattmann, Ph.D.
>> > > > Principal Data Scientist, Engineering Administrative Office (3010)
>> > > > Manager, Open Source Projects Formulation and Development Office
>> (8212)
>> > > > NASA Jet Propulsion Laboratory Pasadena, CA 91109 USA
>> > > > Office: 180-503E, Mailstop: 180-503
>> > > > Email: chris.a.mattm...@nasa.gov
>> > > > WWW:  http://sunset.usc.edu/~mattmann/
>> > > > ++
>> > > > Director, Information Retrieval and Data Science Group (IRDS)
>> > > > Adjunct Associate Professor, Computer Science Department
>> > > > University of Southern California, Los Angeles, CA 90089 USA
>> > > > WWW: http://irds.usc.edu/
>> > > > ++
>> > > >
>> > > >
>> > > > On 12/1/16, 1:15 PM, "Matt Post" <p...@cs.jhu.edu> wrote:
>> > > >
>> > > > John,
>> > > >
>> > > > Thanks for sharing, this is really helpful. I didn't realize
>> that
>> > > > Marcello was involved.
>> > > >
>> > > > I think we can identify with the NMT danger. I still think there
>> > is a
>> > > > big niche that deep learning approaches won't reach for a few years,
>> > > until
>> > > > GPUs become super prevalent. Which is why I like ModernMT's
>> approaches,
>> > > > which overlap with many of the things I've been thinking. One thing
>> I
&g

Re: modernmt

2017-07-02 Thread John Hewitt
I've talked with the ModernMT people; they're well aware that they're in a
neural MT world, and they also know that there's a sizable market for
non-neural MT solutions.
To back this up -- Philipp Koehn gave a talk in March on comparing
phrase-based, syntax-based, and neural MT in low-resource settings, that
is, when the amount of bilingual text to train on is small.

Neural MT needs (if I remember correctly) about 1,000,000 tokens of
training data to outpace syntax-based MT.
Many language pairs (and, for that matter, domains within a single language
pair) do not meet that requirement, and in those cases syntax-based MT
performs best.

That being said, there are some cool opportunities to combine neural and
syntax-based MT. I can't commit the work hours right now, necessarily, but
I've worked with xnmt <https://github.com/neulab/xnmt>, an MIT-licensed
neural MT library that is purpose-built to be highly modular. It may offer
some good opportunities to make an ensemble system.

On Sun, Jul 2, 2017 at 4:22 AM, Tommaso Teofili <tommaso.teof...@gmail.com>
wrote:

> I think it's interesting as it extends some features that also Joshua has,
> it's open source and has good results compared with NMT.
>
> Tommaso
>
> Il giorno sab 1 lug 2017 alle ore 18:56 Suneel Marthi <
> suneel.mar...@gmail.com> ha scritto:
>
> > Is this the latest/greatest paper around MT @tommaso ??
> >
> > On Sat, Jul 1, 2017 at 7:55 AM, Tommaso Teofili <
> tommaso.teof...@gmail.com
> > >
> > wrote:
> >
> > > I accidentally found the paper about mmt [1]
> > >
> > > [1] :
> > > https://ufal.mff.cuni.cz/eamt2017/user-project-product-
> > > papers/papers/user/EAMT2017_paper_88.pdf
> > >
> > > Il giorno gio 1 dic 2016 alle ore 22:19 Mattmann, Chris A (3010) <
> > > chris.a.mattm...@jpl.nasa.gov> ha scritto:
> > >
> > > > Guys I want to point you at the DARPA D3M program:
> > > >
> > > > http://www.darpa.mil/program/data-driven-discovery-of-models
> > > >
> > > > I’m part of the Government Team for the program. This will be a good
> > > > connection
> > > > to have b/c it’s focused on automatically doing model and code
> building
> > > > for ML based
> > > > approaches.
> > > >
> > > >
> > > > ++
> > > > Chris Mattmann, Ph.D.
> > > > Principal Data Scientist, Engineering Administrative Office (3010)
> > > > Manager, Open Source Projects Formulation and Development Office
> (8212)
> > > > NASA Jet Propulsion Laboratory Pasadena, CA 91109 USA
> > > > Office: 180-503E, Mailstop: 180-503
> > > > Email: chris.a.mattm...@nasa.gov
> > > > WWW:  http://sunset.usc.edu/~mattmann/
> > > > ++
> > > > Director, Information Retrieval and Data Science Group (IRDS)
> > > > Adjunct Associate Professor, Computer Science Department
> > > > University of Southern California, Los Angeles, CA 90089 USA
> > > > WWW: http://irds.usc.edu/
> > > > ++
> > > >
> > > >
> > > > On 12/1/16, 1:15 PM, "Matt Post" <p...@cs.jhu.edu> wrote:
> > > >
> > > > John,
> > > >
> > > > Thanks for sharing, this is really helpful. I didn't realize that
> > > > Marcello was involved.
> > > >
> > > > I think we can identify with the NMT danger. I still think there
> > is a
> > > > big niche that deep learning approaches won't reach for a few years,
> > > until
> > > > GPUs become super prevalent. Which is why I like ModernMT's
> approaches,
> > > > which overlap with many of the things I've been thinking. One thing I
> > > > really like is there automatic context-switching approach. This is a
> > > great
> > > > way to build general-purpose models, and I'd like to mimic it. I have
> > > some
> > > > general ideas about how this should be implemented but am also
> looking
> > > into
> > > > the literature here.
> > > >
> > > > matt
> > > >
> > > >
> > > > > On Dec 1, 2016, at 1:46 PM, John Hewitt <
> john...@seas.upenn.edu>
> > > > wrote:
> > > > >
> > > > > I had a few good conversations over dinner with this team at
> AMT

Re: modernmt

2017-07-02 Thread Tommaso Teofili
I think it's interesting as it extends some features that also Joshua has,
it's open source and has good results compared with NMT.

Tommaso

Il giorno sab 1 lug 2017 alle ore 18:56 Suneel Marthi <
suneel.mar...@gmail.com> ha scritto:

> Is this the latest/greatest paper around MT @tommaso ??
>
> On Sat, Jul 1, 2017 at 7:55 AM, Tommaso Teofili <tommaso.teof...@gmail.com
> >
> wrote:
>
> > I accidentally found the paper about mmt [1]
> >
> > [1] :
> > https://ufal.mff.cuni.cz/eamt2017/user-project-product-
> > papers/papers/user/EAMT2017_paper_88.pdf
> >
> > Il giorno gio 1 dic 2016 alle ore 22:19 Mattmann, Chris A (3010) <
> > chris.a.mattm...@jpl.nasa.gov> ha scritto:
> >
> > > Guys I want to point you at the DARPA D3M program:
> > >
> > > http://www.darpa.mil/program/data-driven-discovery-of-models
> > >
> > > I’m part of the Government Team for the program. This will be a good
> > > connection
> > > to have b/c it’s focused on automatically doing model and code building
> > > for ML based
> > > approaches.
> > >
> > >
> > > ++
> > > Chris Mattmann, Ph.D.
> > > Principal Data Scientist, Engineering Administrative Office (3010)
> > > Manager, Open Source Projects Formulation and Development Office (8212)
> > > NASA Jet Propulsion Laboratory Pasadena, CA 91109 USA
> > > Office: 180-503E, Mailstop: 180-503
> > > Email: chris.a.mattm...@nasa.gov
> > > WWW:  http://sunset.usc.edu/~mattmann/
> > > ++
> > > Director, Information Retrieval and Data Science Group (IRDS)
> > > Adjunct Associate Professor, Computer Science Department
> > > University of Southern California, Los Angeles, CA 90089 USA
> > > WWW: http://irds.usc.edu/
> > > ++
> > >
> > >
> > > On 12/1/16, 1:15 PM, "Matt Post" <p...@cs.jhu.edu> wrote:
> > >
> > > John,
> > >
> > > Thanks for sharing, this is really helpful. I didn't realize that
> > > Marcello was involved.
> > >
> > > I think we can identify with the NMT danger. I still think there
> is a
> > > big niche that deep learning approaches won't reach for a few years,
> > until
> > > GPUs become super prevalent. Which is why I like ModernMT's approaches,
> > > which overlap with many of the things I've been thinking. One thing I
> > > really like is there automatic context-switching approach. This is a
> > great
> > > way to build general-purpose models, and I'd like to mimic it. I have
> > some
> > > general ideas about how this should be implemented but am also looking
> > into
> > > the literature here.
> > >
> > > matt
> > >
> > >
> > > > On Dec 1, 2016, at 1:46 PM, John Hewitt <john...@seas.upenn.edu>
> > > wrote:
> > > >
> > > > I had a few good conversations over dinner with this team at AMTA
> > in
> > > Austin
> > > > in October.
> > > > They seem to be in the interesting position where their work is
> > > good, but
> > > > is in danger of being superseded by neural MT as they come out of
> > > the gate.
> > > > Clearly, it has benefits over NMT, and is easier to adopt, but
> may
> > > not be
> > > > the winner over the long run.
> > > >
> > > > Here's the link
> > > > <
> > > https://amtaweb.org/wp-content/uploads/2016/11/MMT_
> > Tutorial_FedericoTrombetti_wide-cover.pdf
> > > >
> > > > to their AMTA tutorial.
> > > >
> > > > -John
> > > >
> > > > On Thu, Dec 1, 2016 at 10:17 AM, Mattmann, Chris A (3010) <
> > > > chris.a.mattm...@jpl.nasa.gov> wrote:
> > > >
> > > >> Wow seems like this kind of overlaps with BigTranslate as well..
> > > thanks
> > > >> for passing
> > > >> along Matt
> > > >>
> > > >> 
> > ++
> > > >> Chris Mattmann, Ph.D.
> > > >> Principal Data Scientist, Engineering Administrative Office
> (3010)
> > > >> Ma

Re: modernmt

2017-07-01 Thread Suneel Marthi
Is this the latest/greatest paper around MT @tommaso ??

On Sat, Jul 1, 2017 at 7:55 AM, Tommaso Teofili <tommaso.teof...@gmail.com>
wrote:

> I accidentally found the paper about mmt [1]
>
> [1] :
> https://ufal.mff.cuni.cz/eamt2017/user-project-product-
> papers/papers/user/EAMT2017_paper_88.pdf
>
> Il giorno gio 1 dic 2016 alle ore 22:19 Mattmann, Chris A (3010) <
> chris.a.mattm...@jpl.nasa.gov> ha scritto:
>
> > Guys I want to point you at the DARPA D3M program:
> >
> > http://www.darpa.mil/program/data-driven-discovery-of-models
> >
> > I’m part of the Government Team for the program. This will be a good
> > connection
> > to have b/c it’s focused on automatically doing model and code building
> > for ML based
> > approaches.
> >
> >
> > ++
> > Chris Mattmann, Ph.D.
> > Principal Data Scientist, Engineering Administrative Office (3010)
> > Manager, Open Source Projects Formulation and Development Office (8212)
> > NASA Jet Propulsion Laboratory Pasadena, CA 91109 USA
> > Office: 180-503E, Mailstop: 180-503
> > Email: chris.a.mattm...@nasa.gov
> > WWW:  http://sunset.usc.edu/~mattmann/
> > ++
> > Director, Information Retrieval and Data Science Group (IRDS)
> > Adjunct Associate Professor, Computer Science Department
> > University of Southern California, Los Angeles, CA 90089 USA
> > WWW: http://irds.usc.edu/
> > ++
> >
> >
> > On 12/1/16, 1:15 PM, "Matt Post" <p...@cs.jhu.edu> wrote:
> >
> > John,
> >
> > Thanks for sharing, this is really helpful. I didn't realize that
> > Marcello was involved.
> >
> > I think we can identify with the NMT danger. I still think there is a
> > big niche that deep learning approaches won't reach for a few years,
> until
> > GPUs become super prevalent. Which is why I like ModernMT's approaches,
> > which overlap with many of the things I've been thinking. One thing I
> > really like is there automatic context-switching approach. This is a
> great
> > way to build general-purpose models, and I'd like to mimic it. I have
> some
> > general ideas about how this should be implemented but am also looking
> into
> > the literature here.
> >
> > matt
> >
> >
> > > On Dec 1, 2016, at 1:46 PM, John Hewitt <john...@seas.upenn.edu>
> > wrote:
> > >
> > > I had a few good conversations over dinner with this team at AMTA
> in
> > Austin
> > > in October.
> > > They seem to be in the interesting position where their work is
> > good, but
> > > is in danger of being superseded by neural MT as they come out of
> > the gate.
> > > Clearly, it has benefits over NMT, and is easier to adopt, but may
> > not be
> > > the winner over the long run.
> > >
> > > Here's the link
> > > <
> > https://amtaweb.org/wp-content/uploads/2016/11/MMT_
> Tutorial_FedericoTrombetti_wide-cover.pdf
> > >
> > > to their AMTA tutorial.
> > >
> > > -John
> > >
> > > On Thu, Dec 1, 2016 at 10:17 AM, Mattmann, Chris A (3010) <
> > > chris.a.mattm...@jpl.nasa.gov> wrote:
> > >
> > >> Wow seems like this kind of overlaps with BigTranslate as well..
> > thanks
> > >> for passing
> > >> along Matt
> > >>
> > >> 
> ++
> > >> Chris Mattmann, Ph.D.
> > >> Principal Data Scientist, Engineering Administrative Office (3010)
> > >> Manager, Open Source Projects Formulation and Development Office
> > (8212)
> > >> NASA Jet Propulsion Laboratory Pasadena, CA 91109 USA
> > >> Office: 180-503E, Mailstop: 180-503
> > >> Email: chris.a.mattm...@nasa.gov
> > >> WWW:  http://sunset.usc.edu/~mattmann/
> > >> 
> ++
> > >> Director, Information Retrieval and Data Science Group (IRDS)
> > >> Adjunct Associate Professor, Computer Science Department
> > >> University of Southern California, Los Angeles, CA 90089 USA
> > >> WWW: http://irds.usc.edu/
> > >> 
> ++
> > >>
> > >>
> > >> On 12/1/16, 4:47 AM, "Matt Post" <p...@cs.jhu.edu> wrote:
> > >>
> > >>Just came across this, and it's really cool:
> > >>
> > >>https://github.com/ModernMT/MMT
> > >>
> > >>See the README for some great use cases. I'm surprised I'd
> never
> > heard
> > >> of this before as it's EU funded and associated with U Edinburgh.
> > >>
> > >>
> >
> >
> >
> >
>


Re: modernmt

2017-07-01 Thread Chris Mattmann
Thanks for sharing Tomasso!



On 7/1/17, 4:55 AM, "Tommaso Teofili" <tommaso.teof...@gmail.com> wrote:

I accidentally found the paper about mmt [1]

[1] :

https://ufal.mff.cuni.cz/eamt2017/user-project-product-papers/papers/user/EAMT2017_paper_88.pdf

Il giorno gio 1 dic 2016 alle ore 22:19 Mattmann, Chris A (3010) <
chris.a.mattm...@jpl.nasa.gov> ha scritto:

> Guys I want to point you at the DARPA D3M program:
>
> http://www.darpa.mil/program/data-driven-discovery-of-models
>
> I’m part of the Government Team for the program. This will be a good
> connection
> to have b/c it’s focused on automatically doing model and code building
> for ML based
> approaches.
>
>
> ++
> Chris Mattmann, Ph.D.
> Principal Data Scientist, Engineering Administrative Office (3010)
> Manager, Open Source Projects Formulation and Development Office (8212)
> NASA Jet Propulsion Laboratory Pasadena, CA 91109 USA
> Office: 180-503E, Mailstop: 180-503
> Email: chris.a.mattm...@nasa.gov
> WWW:  http://sunset.usc.edu/~mattmann/
> ++
> Director, Information Retrieval and Data Science Group (IRDS)
> Adjunct Associate Professor, Computer Science Department
> University of Southern California, Los Angeles, CA 90089 USA
> WWW: http://irds.usc.edu/
> ++
>
>
> On 12/1/16, 1:15 PM, "Matt Post" <p...@cs.jhu.edu> wrote:
>
> John,
>
> Thanks for sharing, this is really helpful. I didn't realize that
> Marcello was involved.
>
> I think we can identify with the NMT danger. I still think there is a
> big niche that deep learning approaches won't reach for a few years, until
> GPUs become super prevalent. Which is why I like ModernMT's approaches,
> which overlap with many of the things I've been thinking. One thing I
> really like is there automatic context-switching approach. This is a great
> way to build general-purpose models, and I'd like to mimic it. I have some
> general ideas about how this should be implemented but am also looking 
into
> the literature here.
>
> matt
>
>
> > On Dec 1, 2016, at 1:46 PM, John Hewitt <john...@seas.upenn.edu>
> wrote:
> >
> > I had a few good conversations over dinner with this team at AMTA in
> Austin
> > in October.
> > They seem to be in the interesting position where their work is
> good, but
> > is in danger of being superseded by neural MT as they come out of
> the gate.
> > Clearly, it has benefits over NMT, and is easier to adopt, but may
> not be
> > the winner over the long run.
> >
> > Here's the link
> > <
> 
https://amtaweb.org/wp-content/uploads/2016/11/MMT_Tutorial_FedericoTrombetti_wide-cover.pdf
> >
> > to their AMTA tutorial.
> >
> > -John
> >
> > On Thu, Dec 1, 2016 at 10:17 AM, Mattmann, Chris A (3010) <
> > chris.a.mattm...@jpl.nasa.gov> wrote:
> >
> >> Wow seems like this kind of overlaps with BigTranslate as well..
> thanks
> >> for passing
> >> along Matt
> >>
> >> ++
> >> Chris Mattmann, Ph.D.
> >> Principal Data Scientist, Engineering Administrative Office (3010)
> >> Manager, Open Source Projects Formulation and Development Office
> (8212)
> >> NASA Jet Propulsion Laboratory Pasadena, CA 91109 USA
> >> Office: 180-503E, Mailstop: 180-503
> >> Email: chris.a.mattm...@nasa.gov
> >> WWW:  http://sunset.usc.edu/~mattmann/
> >> ++
> >> Director, Information Retrieval and Data Science Group (IRDS)
> >> Adjunct Associate Professor, Computer Science Department
> >> University of Southern California, Los Angeles, CA 90089 USA
> >> WWW: http://irds.usc.edu/
> >> ++
> >>
> >>
> >> On 12/1/16, 4:47 AM, "Matt Post" <p...@cs.jhu.edu> wrote:
> >>
> >>Just came across this, and it's really cool:
> >>
> >>https://github.com/ModernMT/MMT
> >>
> >>See the README for some great use cases. I'm surprised I'd never
> heard
> >> of this before as it's EU funded and associated with U Edinburgh.
> >>
> >>
>
>
>
>





Re: modernmt

2017-07-01 Thread Tommaso Teofili
I accidentally found the paper about mmt [1]

[1] :
https://ufal.mff.cuni.cz/eamt2017/user-project-product-papers/papers/user/EAMT2017_paper_88.pdf

Il giorno gio 1 dic 2016 alle ore 22:19 Mattmann, Chris A (3010) <
chris.a.mattm...@jpl.nasa.gov> ha scritto:

> Guys I want to point you at the DARPA D3M program:
>
> http://www.darpa.mil/program/data-driven-discovery-of-models
>
> I’m part of the Government Team for the program. This will be a good
> connection
> to have b/c it’s focused on automatically doing model and code building
> for ML based
> approaches.
>
>
> ++
> Chris Mattmann, Ph.D.
> Principal Data Scientist, Engineering Administrative Office (3010)
> Manager, Open Source Projects Formulation and Development Office (8212)
> NASA Jet Propulsion Laboratory Pasadena, CA 91109 USA
> Office: 180-503E, Mailstop: 180-503
> Email: chris.a.mattm...@nasa.gov
> WWW:  http://sunset.usc.edu/~mattmann/
> ++
> Director, Information Retrieval and Data Science Group (IRDS)
> Adjunct Associate Professor, Computer Science Department
> University of Southern California, Los Angeles, CA 90089 USA
> WWW: http://irds.usc.edu/
> ++
>
>
> On 12/1/16, 1:15 PM, "Matt Post" <p...@cs.jhu.edu> wrote:
>
> John,
>
> Thanks for sharing, this is really helpful. I didn't realize that
> Marcello was involved.
>
> I think we can identify with the NMT danger. I still think there is a
> big niche that deep learning approaches won't reach for a few years, until
> GPUs become super prevalent. Which is why I like ModernMT's approaches,
> which overlap with many of the things I've been thinking. One thing I
> really like is there automatic context-switching approach. This is a great
> way to build general-purpose models, and I'd like to mimic it. I have some
> general ideas about how this should be implemented but am also looking into
> the literature here.
>
> matt
>
>
> > On Dec 1, 2016, at 1:46 PM, John Hewitt <john...@seas.upenn.edu>
> wrote:
> >
> > I had a few good conversations over dinner with this team at AMTA in
> Austin
> > in October.
> > They seem to be in the interesting position where their work is
> good, but
> > is in danger of being superseded by neural MT as they come out of
> the gate.
> > Clearly, it has benefits over NMT, and is easier to adopt, but may
> not be
> > the winner over the long run.
> >
> > Here's the link
> > <
> https://amtaweb.org/wp-content/uploads/2016/11/MMT_Tutorial_FedericoTrombetti_wide-cover.pdf
> >
> > to their AMTA tutorial.
> >
> > -John
> >
> > On Thu, Dec 1, 2016 at 10:17 AM, Mattmann, Chris A (3010) <
> > chris.a.mattm...@jpl.nasa.gov> wrote:
> >
> >> Wow seems like this kind of overlaps with BigTranslate as well..
> thanks
> >> for passing
> >> along Matt
> >>
> >> ++
> >> Chris Mattmann, Ph.D.
> >> Principal Data Scientist, Engineering Administrative Office (3010)
> >> Manager, Open Source Projects Formulation and Development Office
> (8212)
> >> NASA Jet Propulsion Laboratory Pasadena, CA 91109 USA
> >> Office: 180-503E, Mailstop: 180-503
> >> Email: chris.a.mattm...@nasa.gov
> >> WWW:  http://sunset.usc.edu/~mattmann/
> >> ++
> >> Director, Information Retrieval and Data Science Group (IRDS)
> >> Adjunct Associate Professor, Computer Science Department
> >> University of Southern California, Los Angeles, CA 90089 USA
> >> WWW: http://irds.usc.edu/
> >> ++
> >>
> >>
> >> On 12/1/16, 4:47 AM, "Matt Post" <p...@cs.jhu.edu> wrote:
> >>
> >>Just came across this, and it's really cool:
> >>
> >>https://github.com/ModernMT/MMT
> >>
> >>See the README for some great use cases. I'm surprised I'd never
> heard
> >> of this before as it's EU funded and associated with U Edinburgh.
> >>
> >>
>
>
>
>


Re: modernmt

2016-12-01 Thread Matt Post
John,

Thanks for sharing, this is really helpful. I didn't realize that Marcello was 
involved.

I think we can identify with the NMT danger. I still think there is a big niche 
that deep learning approaches won't reach for a few years, until GPUs become 
super prevalent. Which is why I like ModernMT's approaches, which overlap with 
many of the things I've been thinking. One thing I really like is there 
automatic context-switching approach. This is a great way to build 
general-purpose models, and I'd like to mimic it. I have some general ideas 
about how this should be implemented but am also looking into the literature 
here.

matt


> On Dec 1, 2016, at 1:46 PM, John Hewitt <john...@seas.upenn.edu> wrote:
> 
> I had a few good conversations over dinner with this team at AMTA in Austin
> in October.
> They seem to be in the interesting position where their work is good, but
> is in danger of being superseded by neural MT as they come out of the gate.
> Clearly, it has benefits over NMT, and is easier to adopt, but may not be
> the winner over the long run.
> 
> Here's the link
> <https://amtaweb.org/wp-content/uploads/2016/11/MMT_Tutorial_FedericoTrombetti_wide-cover.pdf>
> to their AMTA tutorial.
> 
> -John
> 
> On Thu, Dec 1, 2016 at 10:17 AM, Mattmann, Chris A (3010) <
> chris.a.mattm...@jpl.nasa.gov> wrote:
> 
>> Wow seems like this kind of overlaps with BigTranslate as well.. thanks
>> for passing
>> along Matt
>> 
>> ++
>> Chris Mattmann, Ph.D.
>> Principal Data Scientist, Engineering Administrative Office (3010)
>> Manager, Open Source Projects Formulation and Development Office (8212)
>> NASA Jet Propulsion Laboratory Pasadena, CA 91109 USA
>> Office: 180-503E, Mailstop: 180-503
>> Email: chris.a.mattm...@nasa.gov
>> WWW:  http://sunset.usc.edu/~mattmann/
>> ++
>> Director, Information Retrieval and Data Science Group (IRDS)
>> Adjunct Associate Professor, Computer Science Department
>> University of Southern California, Los Angeles, CA 90089 USA
>> WWW: http://irds.usc.edu/
>> ++
>> 
>> 
>> On 12/1/16, 4:47 AM, "Matt Post" <p...@cs.jhu.edu> wrote:
>> 
>>Just came across this, and it's really cool:
>> 
>>https://github.com/ModernMT/MMT
>> 
>>See the README for some great use cases. I'm surprised I'd never heard
>> of this before as it's EU funded and associated with U Edinburgh.
>> 
>> 



Re: modernmt

2016-12-01 Thread John Hewitt
I had a few good conversations over dinner with this team at AMTA in Austin
in October.
They seem to be in the interesting position where their work is good, but
is in danger of being superseded by neural MT as they come out of the gate.
Clearly, it has benefits over NMT, and is easier to adopt, but may not be
the winner over the long run.

Here's the link
<https://amtaweb.org/wp-content/uploads/2016/11/MMT_Tutorial_FedericoTrombetti_wide-cover.pdf>
to their AMTA tutorial.

-John

On Thu, Dec 1, 2016 at 10:17 AM, Mattmann, Chris A (3010) <
chris.a.mattm...@jpl.nasa.gov> wrote:

> Wow seems like this kind of overlaps with BigTranslate as well.. thanks
> for passing
> along Matt
>
> ++
> Chris Mattmann, Ph.D.
> Principal Data Scientist, Engineering Administrative Office (3010)
> Manager, Open Source Projects Formulation and Development Office (8212)
> NASA Jet Propulsion Laboratory Pasadena, CA 91109 USA
> Office: 180-503E, Mailstop: 180-503
> Email: chris.a.mattm...@nasa.gov
> WWW:  http://sunset.usc.edu/~mattmann/
> ++
> Director, Information Retrieval and Data Science Group (IRDS)
> Adjunct Associate Professor, Computer Science Department
> University of Southern California, Los Angeles, CA 90089 USA
> WWW: http://irds.usc.edu/
> ++
>
>
> On 12/1/16, 4:47 AM, "Matt Post" <p...@cs.jhu.edu> wrote:
>
> Just came across this, and it's really cool:
>
> https://github.com/ModernMT/MMT
>
> See the README for some great use cases. I'm surprised I'd never heard
> of this before as it's EU funded and associated with U Edinburgh.
>
>


Re: modernmt

2016-12-01 Thread Mattmann, Chris A (3010)
Wow seems like this kind of overlaps with BigTranslate as well.. thanks for 
passing
along Matt

++
Chris Mattmann, Ph.D.
Principal Data Scientist, Engineering Administrative Office (3010)
Manager, Open Source Projects Formulation and Development Office (8212)
NASA Jet Propulsion Laboratory Pasadena, CA 91109 USA
Office: 180-503E, Mailstop: 180-503
Email: chris.a.mattm...@nasa.gov
WWW:  http://sunset.usc.edu/~mattmann/
++
Director, Information Retrieval and Data Science Group (IRDS)
Adjunct Associate Professor, Computer Science Department
University of Southern California, Los Angeles, CA 90089 USA
WWW: http://irds.usc.edu/
++
 

On 12/1/16, 4:47 AM, "Matt Post" <p...@cs.jhu.edu> wrote:

Just came across this, and it's really cool:

https://github.com/ModernMT/MMT

See the README for some great use cases. I'm surprised I'd never heard of 
this before as it's EU funded and associated with U Edinburgh.



Re: modernmt

2016-12-01 Thread Tommaso Teofili
wow I had never heard of it either, but I do know one of the committers :)

Il giorno gio 1 dic 2016 alle ore 13:49 Matt Post <p...@cs.jhu.edu> ha
scritto:

> Just came across this, and it's really cool:
>
> https://github.com/ModernMT/MMT
>
> See the README for some great use cases. I'm surprised I'd never heard of
> this before as it's EU funded and associated with U Edinburgh.


modernmt

2016-12-01 Thread Matt Post
Just came across this, and it's really cool:

https://github.com/ModernMT/MMT

See the README for some great use cases. I'm surprised I'd never heard of this 
before as it's EU funded and associated with U Edinburgh.