Re: modernmt

2017-07-01 Thread Suneel Marthi
Is this the latest/greatest paper around MT @tommaso ??

On Sat, Jul 1, 2017 at 7:55 AM, Tommaso Teofili 
wrote:

> I accidentally found the paper about mmt [1]
>
> [1] :
> https://ufal.mff.cuni.cz/eamt2017/user-project-product-
> papers/papers/user/EAMT2017_paper_88.pdf
>
> Il giorno gio 1 dic 2016 alle ore 22:19 Mattmann, Chris A (3010) <
> chris.a.mattm...@jpl.nasa.gov> ha scritto:
>
> > Guys I want to point you at the DARPA D3M program:
> >
> > http://www.darpa.mil/program/data-driven-discovery-of-models
> >
> > I’m part of the Government Team for the program. This will be a good
> > connection
> > to have b/c it’s focused on automatically doing model and code building
> > for ML based
> > approaches.
> >
> >
> > ++
> > Chris Mattmann, Ph.D.
> > Principal Data Scientist, Engineering Administrative Office (3010)
> > Manager, Open Source Projects Formulation and Development Office (8212)
> > NASA Jet Propulsion Laboratory Pasadena, CA 91109 USA
> > Office: 180-503E, Mailstop: 180-503
> > Email: chris.a.mattm...@nasa.gov
> > WWW:  http://sunset.usc.edu/~mattmann/
> > ++
> > Director, Information Retrieval and Data Science Group (IRDS)
> > Adjunct Associate Professor, Computer Science Department
> > University of Southern California, Los Angeles, CA 90089 USA
> > WWW: http://irds.usc.edu/
> > ++
> >
> >
> > On 12/1/16, 1:15 PM, "Matt Post"  wrote:
> >
> > John,
> >
> > Thanks for sharing, this is really helpful. I didn't realize that
> > Marcello was involved.
> >
> > I think we can identify with the NMT danger. I still think there is a
> > big niche that deep learning approaches won't reach for a few years,
> until
> > GPUs become super prevalent. Which is why I like ModernMT's approaches,
> > which overlap with many of the things I've been thinking. One thing I
> > really like is there automatic context-switching approach. This is a
> great
> > way to build general-purpose models, and I'd like to mimic it. I have
> some
> > general ideas about how this should be implemented but am also looking
> into
> > the literature here.
> >
> > matt
> >
> >
> > > On Dec 1, 2016, at 1:46 PM, John Hewitt 
> > wrote:
> > >
> > > I had a few good conversations over dinner with this team at AMTA
> in
> > Austin
> > > in October.
> > > They seem to be in the interesting position where their work is
> > good, but
> > > is in danger of being superseded by neural MT as they come out of
> > the gate.
> > > Clearly, it has benefits over NMT, and is easier to adopt, but may
> > not be
> > > the winner over the long run.
> > >
> > > Here's the link
> > > <
> > https://amtaweb.org/wp-content/uploads/2016/11/MMT_
> Tutorial_FedericoTrombetti_wide-cover.pdf
> > >
> > > to their AMTA tutorial.
> > >
> > > -John
> > >
> > > On Thu, Dec 1, 2016 at 10:17 AM, Mattmann, Chris A (3010) <
> > > chris.a.mattm...@jpl.nasa.gov> wrote:
> > >
> > >> Wow seems like this kind of overlaps with BigTranslate as well..
> > thanks
> > >> for passing
> > >> along Matt
> > >>
> > >> 
> ++
> > >> Chris Mattmann, Ph.D.
> > >> Principal Data Scientist, Engineering Administrative Office (3010)
> > >> Manager, Open Source Projects Formulation and Development Office
> > (8212)
> > >> NASA Jet Propulsion Laboratory Pasadena, CA 91109 USA
> > >> Office: 180-503E, Mailstop: 180-503
> > >> Email: chris.a.mattm...@nasa.gov
> > >> WWW:  http://sunset.usc.edu/~mattmann/
> > >> 
> ++
> > >> Director, Information Retrieval and Data Science Group (IRDS)
> > >> Adjunct Associate Professor, Computer Science Department
> > >> University of Southern California, Los Angeles, CA 90089 USA
> > >> WWW: http://irds.usc.edu/
> > >> 
> ++
> > >>
> > >>
> > >> On 12/1/16, 4:47 AM, "Matt Post"  wrote:
> > >>
> > >>Just came across this, and it's really cool:
> > >>
> > >>https://github.com/ModernMT/MMT
> > >>
> > >>See the README for some great use cases. I'm surprised I'd
> never
> > heard
> > >> of this before as it's EU funded and associated with U Edinburgh.
> > >>
> > >>
> >
> >
> >
> >
>


Re: [ANNOUNCE] - Apache Joshua 6.1 incubating release

2017-06-22 Thread Suneel Marthi
Congrats on the release.

I have been a silent lurker on this channel since I first heard of Joshua
last September at Amazon, Berlin.

Tommaso and myself recently did a talk at Berlin Buzzwords 2017 -
'Embracing Diversity - searching over multiple languages' [1]
using Apache Joshua for Machine Translation, and Apache OpenNLP for
Language detection.

I have been wondering how much of the present VLPS can be replaced by
OpenNLP with Flink/Beam pipelines.
I did a talk last week at Hadoop Summit, San Jose about 'Large Scale Text
processing with Apache OpenNLP and Apache Flink [2].

Also that Thrax which is presently MapReduce based, can definitely be
ported over to modern streaming distributed frameworks like Flink/Kafka
Streams/Beam.


[1]
https://www.youtube.com/watch?v=ZrWxySF-9KY=20=2s=PLq-odUc2x7i-9Nijx-WfoRMoAfHC9XzTt
[2] https://www.slideshare.net/SuneelMarthi/large-scale-text-processing


Regards,
Suneel

On Fri, Jun 23, 2017 at 1:44 AM, Tommaso Teofili 
wrote:

> Il giorno gio 22 giu 2017 alle ore 18:31 John Hewitt <
> john...@seas.upenn.edu>
> ha scritto:
>
> > Related note: I've begun to announce to the Penn NLP communities; I can
> > talk to Mark Liberman at the LDC about getting a note in there as well.
> >
>
> cool, thanks John!
>
>
> >
> > -John
> >
> > On Thu, Jun 22, 2017 at 10:11 AM, lewis john mcgibbney <
> lewi...@apache.org
> > >
> > wrote:
> >
> > > Hi Tommaso,
> > > EXCELLENT :)
> > > @Matt are you able to Tweet this out and make some tags?
> > > @Tommaso, where else did you announce this? Is it possible for us to
> make
> > > some more noise on various other communication forums/channels?
> >
>
> I am not sure where else it'd make sense to spread the word, perhaps some
> "candidates" could be Lucene or OpenNLP lists.
>
>
> > > This is brilliant news. Thank you Tommaso for being persistent with the
> > > release process, I am glad that we were able to recover the artifacts.
> > > Lewis
> >
>
> yeah, thanks a lot for your help there Lewis.
>
> Regards,
> Tommaso
>
>
> > >
> > > On Thu, Jun 22, 2017 at 5:55 AM, <
> > > dev-digest-h...@joshua.incubator.apache.org> wrote:
> > >
> > > >
> > > > From: Tommaso Teofili 
> > > > To: annou...@apache.org
> > > > Cc: "dev@joshua.incubator.apache.org"  org
> > >
> > > > Bcc:
> > > > Date: Thu, 22 Jun 2017 12:54:49 +
> > > > Subject: [ANNOUNCE] - Apache Joshua 6.1 incubating release
> > > > Hi Folks,
> > > >
> > > > The Apache Joshua team (PPMC) is pleased to announce the immediate
> > > > availability of Apache Joshua 6.1 (incubating).
> > > >
> > > > Apache Joshua is a statistical machine translation decoder for
> > > > phrase-based, hierarchical, and syntax-based machine translation,
> > written
> > > > in Java.
> > > >
> > > > Apache Joshua is released as both source code, downloads for which
> can
> > be
> > > > found at ASF dist download site [0] as well as Maven artifacts which
> > can
> > > be
> > > > found on Maven central [1].
> > > >
> > > > The full Jira release report can be found here [3].
> > > >
> > > > Thank you,
> > > > Tommaso (on behalf of Apache Joshua PPMC)
> > > >
> > > > — DISCLAIMER Apache Joshua is an effort undergoing incubation at The
> > > Apache
> > > > Software Foundation (ASF), sponsored by the Apache Incubator PMC.
> > > > Incubation is required of all newly accepted projects until a further
> > > > review indicates that the infrastructure, communications, and
> decision
> > > > making process have stabilized in a manner consistent with other
> > > successful
> > > > ASF projects. While incubation status is not necessarily a reflection
> > of
> > > > the completeness or stability of the code,it does indicate that the
> > > project
> > > > has yet to be fully endorsed by the ASF.
> > > >
> > > > [0] http://apache.org/dist/incubator/joshua/6.1/
> > > > [1] http://search.maven.org/#search|ga|1|g%3A%22org.apache.joshua%22
> > 
> > > > [3]
> > > > https://issues.apache.org/jira/secure/ReleaseNote.jspa?
> > > > projectId=12319720=12335049
> > > >
> > > >
> > >
> > >
> > > --
> > > http://home.apache.org/~lewismc/
> > > @hectorMcSpector
> > > http://www.linkedin.com/in/lmcgibbney
> > >
> >
>