Thanks JB, I've missed that one. I suggest we continue this in the ticket comments.
Thanks, Amit On Thu, Jul 7, 2016 at 11:05 PM Jean-Baptiste Onofré <[email protected]> wrote: > Hi Amit, > > I think your proposal is related to: > > https://issues.apache.org/jira/browse/BEAM-320 > > As described in the Jira, I'm planning to provide (in dedicated Maven > modules) is a Beam distribution including: > - an uber jar to wrap the dependencies > - the underlying runtime backends > - etc > > Regards > JB > > On 07/07/2016 07:49 PM, Amit Sela wrote: > > Hi everyone, > > > > Lately I've encountered a number of issues concerning the fact that the > > Spark runner does not package Spark along with it and forcing people to > do > > this on their own. > > In addition, this seems to get in the way of having beam-examples > executed > > against the Spark runner, again because it would have to add Spark > > dependencies. > > > > When running on a cluster (which I guess was the original goal here), it > is > > recommended to have Spark provided by the cluster - this makes sense for > > Spark clusters and more so for Spark + YARN clusters where you might have > > your Spark built against a specific Hadoop version or using a vendor > > distribution. > > > > In order to make the runner more accessible to new adopters, I suggest to > > consider releasing a "spark-included" artifact as well. > > > > Thoughts ? > > > > Thanks, > > Amit > > > > -- > Jean-Baptiste Onofré > [email protected] > http://blog.nanthrax.net > Talend - http://www.talend.com >
