Ah, thanks.
On Tue, Aug 26, 2014 at 7:32 PM, Nan Zhu <zhunanmcg...@gmail.com> wrote: > Hi, Victor, > > the issue for you to have different version in driver and cluster is that > you the master will shutdown your application due to the inconsistent > SerialVersionID in ExecutorState > > Best, > > -- > Nan Zhu > > On Tuesday, August 26, 2014 at 10:10 PM, Matei Zaharia wrote: > > Things will definitely compile, and apps compiled on 1.0.0 should even be > able to link against 1.0.2 without recompiling. The only problem is if you > run your driver with 1.0.0 on its classpath, but the cluster has 1.0.2 in > executors. > > For Mesos and YARN vs standalone, the difference is that they just have > more features, at the expense of more complicated setup. For example, they > have richer support for cross-application sharing (see > https://spark.apache.org/docs/latest/job-scheduling.html), and the > ability to run non-Spark applications on the same cluster. > > Matei > > On August 26, 2014 at 6:53:33 PM, Victor Tso-Guillen (v...@paxata.com) > wrote: > > Yes, we are standalone right now. Do you have literature why one would > want to consider Mesos or YARN for Spark deployments? > > Sounds like I should try upgrading my project and seeing if everything > compiles without modification. Then I can connect to an existing 1.0.0 > cluster and see what what happens... > > Thanks, Matei :) > > > On Tue, Aug 26, 2014 at 6:37 PM, Matei Zaharia <matei.zaha...@gmail.com> > wrote: > > Is this a standalone mode cluster? We don't currently make this > guarantee, though it will likely work in 1.0.0 to 1.0.2. The problem though > is that the standalone mode grabs the executors' version of Spark code from > what's installed on the cluster, while your driver might be built against > another version. On YARN and Mesos, you can more easily mix different > versions of Spark, since each application ships its own Spark JAR (or > references one from a URL), and this is used for both the driver and > executors. > > Matei > > On August 26, 2014 at 6:10:57 PM, Victor Tso-Guillen (v...@paxata.com) > wrote: > > I wanted to make sure that there's full compatibility between minor > releases. I have a project that has a dependency on spark-core so that it > can be a driver program and that I can test locally. However, when > connecting to a cluster you don't necessarily know what version you're > connecting to. Is a 1.0.0 cluster binary compatible with a 1.0.2 driver > program? Is a 1.0.0 driver program binary compatible with a 1.0.2 cluster? > > > >