pyspark and R On Mon, Apr 4, 2016 at 9:59 PM, Marcelo Vanzin <van...@cloudera.com> wrote:
> No, tests (except pyspark) should work without having to package anything > first. > > On Mon, Apr 4, 2016 at 9:58 PM, Koert Kuipers <ko...@tresata.com> wrote: > > do i need to run sbt package before doing tests? > > > > On Mon, Apr 4, 2016 at 11:00 PM, Marcelo Vanzin <van...@cloudera.com> > wrote: > >> > >> Hey all, > >> > >> We merged SPARK-13579 today, and if you're like me and have your > >> hands automatically type "sbt assembly" anytime you're building Spark, > >> that won't work anymore. > >> > >> You should now use "sbt package"; you'll still need "sbt assembly" if > >> you require one of the remaining assemblies (streaming connectors, > >> yarn shuffle service). > >> > >> > >> -- > >> Marcelo > >> > >> --------------------------------------------------------------------- > >> To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org > >> For additional commands, e-mail: dev-h...@spark.apache.org > >> > > > > > > -- > Marcelo > > --------------------------------------------------------------------- > To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org > For additional commands, e-mail: dev-h...@spark.apache.org > >