pyspark and R

On Mon, Apr 4, 2016 at 9:59 PM, Marcelo Vanzin <van...@cloudera.com> wrote:

> No, tests (except pyspark) should work without having to package anything
> first.
>
> On Mon, Apr 4, 2016 at 9:58 PM, Koert Kuipers <ko...@tresata.com> wrote:
> > do i need to run sbt package before doing tests?
> >
> > On Mon, Apr 4, 2016 at 11:00 PM, Marcelo Vanzin <van...@cloudera.com>
> wrote:
> >>
> >> Hey all,
> >>
> >> We merged  SPARK-13579 today, and if you're like me and have your
> >> hands automatically type "sbt assembly" anytime you're building Spark,
> >> that won't work anymore.
> >>
> >> You should now use "sbt package"; you'll still need "sbt assembly" if
> >> you require one of the remaining assemblies (streaming connectors,
> >> yarn shuffle service).
> >>
> >>
> >> --
> >> Marcelo
> >>
> >> ---------------------------------------------------------------------
> >> To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org
> >> For additional commands, e-mail: dev-h...@spark.apache.org
> >>
> >
>
>
>
> --
> Marcelo
>
> ---------------------------------------------------------------------
> To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org
> For additional commands, e-mail: dev-h...@spark.apache.org
>
>

Reply via email to