Re: installing packages with pyspark

2016-03-19 Thread Felix Cheung
March 17, 2016 6:40 PM Subject: Re: installing packages with pyspark To: Ajinkya Kale <kaleajin...@gmail.com> Cc: <user@spark.apache.org> Hi, regarding 1, packages are resolved locally. That means that when you specify a package, spark-submit will resolv

Re: installing packages with pyspark

2016-03-19 Thread Franc Carter
ames that are Spark packages, you could also use > --packages in the command line of spark-submit or pyspark. See > http://spark.apache.org/docs/latest/submitting-applications.html > > _____ > From: Jakob Odersky <ja...@odersky.com> > Sent: T

Re: installing packages with pyspark

2016-03-19 Thread Felix Cheung
m> > Sent: Thursday, March 17, 2016 6:40 PM > Subject: Re: installing packages with pyspark > To: Ajinkya Kale <kaleajin...@gmail.com> > Cc: <user@spark.apache.org> > > > > Hi, > regarding 1, packages are resolved locally. That means that when you &g

Re: installing packages with pyspark

2016-03-19 Thread Ajinkya Kale
mitting-applications.html > > _ > From: Jakob Odersky <ja...@odersky.com> > Sent: Thursday, March 17, 2016 6:40 PM > Subject: Re: installing packages with pyspark > To: Ajinkya Kale <kaleajin...@gmail.com> > Cc: <user@spark.apache.org> > > > Hi, > re

Re: installing packages with pyspark

2016-03-19 Thread Jakob Odersky
Hi, regarding 1, packages are resolved locally. That means that when you specify a package, spark-submit will resolve the dependencies and download any jars on the local machine, before shipping* them to the cluster. So, without a priori knowledge of dataproc clusters, it should be no different to

Re: installing packages with pyspark

2016-03-19 Thread Franc Carter
test/submitting-applications.html > > _ > From: Jakob Odersky <ja...@odersky.com> > Sent: Thursday, March 17, 2016 6:40 PM > Subject: Re: installing packages with pyspark > To: Ajinkya Kale <kaleajin...@gmail.com> > Cc: <user@spark.apache.or

Re: installing packages with pyspark

2016-03-19 Thread Jakob Odersky
line of spark-submit or pyspark. See >> http://spark.apache.org/docs/latest/submitting-applications.html >> >> _ >> From: Jakob Odersky <ja...@odersky.com> >> Sent: Thursday, March 17, 2016 6:40 PM >> Subject: Re: installing pa

installing packages with pyspark

2016-03-19 Thread Ajinkya Kale
Hi all, I had couple of questions. 1. Is there documentation on how to add the graphframes or any other package for that matter on the google dataproc managed spark clusters ? 2. Is there a way to add a package to an existing pyspark context through a jupyter notebook ? --aj