you can set property spark.jars

Manuel Sopena Ballesteros <manuel...@garvan.org.au> 于2019年11月15日周五 上午9:30写道:

> Dear zeppelin community,
>
>
>
> I need to send some parameters to pyspark so it can find extra jars.
>
>
>
> This is an example of the parameters I need to send to pyspark:
>
>
>
> pyspark \
>
>   --jars
> /share/ClusterShare/anaconda3/envs/python37/lib/python3.7/site-packages/hail/hail-all-spark.jar
> \
>
>   --conf
> spark.driver.extraClassPath=/share/ClusterShare/anaconda3/envs/python37/lib/python3.7/site-packages/hail/hail-all-spark.jar
> \
>
>   --conf
> spark.executor.extraClassPath=/share/ClusterShare/anaconda3/envs/python37/lib/python3.7/site-packages/hail/hail-all-spark.jar
> \
>
>   --conf spark.serializer=org.apache.spark.serializer.KryoSerializer \
>
>   --conf spark.kryo.registrator=is.hail.kryo.HailKryoRegistrator
>
>
>
> How could I configure my spark interpreter to do this?
>
>
>
> Thank you very much
> NOTICE
> Please consider the environment before printing this email. This message
> and any attachments are intended for the addressee named and may contain
> legally privileged/confidential/copyright information. If you are not the
> intended recipient, you should not read, use, disclose, copy or distribute
> this communication. If you have received this message in error please
> notify us at once by return email and then delete both messages. We accept
> no liability for the distribution of viruses or similar in electronic
> communications. This notice should not be removed.
>


-- 
Best Regards

Jeff Zhang

Reply via email to