Github user lianhuiwang commented on the pull request:

    https://github.com/apache/spark/pull/5580#issuecomment-96369904
  
    @andrewor14 for second question,i add two things for it.one is i add zip 
pyspark archives to pyspark/lib when we build spark jar. other is in submit if 
PYSPARK_ARCHIVES_PATH does not exist and pyspark.zip does not exist, then we 
zip archives to pyspark/lib.
    and i add a conf 'spark.submit.pyArchives' to store pyspark archives that 
let Client know pyArchives.we cannot use PYSPARK_ARCHIVES_PATH env because 
Client and spark-submit is one process and when we set PYSPARK_ARCHIVES_PATH in 
submit, Client cannot get it.
    One thing to note that if we install PYSPARKPATH on every node, now we need 
to set local pyspark archives to PYSPARK_ARCHIVES_PATH. because spark-submit 
will check whether PYSPARK_ARCHIVES_PATH exists. 


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at [email protected] or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to