Github user yanboliang commented on the issue:

    https://github.com/apache/spark/pull/16214
  
    @mengxr I think there is no (at least I did not found) concurrency issue 
when multiple executors on the same machine, since native R has lock to protect 
this. You can verify it by installing package to a shared directory in a single 
node in multi-thread way. However, we should recommend to install package to an 
executor associated directory which means different executor has its own R lib 
directory even if they are in the same machine. For YARN mode, it works well 
since each executor has its own R lib directory natively. 
    @felixcheung I updated the examples and add corresponding session in SparkR 
user guide. I'm not sure whether such an example is appropriate, since it has 
many dependencies on environment. But I think for an interactive analytical 
tool, installing packages across the session is not very rare. Still, I'm open 
to hear your thought, thanks.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at [email protected] or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to