Github user vanzin commented on the pull request:

    https://github.com/apache/spark/pull/2739#issuecomment-60275209
  
    Hi @jacek-lewandowski,
    
    Now that I finally noticed you built this on top of branch-1.1, some of the 
choices you made make a lot more sense. (I always assume people are working on 
master, since it's generally preferable to add new features to master first.)
    
    One huge difference in master, which lead to a lot of my comments, is 
SPARK-2098. That fix added the ability of all daemons - including Master and 
Worker - to read the spark-defaults.conf file. So, if you build on top of that, 
you need zero code dealing with loading config data, and can rely on SparkConf 
for everything. Then, you could have something like:
    
        class SSLOptions(conf: SparkConf, module: String)
    
    That would load options like this:
    
        sslEnabled = conf.getOption(s"spark.$module.ssl.enabled")
          .orElse(conf.getOption("spark.ssl.enabled"))
          .getOrElse(false)
    
    Then you have module-specific configuration and a global fallback. What do 
you think?
    
    On the subject of distributing the configuration, I think it's sort of ok 
to rely on that, for the time being, for standalone mode. Long term, it would 
be better to allow each job to be able to distribute its own configuration, so 
that it's easy for admins and users to use different certificates for the 
daemons and for the jobs, for example.
    
    On Yarn, I still believe we should not have this requirement - since when 
using Spark-on-Yarn, Spark is kind of a client-side thing and shouldn't require 
any changes in the cluster . The needed files should be distributed 
automatically by Spark and made available to executors. That should be doable 
by disabling certificate validation (so that the hostnames don't matter) or 
using wildcard certificates (assuming everything is in the same sub-domain). If 
that's not enough to cover all user cases, we can leave other enhancements for 
later.
    
    I'm not familiar enough with mesos to be able to suggest anything.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org

Reply via email to