Github user tgravescs commented on the pull request:
https://github.com/apache/spark/pull/560#issuecomment-46603671
so the example I gave above is on the master branch and the configs I set
show up on both driver and executors. I'm not concerned with configs that
don't start with "spark." as those aren't spark configs. When you say it
doesn't show up on executors how are you checking? You don't see the logs in
the executors that I listed? I ran on both 0.23 and 2.4 clusters. So I'm
wondering why its not working for you.
Note that most configs get converted into SparkConf and sent to the
executor via akka when it registers so they won't show up in the processes line
with -D's or via getProperty. The security settings are special as they are
needed before the registration happens.
Again I'm only concerned with actual spark configs (spark.*) and I'm only
concerned about the spark framework properly reading them. I'm not concerned
with application code reading them.
Another example you can use is :
export SPARK_JAVA_OPTS="-Dspark.authenticate=true
-Dspark.ui.acls.enable=true -Dspark.akka.threads=10
-Dspark.akka.logAkkaConfig=true"
make sure that the akka settings get logged at the beginning of the
executor processes.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at [email protected] or file a JIRA ticket
with INFRA.
---