Github user pwendell commented on the pull request:
https://github.com/apache/spark/pull/1253#issuecomment-49570062
IMO `-D` does not have the right semantics here because the user isn't
logically setting java properties for the submission tool, they are setting
spark configuration properties for their application. The application might run
totally remotely, for instance, so why should the user expect that a -D set on
the submission site gets packaged up and sent to the remote launcher. Also
confusing is that we'd only really triage -D options that are spark properties,
not other ones, so the semantics would differ depending on whether the user
happened to set a java property that started with `spark`. For these reasons I
feel it's better to just have an explicit config-related flag.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at [email protected] or file a JIRA ticket
with INFRA.
---