GitHub user andrewor14 opened a pull request:

    https://github.com/apache/spark/pull/2067

    [SPARK-3140] Clarify confusing PySpark exception message

    We read the py4j port from the stdout of the `bin/spark-submit` subprocess. 
If there is interference in stdout (e.g. a random echo in `spark-submit`), we 
throw an exception with a warning message. We do not, however, distinguish 
between this case from the case where no stdout is produced at all.
    
    I wasted a non-trivial amount of time being baffled by this exception in 
search of places where I print random whitespace (in vain, of course). A 
clearer exception message that distinguishes between these cases will prevent 
future headaches that I have gone through.

You can merge this pull request into a Git repository by running:

    $ git pull https://github.com/andrewor14/spark python-exception

Alternatively you can review and apply these changes as the patch at:

    https://github.com/apache/spark/pull/2067.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

    This closes #2067
    
----
commit e96a7a0016502820f325c39d71b236b2b39e0cb6
Author: Andrew Or <[email protected]>
Date:   2014-08-20T22:09:33Z

    Distinguish between unexpected output and no output at all

----


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at [email protected] or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to