Github user ankurdave commented on the pull request:

    https://github.com/apache/spark/pull/4136#issuecomment-71286155
  
    @JoshRosen Actually, it seems the test failures still occur, but only when 
I add a [unit 
test](https://github.com/apache/spark/commit/9b225ac3072de522b40b46aba6df1f1c231f13ef#diff-3ade47bc293ef06e43c25f1ac1f6783bR354)
 that sets spark.default.parallelism.
    
    Adding the test causes subsequent tests within the same run to fail with 
exceptions like
    ```
    java.io.IOException: org.apache.spark.SparkException: Failed to get 
broadcast_0_piece0 of broadcast_0
    ```
    and
    ```
    java.io.IOException: PARSING_ERROR(2)
    ```
    The exception traces always occur in TorrentBroadcast.
    
    It seems like setting spark.default.parallelism is causing some kind of 
side effect that corrupts broadcasts in later unit tests, which is strange 
since (1) each unit test should have its own SparkContext and therefore its own 
temp directory, and (2) I'm only passing spark.default.parallelism to 
SparkConf/SparkContext, not setting it as a system property.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at [email protected] or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to