Github user sahilTakiar commented on the issue:
https://github.com/apache/spark/pull/20893
Wrote a test in `SparkLauncherSuite` and was able to replicate the error
from HIVE-18533, and then realized the exception is only logged and then
swallowed. From `SparkContext`
```
if (_dagScheduler != null) {
Utils.tryLogNonFatalError {
_dagScheduler.stop()
}
_dagScheduler = null
}
```
Seems the failures I was seeing in HIVE-18533 are due to something else.
Regardless, this is still probably a good fix since you still don't want to
write to the connection unless its open, but given that the exception is only
logged and not thrown, don't see an easy way to write a test for this.
---
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]