[
https://issues.apache.org/jira/browse/HIVE-12046?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14946785#comment-14946785
]
Xuefu Zhang commented on HIVE-12046:
------------------------------------
Look at the patch again and it seems that it handles the case where the remote
spark client is in a bad state when trying to submit a spark job. This is good.
However, it's unclear that what's going to happen when either
getDefaultParallelism() or getExecutorCount() is called in such a situation.
Also even in case that exectue() is called, the remote client can become bad
right after the isActive() check.
Therefore, I think we need to define a scope for this JIRA. If we want to be
resilient to the connection loss, then we need to consider more cases and the
way to handle it. However, it's also acceptable in my opinion that we can
detect the error and ask user to log out and log in again to get a valid
session. The latter seems simpler and easier.
> Re-create spark client if connection is dropped
> -----------------------------------------------
>
> Key: HIVE-12046
> URL: https://issues.apache.org/jira/browse/HIVE-12046
> Project: Hive
> Issue Type: Bug
> Components: Spark
> Reporter: Jimmy Xiang
> Assignee: Jimmy Xiang
> Priority: Minor
> Fix For: 1.3.0, 2.0.0
>
> Attachments: HIVE-12046.1.patch
>
>
> Currently, if the connection to the spark cluster is dropped, the spark
> client will stay in a bad state. A new Hive session is needed to re-establish
> the connection. It is better to auto reconnect in this case.
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)