[ https://issues.apache.org/jira/browse/AMBARI-24455?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Jayush Luniya resolved AMBARI-24455. ------------------------------------ Resolution: Invalid HDP-3.0 stack definition issue. > Spark service check failure after UI Deploy - non Secure with Ranger/KMS > ------------------------------------------------------------------------ > > Key: AMBARI-24455 > URL: https://issues.apache.org/jira/browse/AMBARI-24455 > Project: Ambari > Issue Type: Bug > Components: ambari-server > Affects Versions: 2.7.1 > Reporter: Srikanth Janardhan > Assignee: Dmytro Grinenko > Priority: Blocker > Fix For: 2.7.1 > > > Spark Service check fails while UI Deploy: > {code} > stderr: > Traceback (most recent call last): > File > "/var/lib/ambari-agent/cache/stacks/HDP/3.0/services/SPARK2/package/scripts/service_check.py", > line 78, in service_check > Execute(cmd, user=params.smoke_user, path=[beeline_cmd], > timeout=CHECK_COMMAND_TIMEOUT_DEFAULT) > File "/usr/lib/ambari-agent/lib/resource_management/core/base.py", line > 166, in __init__ > self.env.run() > File "/usr/lib/ambari-agent/lib/resource_management/core/environment.py", > line 160, in run > self.run_action(resource, action) > File "/usr/lib/ambari-agent/lib/resource_management/core/environment.py", > line 124, in run_action > provider_action() > File > "/usr/lib/ambari-agent/lib/resource_management/core/providers/system.py", > line 263, in action_run > returns=self.resource.returns) > File "/usr/lib/ambari-agent/lib/resource_management/core/shell.py", line > 72, in inner > result = function(command, **kwargs) > File "/usr/lib/ambari-agent/lib/resource_management/core/shell.py", line > 102, in checked_call > tries=tries, try_sleep=try_sleep, > timeout_kill_strategy=timeout_kill_strategy, returns=returns) > File "/usr/lib/ambari-agent/lib/resource_management/core/shell.py", line > 150, in _call_wrapper > result = _call(command, **kwargs_copy) > File "/usr/lib/ambari-agent/lib/resource_management/core/shell.py", line > 314, in _call > raise ExecutionFailed(err_msg, code, out, err) > ExecutionFailed: Execution of '! /usr/hdp/current/spark2-client/bin/beeline > -u > 'jdbc:hive2://ctr-e138-1518143905142-432870-01-000004.hwx.site:10016/default' > transportMode=binary -e '' 2>&1| awk '{print}'|grep -i -e 'Connection > refused' -e 'Invalid URL' -e 'Error: Could not open'' returned 1. ######## > Hortonworks ############# > This is MOTD message, added for testing in qe infra > Error: Could not open client transport with JDBC Uri: > jdbc:hive2://ctr-e138-1518143905142-432870-01-000004.hwx.site:10016/default: > java.net.ConnectException: Connection refused (Connection refused) > (state=08S01,code=0) > Error: Could not open client transport with JDBC Uri: > jdbc:hive2://ctr-e138-1518143905142-432870-01-000004.hwx.site:10016/default: > java.net.ConnectException: Connection refused (Connection refused) > (state=08S01,code=0) > The above exception was the cause of the following exception: > Traceback (most recent call last): > File > "/var/lib/ambari-agent/cache/stacks/HDP/3.0/services/SPARK2/package/scripts/service_check.py", > line 88, in > SparkServiceCheck().execute() > File > "/usr/lib/ambari-agent/lib/resource_management/libraries/script/script.py", > line 353, in execute > method(env) > File > "/var/lib/ambari-agent/cache/stacks/HDP/3.0/services/SPARK2/package/scripts/service_check.py", > line 85, in service_check > raise Fail("Connection to all Spark thrift servers servers failed") > resource_management.core.exceptions.Fail: Connection to all Spark thrift > servers servers failed > stdout: > 2018-08-10 05:22:29,576 - Using hadoop conf dir: > /usr/hdp/3.0.1.0-73/hadoop/conf > 2018-08-10 05:22:29,600 - Execute['curl -s -o /dev/null -w'%{http_code}' > --negotiate -u: -k > http://ctr-e138-1518143905142-432870-01-000004.hwx.site:18081 | grep 200'] > {'logoutput': True, 'tries': 5, 'user': 'ambari-qa', 'try_sleep': 3} > ######## Hortonworks ############# > This is MOTD message, added for testing in qe infra > 200 > 2018-08-10 05:22:29,794 - Execute['curl -s -o /dev/null -w'%{http_code}' > --negotiate -u: -k > http://ctr-e138-1518143905142-432870-01-000004.hwx.site:8999/sessions | grep > 200'] {'logoutput': True, 'tries': 3, 'user': 'ambari-qa', 'try_sleep': 1} > ######## Hortonworks ############# > This is MOTD message, added for testing in qe infra > 200 > 2018-08-10 05:22:29,843 - Execute['! > /usr/hdp/current/spark2-client/bin/beeline -u > 'jdbc:hive2://ctr-e138-1518143905142-432870-01-000004.hwx.site:10016/default' > transportMode=binary -e '' 2>&1| awk '{print}'|grep -i -e 'Connection > refused' -e 'Invalid URL' -e 'Error: Could not open''] {'path': > [u'/usr/hdp/current/spark2-client/bin/beeline'], 'user': 'ambari-qa', > 'timeout': 60.0} > Command failed after 1 tries > {code} > I checked the schedule of commands and observed that Spark Thrift server > start is scheduled to run after Spark2 service check is executed. -- This message was sent by Atlassian JIRA (v7.6.3#76005)