[ 
https://issues.apache.org/jira/browse/AMBARI-23902?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ishan Bhatt updated AMBARI-23902:
---------------------------------
    Fix Version/s:     (was: 2.7.0)
                   2.7.1

> YARN service check failure with no permission to submit jobs for ambari-qa 
> user
> -------------------------------------------------------------------------------
>
>                 Key: AMBARI-23902
>                 URL: https://issues.apache.org/jira/browse/AMBARI-23902
>             Project: Ambari
>          Issue Type: Bug
>          Components: ambari-server
>    Affects Versions: 2.7.0
>            Reporter: Srikanth Janardhan
>            Assignee: Dmitry Lysnichenko
>            Priority: Blocker
>             Fix For: 2.7.1
>
>
> Service check for YARN failed:
> {code}
> stderr: 
> Traceback (most recent call last):
>   File 
> "/var/lib/ambari-agent/cache/stacks/HDP/3.0/services/YARN/package/scripts/service_check.py",
>  line 185, in 
>     ServiceCheck().execute()
>   File 
> "/usr/lib/ambari-agent/lib/resource_management/libraries/script/script.py", 
> line 353, in execute
>     method(env)
>   File 
> "/var/lib/ambari-agent/cache/stacks/HDP/3.0/services/YARN/package/scripts/service_check.py",
>  line 121, in service_check
>     user=params.smokeuser,
>   File "/usr/lib/ambari-agent/lib/resource_management/core/shell.py", line 
> 72, in inner
>     result = function(command, **kwargs)
>   File "/usr/lib/ambari-agent/lib/resource_management/core/shell.py", line 
> 102, in checked_call
>     tries=tries, try_sleep=try_sleep, 
> timeout_kill_strategy=timeout_kill_strategy, returns=returns)
>   File "/usr/lib/ambari-agent/lib/resource_management/core/shell.py", line 
> 150, in _call_wrapper
>     result = _call(command, **kwargs_copy)
>   File "/usr/lib/ambari-agent/lib/resource_management/core/shell.py", line 
> 308, in _call
>     raise ExecutionFailed(err_msg, code, out, err)
> resource_management.core.exceptions.ExecutionFailed: Execution of 
> '/usr/bin/kinit -kt /etc/security/keytabs/smokeuser.headless.keytab 
> [email protected]; yarn 
> org.apache.hadoop.yarn.applications.distributedshell.Client -shell_command ls 
> -num_containers 1 -jar 
> /usr/hdp/current/hadoop-yarn-client/hadoop-yarn-applications-distributedshell.jar
>  -timeout 300000 --queue default' returned 1. ######## Hortonworks 
> #############
> This is MOTD message, added for testing in qe infra
> 18/05/18 11:33:09 INFO distributedshell.Client: Initializing Client
> 18/05/18 11:33:09 INFO distributedshell.Client: Running Client
> 18/05/18 11:33:09 INFO client.AHSProxy: Connecting to Application History 
> server at ctr-e138-1518143905142-317995-01-000004.hwx.site/172.27.56.82:10200
> 18/05/18 11:33:09 INFO client.ConfiguredRMFailoverProxyProvider: Failing over 
> to rm2
> 18/05/18 11:33:10 INFO distributedshell.Client: Got Cluster metric info from 
> ASM, numNodeManagers=3
> 18/05/18 11:33:10 INFO distributedshell.Client: Got Cluster node info from ASM
> 18/05/18 11:33:10 INFO distributedshell.Client: Got node report from ASM for, 
> nodeId=ctr-e138-1518143905142-317995-01-000009.hwx.site:25454, 
> nodeAddress=ctr-e138-1518143905142-317995-01-000009.hwx.site:8042, 
> nodeRackName=/default-rack, nodeNumContainers=1
> 18/05/18 11:33:10 INFO distributedshell.Client: Got node report from ASM for, 
> nodeId=ctr-e138-1518143905142-317995-01-000006.hwx.site:25454, 
> nodeAddress=ctr-e138-1518143905142-317995-01-000006.hwx.site:8042, 
> nodeRackName=/default-rack, nodeNumContainers=0
> 18/05/18 11:33:10 INFO distributedshell.Client: Got node report from ASM for, 
> nodeId=ctr-e138-1518143905142-317995-01-000007.hwx.site:25454, 
> nodeAddress=ctr-e138-1518143905142-317995-01-000007.hwx.site:8042, 
> nodeRackName=/default-rack, nodeNumContainers=0
> 18/05/18 11:33:10 INFO distributedshell.Client: Queue info, 
> queueName=default, queueCurrentCapacity=0.027777778, queueMaxCapacity=1.0, 
> queueApplicationCount=0, queueChildQueueCount=0
> 18/05/18 11:33:10 INFO distributedshell.Client: Max mem capability of 
> resources in this cluster 12288
> 18/05/18 11:33:10 INFO distributedshell.Client: Max virtual cores capability 
> of resources in this cluster 1
> 18/05/18 11:33:10 WARN distributedshell.Client: AM Memory not specified, use 
> 100 mb as AM memory
> 18/05/18 11:33:10 WARN distributedshell.Client: AM vcore not specified, use 1 
> mb as AM vcores
> 18/05/18 11:33:10 WARN distributedshell.Client: AM Resource capability=
> 18/05/18 11:33:10 INFO distributedshell.Client: Copy App Master jar from 
> local filesystem and add to local environment
> 18/05/18 11:33:11 INFO distributedshell.Client: Set the environment for the 
> application master
> 18/05/18 11:33:11 INFO distributedshell.Client: Setting up app master command
> 18/05/18 11:33:11 INFO distributedshell.Client: Completed setting up app 
> master command {{JAVA_HOME}}/bin/java -Xmx100m 
> org.apache.hadoop.yarn.applications.distributedshell.ApplicationMaster 
> --container_type GUARANTEED --container_memory 10 --container_vcores 1 
> --num_containers 1 --priority 0 1>/AppMaster.stdout 2>/AppMaster.stderr 
> 18/05/18 11:33:11 INFO hdfs.DFSClient: Created token for ambari-qa: 
> HDFS_DELEGATION_TOKEN [email protected], renewer=cstm-yarn, 
> realUser=, issueDate=1526643191972, maxDate=1527247991972, sequenceNumber=27, 
> masterKeyId=4 on ha-hdfs:mycluster
> 18/05/18 11:33:12 INFO distributedshell.Client: Got dt for hdfs://mycluster; 
> Kind: HDFS_DELEGATION_TOKEN, Service: ha-hdfs:mycluster, Ident: (token for 
> ambari-qa: HDFS_DELEGATION_TOKEN [email protected], 
> renewer=cstm-yarn, realUser=, issueDate=1526643191972, maxDate=1527247991972, 
> sequenceNumber=27, masterKeyId=4)
> 18/05/18 11:33:12 INFO distributedshell.Client: Submitting application to ASM
> 18/05/18 11:33:12 INFO impl.TimelineClientImpl: Timeline service address: 
> ctr-e138-1518143905142-317995-01-000004.hwx.site:8188
> 18/05/18 11:33:12 ERROR distributedshell.Client: Error running Client
> org.apache.hadoop.yarn.exceptions.YarnException: 
> org.apache.hadoop.security.AccessControlException: User ambari-qa does not 
> have permission to submit application_1526608768177_0003 to queue default
>       at 
> org.apache.hadoop.yarn.ipc.RPCUtil.getRemoteException(RPCUtil.java:38)
>       at 
> org.apache.hadoop.yarn.server.resourcemanager.RMAppManager.createAndPopulateNewRMApp(RMAppManager.java:436)
>       at 
> org.apache.hadoop.yarn.server.resourcemanager.RMAppManager.submitApplication(RMAppManager.java:321)
>       at 
> org.apache.hadoop.yarn.server.resourcemanager.ClientRMService.submitApplication(ClientRMService.java:645)
>       at 
> org.apache.hadoop.yarn.api.impl.pb.service.ApplicationClientProtocolPBServiceImpl.submitApplication(ApplicationClientProtocolPBServiceImpl.java:277)
>       at 
> org.apache.hadoop.yarn.proto.ApplicationClientProtocol$ApplicationClientProtocolService$2.callBlockingMethod(ApplicationClientProtocol.java:563)
>       at 
> org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:523)
>       at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:991)
>       at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:872)
>       at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:818)
>       at java.security.AccessController.doPrivileged(Native Method)
>       at javax.security.auth.Subject.doAs(Subject.java:422)
>       at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1682)
>       at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2678)
> Caused by: org.apache.hadoop.security.AccessControlException: User ambari-qa 
> does not have permission to submit application_1526608768177_0003 to queue 
> default
>       at 
> org.apache.hadoop.yarn.server.resourcemanager.RMAppManager.createAndPopulateNewRMApp(RMAppManager.java:438)
>       ... 12 more
>       at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
>       at 
> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
>       at 
> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
>       at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
>       at 
> org.apache.hadoop.yarn.ipc.RPCUtil.instantiateException(RPCUtil.java:53)
>       at 
> org.apache.hadoop.yarn.ipc.RPCUtil.instantiateYarnException(RPCUtil.java:75)
>       at 
> org.apache.hadoop.yarn.ipc.RPCUtil.unwrapAndThrowException(RPCUtil.java:116)
>       at 
> org.apache.hadoop.yarn.api.impl.pb.client.ApplicationClientProtocolPBClientImpl.submitApplication(ApplicationClientProtocolPBClientImpl.java:292)
>       at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>       at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>       at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>       at java.lang.reflect.Method.invoke(Method.java:498)
>       at 
> org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:422)
>       at 
> org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeMethod(RetryInvocationHandler.java:165)
>       at 
> org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invoke(RetryInvocationHandler.java:157)
>       at 
> org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeOnce(RetryInvocationHandler.java:95)
>       at 
> org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:359)
>       at com.sun.proxy.$Proxy9.submitApplication(Unknown Source)
>       at 
> org.apache.hadoop.yarn.client.api.impl.YarnClientImpl.submitApplication(YarnClientImpl.java:286)
>       at 
> org.apache.hadoop.yarn.applications.distributedshell.Client.run(Client.java:981)
>       at 
> org.apache.hadoop.yarn.applications.distributedshell.Client.main(Client.java:265)
> Caused by: 
> org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.yarn.exceptions.YarnException):
>  org.apache.hadoop.security.AccessControlException: User ambari-qa does not 
> have permission to submit application_1526608768177_0003 to queue default
>       at 
> org.apache.hadoop.yarn.ipc.RPCUtil.getRemoteException(RPCUtil.java:38)
>       at 
> org.apache.hadoop.yarn.server.resourcemanager.RMAppManager.createAndPopulateNewRMApp(RMAppManager.java:436)
>       at 
> org.apache.hadoop.yarn.server.resourcemanager.RMAppManager.submitApplication(RMAppManager.java:321)
>       at 
> org.apache.hadoop.yarn.server.resourcemanager.ClientRMService.submitApplication(ClientRMService.java:645)
>       at 
> org.apache.hadoop.yarn.api.impl.pb.service.ApplicationClientProtocolPBServiceImpl.submitApplication(ApplicationClientProtocolPBServiceImpl.java:277)
>       at 
> org.apache.hadoop.yarn.proto.ApplicationClientProtocol$ApplicationClientProtocolService$2.callBlockingMethod(ApplicationClientProtocol.java:563)
>       at 
> org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:523)
>       at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:991)
>       at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:872)
>       at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:818)
>       at java.security.AccessController.doPrivileged(Native Method)
>       at javax.security.auth.Subject.doAs(Subject.java:422)
>       at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1682)
>       at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2678)
> Caused by: org.apache.hadoop.security.AccessControlException: User ambari-qa 
> does not have permission to submit application_1526608768177_0003 to queue 
> default
>       at 
> org.apache.hadoop.yarn.server.resourcemanager.RMAppManager.createAndPopulateNewRMApp(RMAppManager.java:438)
>       ... 12 more
>       at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1491)
>       at org.apache.hadoop.ipc.Client.call(Client.java:1437)
>       at org.apache.hadoop.ipc.Client.call(Client.java:1347)
>       at 
> org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:228)
>       at 
> org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:116)
>       at com.sun.proxy.$Proxy8.submitApplication(Unknown Source)
>       at 
> org.apache.hadoop.yarn.api.impl.pb.client.ApplicationClientProtocolPBClientImpl.submitApplication(ApplicationClientProtocolPBClientImpl.java:289)
>       ... 13 more
>  stdout:
> 2018-05-18 11:33:06,916 - Using hadoop conf dir: 
> /usr/hdp/3.0.0.0-1356/hadoop/conf
> 2018-05-18 11:33:06,917 - Stack Feature Version Info: Cluster Stack=3.0, 
> Command Stack=None, Command Version=3.0.0.0-1356 -> 3.0.0.0-1356
> 2018-05-18 11:33:07,046 - Using hadoop conf dir: 
> /usr/hdp/3.0.0.0-1356/hadoop/conf
> 2018-05-18 11:33:07,107 - HdfsResource['/user/ambari-qa'] 
> {'security_enabled': True, 'hadoop_bin_dir': 
> '/usr/hdp/3.0.0.0-1356/hadoop/bin', 'keytab': 
> '/etc/security/keytabs/hdfs.headless.keytab', 'dfs_type': '', 'default_fs': 
> 'hdfs://mycluster', 'hdfs_resource_ignore_file': 
> '/var/lib/ambari-agent/data/.hdfs_resource_ignore', 'hdfs_site': ..., 
> 'kinit_path_local': '/usr/bin/kinit', 'principal_name': 
> '[email protected]', 'user': 'cstm-hdfs', 'owner': 'ambari-qa', 
> 'hadoop_conf_dir': '/usr/hdp/3.0.0.0-1356/hadoop/conf', 'type': 'directory', 
> 'action': ['create_on_execute'], 'immutable_paths': [u'/apps/hive/warehouse', 
> u'/mr-history/done', u'/app-logs', u'/tmp'], 'mode': 0770}
> 2018-05-18 11:33:07,122 - Execute['/usr/bin/kinit -kt 
> /etc/security/keytabs/hdfs.headless.keytab [email protected]'] {'user': 
> 'cstm-hdfs'}
> 2018-05-18 11:33:07,233 - call['ambari-sudo.sh su cstm-hdfs -l -s /bin/bash 
> -c 'curl --negotiate -u : -s 
> '"'"'http://ctr-e138-1518143905142-317995-01-000004.hwx.site:20070/jmx?qry=Hadoop:service=NameNode,name=FSNamesystem'"'"'
>  1>/tmp/tmpH0Q_aW 2>/tmp/tmpmFbDsW''] {'quiet': False}
> 2018-05-18 11:33:07,340 - call returned (0, '######## Hortonworks 
> #############\nThis is MOTD message, added for testing in qe infra')
> 2018-05-18 11:33:07,341 - get_user_call_output returned (0, u'{\n  "beans" : 
> [ {\n    "name" : "Hadoop:service=NameNode,name=FSNamesystem",\n    
> "modelerType" : "FSNamesystem",\n    "tag.Context" : "dfs",\n    
> "tag.HAState" : "standby",\n    "tag.TotalSyncTimes" : "",\n    
> "tag.Hostname" : "ctr-e138-1518143905142-317995-01-000004.hwx.site",\n    
> "MissingBlocks" : 0,\n    "MissingReplOneBlocks" : 0,\n    
> "ExpiredHeartbeats" : 0,\n    "TransactionsSinceLastCheckpoint" : 5054,\n    
> "TransactionsSinceLastLogRoll" : 0,\n    "LastWrittenTransactionId" : 6338,\n 
>    "LastCheckpointTime" : 1526630224438,\n    "CapacityTotal" : 
> 1656383242752,\n    "CapacityTotalGB" : 1543.0,\n    "CapacityUsed" : 
> 3494182380,\n    "CapacityUsedGB" : 3.0,\n    "CapacityRemaining" : 
> 1549721537687,\n    "ProvidedCapacityTotal" : 0,\n    "CapacityRemainingGB" : 
> 1443.0,\n    "CapacityUsedNonDFS" : 0,\n    "TotalLoad" : 68,\n    
> "SnapshottableDirectories" : 0,\n    "Snapshots" : 0,\n    
> "NumEncryptionZones" : 0,\n    "LockQueueLength" : 0,\n    "BlocksTotal" : 
> 258,\n    "NumFilesUnderConstruction" : 11,\n    "NumActiveClients" : 9,\n    
> "FilesTotal" : 1233,\n    "PendingReplicationBlocks" : 0,\n    
> "PendingReconstructionBlocks" : 0,\n    "UnderReplicatedBlocks" : 0,\n    
> "LowRedundancyBlocks" : 0,\n    "CorruptBlocks" : 0,\n    
> "ScheduledReplicationBlocks" : 0,\n    "PendingDeletionBlocks" : 0,\n    
> "LowRedundancyReplicatedBlocks" : 0,\n    "CorruptReplicatedBlocks" : 0,\n    
> "MissingReplicatedBlocks" : 0,\n    "MissingReplicationOneBlocks" : 0,\n    
> "BytesInFutureReplicatedBlocks" : 0,\n    "PendingDeletionReplicatedBlocks" : 
> 0,\n    "TotalReplicatedBlocks" : 258,\n    "LowRedundancyECBlockGroups" : 
> 0,\n    "CorruptECBlockGroups" : 0,\n    "MissingECBlockGroups" : 0,\n    
> "BytesInFutureECBlockGroups" : 0,\n    "PendingDeletionECBlocks" : 0,\n    
> "TotalECBlockGroups" : 0,\n    "ExcessBlocks" : 0,\n    
> "NumTimedOutPendingReconstructions" : 0,\n    "PostponedMisreplicatedBlocks" 
> : 0,\n    "PendingDataNodeMessageCount" : 1,\n    
> "MillisSinceLastLoadedEdits" : 71633,\n    "BlockCapacity" : 4194304,\n    
> "NumLiveDataNodes" : 7,\n    "NumDeadDataNodes" : 0,\n    
> "NumDecomLiveDataNodes" : 0,\n    "NumDecomDeadDataNodes" : 0,\n    
> "VolumeFailuresTotal" : 0,\n    "EstimatedCapacityLostTotal" : 0,\n    
> "NumDecommissioningDataNodes" : 0,\n    "StaleDataNodes" : 0,\n    
> "NumStaleStorages" : 0,\n    "TotalSyncCount" : 0,\n    
> "NumInMaintenanceLiveDataNodes" : 0,\n    "NumInMaintenanceDeadDataNodes" : 
> 0,\n    "NumEnteringMaintenanceDataNodes" : 0\n  } ]\n}', u'')
> 2018-05-18 11:33:07,344 - call['ambari-sudo.sh su cstm-hdfs -l -s /bin/bash 
> -c 'curl --negotiate -u : -s 
> '"'"'http://ctr-e138-1518143905142-317995-01-000005.hwx.site:20070/jmx?qry=Hadoop:service=NameNode,name=FSNamesystem'"'"'
>  1>/tmp/tmpj1d3wc 2>/tmp/tmpQ4AEly''] {'quiet': False}
> 2018-05-18 11:33:07,906 - call returned (0, '######## Hortonworks 
> #############\nThis is MOTD message, added for testing in qe infra')
> 2018-05-18 11:33:07,906 - get_user_call_output returned (0, u'{\n  "beans" : 
> [ {\n    "name" : "Hadoop:service=NameNode,name=FSNamesystem",\n    
> "modelerType" : "FSNamesystem",\n    "tag.Context" : "dfs",\n    
> "tag.HAState" : "active",\n    "tag.TotalSyncTimes" : "43 17 ",\n    
> "tag.Hostname" : "ctr-e138-1518143905142-317995-01-000005.hwx.site",\n    
> "MissingBlocks" : 0,\n    "MissingReplOneBlocks" : 0,\n    
> "ExpiredHeartbeats" : 0,\n    "TransactionsSinceLastCheckpoint" : 5082,\n    
> "TransactionsSinceLastLogRoll" : 28,\n    "LastWrittenTransactionId" : 
> 20810,\n    "LastCheckpointTime" : 1526630224790,\n    "CapacityTotal" : 
> 1656383242752,\n    "CapacityTotalGB" : 1543.0,\n    "CapacityUsed" : 
> 3494182380,\n    "CapacityUsedGB" : 3.0,\n    "CapacityRemaining" : 
> 1549721537687,\n    "ProvidedCapacityTotal" : 0,\n    "CapacityRemainingGB" : 
> 1443.0,\n    "CapacityUsedNonDFS" : 0,\n    "TotalLoad" : 68,\n    
> "SnapshottableDirectories" : 0,\n    "Snapshots" : 0,\n    
> "NumEncryptionZones" : 0,\n    "LockQueueLength" : 0,\n    "BlocksTotal" : 
> 258,\n    "NumFilesUnderConstruction" : 11,\n    "NumActiveClients" : 9,\n    
> "FilesTotal" : 1233,\n    "PendingReplicationBlocks" : 0,\n    
> "PendingReconstructionBlocks" : 0,\n    "UnderReplicatedBlocks" : 0,\n    
> "LowRedundancyBlocks" : 0,\n    "CorruptBlocks" : 0,\n    
> "ScheduledReplicationBlocks" : 0,\n    "PendingDeletionBlocks" : 0,\n    
> "LowRedundancyReplicatedBlocks" : 0,\n    "CorruptReplicatedBlocks" : 0,\n    
> "MissingReplicatedBlocks" : 0,\n    "MissingReplicationOneBlocks" : 0,\n    
> "BytesInFutureReplicatedBlocks" : 0,\n    "PendingDeletionReplicatedBlocks" : 
> 0,\n    "TotalReplicatedBlocks" : 258,\n    "LowRedundancyECBlockGroups" : 
> 0,\n    "CorruptECBlockGroups" : 0,\n    "MissingECBlockGroups" : 0,\n    
> "BytesInFutureECBlockGroups" : 0,\n    "PendingDeletionECBlocks" : 0,\n    
> "TotalECBlockGroups" : 0,\n    "ExcessBlocks" : 0,\n    
> "NumTimedOutPendingReconstructions" : 31,\n    "PostponedMisreplicatedBlocks" 
> : 0,\n    "PendingDataNodeMessageCount" : 0,\n    
> "MillisSinceLastLoadedEdits" : 0,\n    "BlockCapacity" : 4194304,\n    
> "NumLiveDataNodes" : 7,\n    "NumDeadDataNodes" : 0,\n    
> "NumDecomLiveDataNodes" : 0,\n    "NumDecomDeadDataNodes" : 0,\n    
> "VolumeFailuresTotal" : 0,\n    "EstimatedCapacityLostTotal" : 0,\n    
> "NumDecommissioningDataNodes" : 0,\n    "StaleDataNodes" : 0,\n    
> "NumStaleStorages" : 0,\n    "TotalSyncCount" : 28,\n    
> "NumInMaintenanceLiveDataNodes" : 0,\n    "NumInMaintenanceDeadDataNodes" : 
> 0,\n    "NumEnteringMaintenanceDataNodes" : 0\n  } ]\n}', u'')
> 2018-05-18 11:33:07,908 - NameNode HA states: active_namenodes = [(u'nn2', 
> 'ctr-e138-1518143905142-317995-01-000005.hwx.site:20070')], standby_namenodes 
> = [(u'nn1', 'ctr-e138-1518143905142-317995-01-000004.hwx.site:20070')], 
> unknown_namenodes = []
> 2018-05-18 11:33:07,909 - call['ambari-sudo.sh su cstm-hdfs -l -s /bin/bash 
> -c 'curl -sS -L -w '"'"'%{http_code}'"'"' -X GET --negotiate -u : 
> '"'"'http://ctr-e138-1518143905142-317995-01-000005.hwx.site:20070/webhdfs/v1/user/ambari-qa?op=GETFILESTATUS'"'"'
>  1>/tmp/tmpyVY03k 2>/tmp/tmpL2L6US''] {'logoutput': None, 'quiet': False}
> 2018-05-18 11:33:08,024 - call returned (0, '######## Hortonworks 
> #############\nThis is MOTD message, added for testing in qe infra')
> 2018-05-18 11:33:08,025 - get_user_call_output returned (0, 
> u'{"FileStatus":{"accessTime":0,"blockSize":0,"childrenNum":1,"fileId":16388,"group":"hdfs","length":0,"modificationTime":1526609917254,"owner":"ambari-qa","pathSuffix":"","permission":"770","replication":0,"storagePolicy":0,"type":"DIRECTORY"}}200',
>  u'')
> 2018-05-18 11:33:08,029 - checked_call['/usr/bin/kinit -kt 
> /etc/security/keytabs/smokeuser.headless.keytab [email protected]; yarn 
> org.apache.hadoop.yarn.applications.distributedshell.Client -shell_command ls 
> -num_containers 1 -jar 
> /usr/hdp/current/hadoop-yarn-client/hadoop-yarn-applications-distributedshell.jar
>  -timeout 300000 --queue default'] {'path': 
> '/usr/sbin:/sbin:/usr/local/bin:/bin:/usr/bin', 'user': 'ambari-qa'}
> Command failed after 1 tries
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

Reply via email to