[ https://issues.apache.org/jira/browse/AMBARI-21544?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16096078#comment-16096078 ]
Hadoop QA commented on AMBARI-21544: ------------------------------------ {color:red}-1 overall{color}. Here are the results of testing the latest attachment http://issues.apache.org/jira/secure/attachment/12878330/AMBARI-21544.patch against trunk revision . {color:green}+1 @author{color}. The patch does not contain any @author tags. {color:red}-1 tests included{color}. The patch doesn't appear to include any new or modified tests. Please justify why no new tests are needed for this patch. Also please list what manual steps were performed to verify this patch. {color:green}+1 release audit{color}. The applied patch does not increase the total number of release audit warnings. {color:green}+1 javac{color}. The applied patch does not increase the total number of javac compiler warnings. {color:green}+1 core tests{color}. The patch passed unit tests in . Console output: https://builds.apache.org/job/Ambari-trunk-test-patch/11833//console This message is automatically generated. > HiveServer2 fails to start with webhdfs call to create /hdp/apps/..jar files > fails with org.apache.hadoop.hdfs.server.namenode.LeaseExpiredException > ------------------------------------------------------------------------------------------------------------------------------------------------------ > > Key: AMBARI-21544 > URL: https://issues.apache.org/jira/browse/AMBARI-21544 > Project: Ambari > Issue Type: Bug > Reporter: Andrew Onischuk > Assignee: Andrew Onischuk > Fix For: 2.5.2 > > Attachments: AMBARI-21544.patch > > > HiveServer2 fails to start with webhdfs call to create /hdp/apps/..jar files > fails with org.apache.hadoop.hdfs.server.namenode.LeaseExpiredException. > Seeing this specifically on HA cluster where one instance of HiveServer2 fails > to start. > HiveServer2 start error... > > > > 2017-07-18 05:27:36,795 - NameNode HA states: active_namenodes = > [(u'nn2', 'ctr-e134-1499953498516-16356-01-000005.hwx.site:20070')], > standby_namenodes = [(u'nn1', > 'ctr-e134-1499953498516-16356-01-000004.hwx.site:20070')], unknown_namenodes > = [] > 2017-07-18 05:27:36,797 - call['ambari-sudo.sh su hdfs -l -s /bin/bash -c > 'curl -sS -L -w '"'"'%{http_code}'"'"' -X GET --negotiate -u : > '"'"'http://ctr-e134-1499953498516-16356-01-000005.hwx.site:20070/webhdfs/v1/hdp/apps/2.6.3.0-61/mapreduce/hadoop-streaming.jar?op=GETFILESTATUS'"'"' > 1>/tmp/tmpvtBOI9 2>/tmp/tmpMbcTp1''] {'logoutput': None, 'quiet': False} > 2017-07-18 05:27:36,885 - call returned (0, '') > 2017-07-18 05:27:36,886 - Creating new file > /hdp/apps/2.6.3.0-61/mapreduce/hadoop-streaming.jar in DFS > 2017-07-18 05:27:36,887 - call['ambari-sudo.sh su hdfs -l -s /bin/bash -c > 'curl -sS -L -w '"'"'%{http_code}'"'"' -X PUT --data-binary > @/usr/hdp/2.6.3.0-61/hadoop-mapreduce/hadoop-streaming.jar -H > '"'"'Content-Type: application/octet-stream'"'"' --negotiate -u : > '"'"'http://ctr-e134-1499953498516-16356-01-000005.hwx.site:20070/webhdfs/v1/hdp/apps/2.6.3.0-61/mapreduce/hadoop-streaming.jar?op=CREATE&overwrite=True&permission=444'"'"' > 1>/tmp/tmpqYkC_P 2>/tmp/tmpT30u8x''] {'logoutput': None, 'quiet': False} > 2017-07-18 05:27:37,135 - call returned (0, '') > .... > self._create_file(self.main_resource.resource.target, > source=self.main_resource.resource.source, mode=self.mode) > File > "/usr/lib/python2.6/site-packages/resource_management/libraries/providers/hdfs_resource.py", > line 423, in _create_file > self.util.run_command(target, 'CREATE', method='PUT', overwrite=True, > assertable_result=False, file_to_put=source, **kwargs) > File > "/usr/lib/python2.6/site-packages/resource_management/libraries/providers/hdfs_resource.py", > line 204, in run_command > raise Fail(err_msg) > resource_management.core.exceptions.Fail: Execution of 'curl -sS -L -w > '%{http_code}' -X PUT --data-binary > @/usr/hdp/2.6.3.0-61/hadoop-mapreduce/hadoop-streaming.jar -H 'Content-Type: > application/octet-stream' --negotiate -u : > 'http://ctr-e134-1499953498516-16356-01-000005.hwx.site:20070/webhdfs/v1/hdp/apps/2.6.3.0-61/mapreduce/hadoop-streaming.jar?op=CREATE&overwrite=True&permission=444'' > returned status_code=403. > { > "RemoteException": { > "exception": "LeaseExpiredException", > "javaClassName": > "org.apache.hadoop.hdfs.server.namenode.LeaseExpiredException", > "message": "No lease on > /hdp/apps/2.6.3.0-61/mapreduce/hadoop-streaming.jar (inode 16566): File does > not exist. Holder DFSClient_NONMAPREDUCE_1130121686_152 does not have any > open files.\n\tat > org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkLease(FSNamesystem.java:3660)\n\tat > > org.apache.hadoop.hdfs.server.namenode.FSNamesystem.analyzeFileState(FSNamesystem.java:3463)\n\tat > > org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getNewBlockTargets(FSNamesystem.java:3301)\n\tat > > org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:3261)\n\tat > > org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:850)\n\tat > > org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:503)\n\tat > > org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)\n\tat > > org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:640)\n\tat > org.apache.hadoop.ipc.RPC$Server.call(RPC.java:982)\n\tat > org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2351)\n\tat > org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2347)\n\tat > java.security.AccessController.doPrivileged(Native Method)\n\tat > javax.security.auth.Subject.doAs(Subject.java:422)\n\tat > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1866)\n\tat > org.apache.hadoop.ipc.Server$Handler.run(Server.java:2345)\n" > > > NameNode log throws LeaseExpiredException... > > > > 2017-07-18 05:27:36,980 INFO > delegation.AbstractDelegationTokenSecretManager > (AbstractDelegationTokenSecretManager.java:createPassword(385)) - Creating > password for identifier: HDFS_DELEGATION_TOKEN token 8 for hdfs, currentKey: 2 > 2017-07-18 05:27:37,054 INFO > delegation.AbstractDelegationTokenSecretManager > (AbstractDelegationTokenSecretManager.java:createPassword(385)) - Creating > password for identifier: HDFS_DELEGATION_TOKEN token 9 for hdfs, currentKey: 2 > 2017-07-18 05:27:37,118 INFO ipc.Server (Server.java:logException(2428)) > - IPC Server handler 32 on 8020, call > org.apache.hadoop.hdfs.protocol.ClientProtocol.addBlock from > 172.27.9.200:45817 Call#2119 Retry#0: > org.apache.hadoop.hdfs.server.namenode.LeaseExpiredException: No lease on > /hdp/apps/2.6.3.0-61/mapreduce/hadoop-streaming.jar (inode 16566): File does > not exist. Holder DFSClient_NONMAPREDUCE_1130121686_152 does not have any > open files. > 2017-07-18 05:27:37,152 INFO hdfs.StateChange > (FSNamesystem.java:logAllocatedBlock(3831)) - BLOCK* allocate > blk_1073741851_1027, replicas=172.27.9.200:1019, 172.27.12.200:1019, > 172.27.24.212:1019 for /hdp/apps/2.6.3.0-61/mapreduce/hadoop-streaming.jar > 2017-07-18 05:27:37,227 INFO hdfs.StateChange > (FSNamesystem.java:completeFile(3724)) - DIR* completeFile: > /hdp/apps/2.6.3.0-61/mapreduce/hadoop-streaming.jar is closed by > DFSClient_NONMAPREDUCE_-1879489015_153 > 2017-07-18 05:27:39,523 INFO BlockStateChange > (BlockManager.java:computeReplicationWorkForBlocks(1648)) - BLOCK* > neededReplications = 0, pendingReplications = 0. > > This is not specific to hadoop-streaming.jar file creation, In other cluster > failure occurs at creating /hdp/apps/2.6.3.0-61/pig/pig.tar.gz file... > > > > 2017-07-18 05:31:50,608 INFO BlockStateChange > (BlockManager.java:computeReplicationWorkForBlocks(1648)) - BLOCK* > neededReplications = 0, pendingReplications = 0. > 2017-07-18 05:31:50,685 INFO > delegation.AbstractDelegationTokenSecretManager > (AbstractDelegationTokenSecretManager.java:createPassword(385)) - Creating > password for identifier: HDFS_DELEGATION_TOKEN token 5 for hdfs, currentKey: 2 > 2017-07-18 05:31:50,690 INFO hdfs.StateChange > (FSNamesystem.java:logAllocatedBlock(3831)) - BLOCK* allocate > blk_1073741848_1024, replicas=172.27.18.201:1019, 172.27.19.4:1019, > 172.27.52.76:1019 for /hdp/apps/2.6.3.0-61/pig/pig.tar.gz > 2017-07-18 05:31:51,228 INFO hdfs.StateChange > (FSNamesystem.java:logAllocatedBlock(3831)) - BLOCK* allocate > blk_1073741849_1025, replicas=172.27.19.4:1019, 172.27.17.134:1019, > 172.27.52.76:1019 for /hdp/apps/2.6.3.0-61/pig/pig.tar.gz > 2017-07-18 05:31:51,298 INFO ipc.Server (Server.java:logException(2428)) > - IPC Server handler 23 on 8020, call > org.apache.hadoop.hdfs.protocol.ClientProtocol.addBlock from > 172.27.18.201:36652 Call#1959 Retry#0: > org.apache.hadoop.hdfs.server.namenode.LeaseExpiredException: No lease on > /hdp/apps/2.6.3.0-61/pig/pig.tar.gz (inode 16561): File does not exist. > Holder DFSClient_NONMAPREDUCE_1849462310_141 does not have any open files. > 2017-07-18 05:31:51,800 INFO hdfs.StateChange > (FSNamesystem.java:logAllocatedBlock(3831)) - BLOCK* allocate > blk_1073741850_1026, replicas=172.27.19.4:1019, 172.27.52.76:1019, > 172.27.18.201:1019 for /hdp/apps/2.6.3.0-61/pig/pig.tar.gz > 2017-07-18 05:31:51,823 INFO hdfs.StateChange > (FSNamesystem.java:completeFile(3724)) - DIR* completeFile: > /hdp/apps/2.6.3.0-61/pig/pig.tar.gz is closed by > DFSClient_NONMAPREDUCE_307451118_147 > -- This message was sent by Atlassian JIRA (v6.4.14#64029)