Ashok created IMPALA-8972:
-----------------------------

             Summary: Impala deamon crashing frequently
                 Key: IMPALA-8972
                 URL: https://issues.apache.org/jira/browse/IMPALA-8972
             Project: IMPALA
          Issue Type: Question
          Components: Infrastructure
    Affects Versions: Impala 2.8.0
         Environment: Impala version 2.8.0-cdh5-INTERNAL RELEASE (build )
            Reporter: Ashok


Hi Team,

 

Impala deamon is crashing frequently and need to restart .

 

Please help in troubleshooting the same 

 

I could see below error messages in deamon logs

 

1.

 

Java exception follows:
org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.hdfs.server.namenode.LeaseExpiredException):
 No lease on /user/hive/warehouse/steelwedge_psnokiadmt_p
rod.db/ppd_ro_im_dpart_bkp/_impala_insert_staging/e847d4231bb8c531_c166c98d00000000/.e847d4231bb8c531-c166c98d00000002_664317806_dir/e847d4231bb8c531-c166c98
d00000002_844965293_data.0.parq (inode 17854099): File does not exist. Holder 
DFSClient_NONMAPREDUCE_-924590406_1 does not have any open files.
 at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkLease(FSNamesystem.java:3635)
 at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.analyzeFileState(FSNamesystem.java:3438)
 at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:3294)
 at 
org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:679)
 at 
org.apache.hadoop.hdfs.server.namenode.AuthorizationProviderProxyClientProtocol.addBlock(AuthorizationProviderProxyClientProtocol.java:214)
 at 
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:489)
 at 
org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
 at 
org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:617)
 at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1073)
 at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2086)
 at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2082)
 at java.security.AccessController.doPrivileged(Native Method)
 at javax.security.auth.Subject.doAs(Subject.java:415)
 at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1698)
 at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2080)

at org.apache.hadoop.ipc.Client.call(Client.java:1472)
 at org.apache.hadoop.ipc.Client.call(Client.java:1409)
 at 
org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:230)
 at com.sun.proxy.$Proxy9.addBlock(Unknown Source)
 at 
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.addBlock(ClientNamenodeProtocolTranslatorPB.java:409)
 at sun.reflect.GeneratedMethodAccessor7.invoke(Unknown Source)
 at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:606)
 at 
org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:256)
 at 
org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:104)
 at com.sun.proxy.$Proxy10.addBlock(Unknown Source)
 at 
org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.locateFollowingBlock(DFSOutputStream.java:1739)
 at 
org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.nextBlockOutputStream(DFSOutputStream.java:1535)
 at 
org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:689)
Wrote minidump to 
/var/log/impala/minidumps/impalad/6352d57e-7493-b4db-27e7f36f-518eec8e.dmp
#
# A fatal error has been detected by the Java Runtime Environment:
#
# SIGSEGV (0xb) at pc=0x00007f42744b72fc, pid=4881, tid=139899197028096
#
# JRE version: Java(TM) SE Runtime Environment (7.0_80-b15) (build 1.7.0_80-b15)
# Java VM: Java HotSpot(TM) 64-Bit Server VM (24.80-b11 mixed mode linux-amd64 
compressed oops)
# Problematic frame:
# C [libkudu_client.so.0+0x27d2fc] void std::_Rb_tree<std::string, 
std::pair<std::string const, std::string>, 
std::_Select1st<std::pair<std::string const,
std::string> >, std::less<std::string>, std::allocator<std::pair<std::string 
const, std::string> > >::_M_insert_unique<std::_Rb_tree_iterator<std::pair<std::
string const, std::string> > >(std::_Rb_tree_iterator<std::pair<std::string 
const, std::string> >, std::_Rb_tree_iterator<std::pair<std::string const, 
std::s
tring> >)+0x2381c

 

2.

 

W0917 22:35:49.505252  1265 BlockReaderFactory.java:778] I/O error constructing 
remote block reader.W0917 22:35:49.505252  1265 BlockReaderFactory.java:778] 
I/O error constructing remote block reader.Java exception 
follows:java.io.IOException: Got error for OP_READ_BLOCK, status=ERROR, 
self=/10.111.92.61:46531, remote=/10.111.92.61:50010, for file 
/user/hive/warehouse/steelwedge_psnokiadmt_prod.db/graph_ppt_in_list/job_id=149042/ea46e454a6357b08-2622f21800000002_731828525_data.0.,
 for pool BP-1380753826-10.128.50.16-1462783635263 block 1081077775_7337065 at 
org.apache.hadoop.hdfs.RemoteBlockReader2.checkSuccess(RemoteBlockReader2.java:467)
 at 
org.apache.hadoop.hdfs.RemoteBlockReader2.newBlockReader(RemoteBlockReader2.java:432)
 at 
org.apache.hadoop.hdfs.BlockReaderFactory.getRemoteBlockReader(BlockReaderFactory.java:881)
 at 
org.apache.hadoop.hdfs.BlockReaderFactory.getRemoteBlockReaderFromTcp(BlockReaderFactory.java:759)
 at 
org.apache.hadoop.hdfs.BlockReaderFactory.build(BlockReaderFactory.java:376) at 
org.apache.hadoop.hdfs.DFSInputStream.blockSeekTo(DFSInputStream.java:662) at 
org.apache.hadoop.hdfs.DFSInputStream.readWithStrategy(DFSInputStream.java:889) 
at org.apache.hadoop.hdfs.DFSInputStream.read(DFSInputStream.java:965) at 
org.apache.hadoop.fs.FSDataInputStream.read(FSDataInputStream.java:147)W0917 
22:35:49.506021  1265 DFSInputStream.java:699] Failed to connect to 
/10.111.92.61:50010 for block, add to deadNodes and continue. 
java.io.IOException: Got error for OP_READ_BLOCK, status=ERROR, 
self=/10.111.92.61:46531, remote=/10.111.92.61:50010, for file 
/user/hive/warehouse/steelwedge_psnokiadmt_prod.db/graph_ppt_in_list/job_id=149042/ea46e454a6357b08-2622f21800000002_731828525_data.0.,
 for pool BP-1380753826-10.128.50.16-1462783635263 block 1081077775_7337065Java 
exception follows:java.io.IOException: Got Aborting Impala for OP_READ_BLOCK, 
status=ERROR, self=/10.111.92.61:46531, remote=/10.111.92.61:50010, for file 
/user/hive/warehouse/steelwedge_psnokiadmt_prod.db/graph_ppt_in_list/job_id=149042/ea46e454a6357b08-2622f21800000002_731828525_data.0.,
 for pool BP-1380753826-10.128.50.16-1462783635263 block 1081077775_7337065 at 
org.apache.hadoop.hdfs.RemoteBlockReader2.checkSuccess(RemoteBlockReader2.java:467)
 at 
org.apache.hadoop.hdfs.RemoteBlockReader2.newBlockReader(RemoteBlockReader2.java:432)
 at 
org.apache.hadoop.hdfs.BlockReaderFactory.getRemoteBlockReader(BlockReaderFactory.java:881)
 at 
org.apache.hadoop.hdfs.BlockReaderFactory.getRemoteBlockReaderFromTcp(BlockReaderFactory.java:759)
 at 
org.apache.hadoop.hdfs.BlockReaderFactory.build(BlockReaderFactory.java:376) at 
org.apache.hadoop.hdfs.DFSInputStream.blockSeekTo(DFSInputStream.java:662) at 
org.apache.hadoop.hdfs.DFSInputStream.readWithStrategy(DFSInputStream.java:889) 
at org.apache.hadoop.hdfs.DFSInputStream.read(DFSInputStream.java:965) at 
org.apache.hadoop.fs.FSDataInputStream.read(FSDataInputStream.java:147)



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to