------ Original Message ------
Date: Wednesday, Oct 4, 2023 01:34:02 PM
From: "Kiyoshi Mizumaru" <kiyoshi.mizum...@gmail.com>
To: "Harry Jamison" <harryjamiso...@yahoo.com.invalid>
Cc: "Hadoop Users" <user@hadoop.apache.org>
Subject: Re: HDFS HA standby
Why don't you try to change the logging level? DEBUG or TRACE would be helpful.On Wed, Oct 4, 2023 at 1:13 PM Harry Jamison <harryjamiso...@yahoo.com.invalid> wrote:I am not sure exactly what the problem is now.My namenode (and I think journal node are getting shut down.Is there a way to tell Why it is getting the shutdown signal?Also the datanode seems to be getting this error
End of File Exception between local host isHere are the logs, and I only see INFO logging, and then a the Shutdown[2023-10-03 20:53:00,873] INFO Initializing quota with 12 thread(s) (org.apache.hadoop.hdfs.server.namenode.FSDirectory)
[2023-10-03 20:53:00,876] INFO Quota initialization completed in 1 milliseconds
name space=2
storage space=0
storage types=RAM_DISK=0, SSD=0, DISK=0, ARCHIVE=0, PROVIDED=0 (org.apache.hadoop.hdfs.server.namenode.FSDirectory)
[2023-10-03 20:53:00,882] INFO Total number of blocks = 0 (org.apache.hadoop.hdfs.server.blockmanagement.BlockManager)
[2023-10-03 20:53:00,884] INFO Starting CacheReplicationMonitor with interval 30000 milliseconds (org.apache.hadoop.hdfs.server.blockmanagement.CacheReplicationMonitor)
[2023-10-03 20:53:00,884] INFO Number of invalid blocks = 0 (org.apache.hadoop.hdfs.server.blockmanagement.BlockManager)
[2023-10-03 20:53:00,884] INFO Number of under-replicated blocks = 0 (org.apache.hadoop.hdfs.server.blockmanagement.BlockManager)
[2023-10-03 20:53:00,884] INFO Number of over-replicated blocks = 0 (org.apache.hadoop.hdfs.server.blockmanagement.BlockManager)
[2023-10-03 20:53:00,884] INFO Number of blocks being written = 0 (org.apache.hadoop.hdfs.server.blockmanagement.BlockManager)
[2023-10-03 20:53:00,884] INFO STATE* Replication Queue initialization scan for invalid, over- and under-replicated blocks completed in 67 msec (org.apache.hadoop.hdfs.StateChange)
[2023-10-03 20:54:16,453] ERROR RECEIVED SIGNAL 15: SIGTERM (org.apache.hadoop.hdfs.server.namenode.NameNode)
[2023-10-03 20:54:16,467] INFO SHUTDOWN_MSG:
/-***********************************************************
SHUTDOWN_MSG: Shutting down NameNode at vmnode1/192.168.1.159
*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-/ (org.apache.hadoop.hdfs.server.namenode.NameNode)
When I start the data node I see this[2023-10-03 20:53:00,882] INFO Namenode Block pool BP-1620264838-192.168.1.159-1696370857417 (Datanode Uuid 66068658-b08b-49cd-aba0-56ac1f29e7d5) service to vmnode1/192.168.1.159:8020 trying to claim ACTIVE state with txid=15 (org.apache.hadoop.hdfs.server.datanode.DataNode)
[2023-10-03 20:53:00,882] INFO Acknowledging ACTIVE Namenode Block pool BP-1620264838-192.168.1.159-1696370857417 (Datanode Uuid 66068658-b08b-49cd-aba0-56ac1f29e7d5) service to vmnode1/192.168.1.159:8020 (org.apache.hadoop.hdfs.server.datanode.DataNode)
[2023-10-03 20:53:00,882] INFO After receiving heartbeat response, updating state of namenode vmnode1:8020 to active (org.apache.hadoop.hdfs.server.datanode.DataNode)
[2023-10-03 20:54:18,771] WARN IOException in offerService (org.apache.hadoop.hdfs.server.datanode.DataNode)
java.io.EOFException: End of File Exception between local host is: "vmnode1/192.168.1.159"; destination host is: "vmnode1":8020; : java.io.EOFException; For more details see: http://wiki.apache.org/hadoop/EOFException
at java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
at java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
at java.base/jdk.internal.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
at java.base/java.lang.reflect.Constructor.newInstance(Constructor.java:490)
at org.apache.hadoop.net.NetUtils.wrapWithMessage(NetUtils.java:930)
at org.apache.hadoop.net.NetUtils.wrapException(NetUtils.java:879)
at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1571)
at org.apache.hadoop.ipc.Client.call(Client.java:1513)
at org.apache.hadoop.ipc.Client.call(Client.java:1410)
at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:258)
at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:139)
at com.sun.proxy.$Proxy19.sendHeartbeat(Unknown Source)
at org.apache.hadoop.hdfs.protocolPB.DatanodeProtocolClientSideTranslatorPB.sendHeartbeat(DatanodeProtocolClientSideTranslatorPB.java:168)
at org.apache.hadoop.hdfs.server.datanode.BPServiceActor.sendHeartBeat(BPServiceActor.java:562)
at org.apache.hadoop.hdfs.server.datanode.BPServiceActor.offerService(BPServiceActor.java:710)
at org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:920)
at java.base/java.lang.Thread.run(Thread.java:829)
Caused by: java.io.EOFException
at java.base/java.io.DataInputStream.readInt(DataInputStream.java:397)
at org.apache.hadoop.ipc.Client$IpcStreams.readResponse(Client.java:1906)
at org.apache.hadoop.ipc.Client$Connection.receiveRpcResponse(Client.java:1187)
at org.apache.hadoop.ipc.Client$Connection.run(Client.java:1078)
unsubscribe me
- unsubscribe me Man-Young Goo
- Unsubscribe me Man-Young Goo
- unsubscribe me Man-Young Goo