Lin Zhang created HDFS-14360:
--------------------------------
Summary: some excptioins happened while using ISA-L
Key: HDFS-14360
URL: https://issues.apache.org/jira/browse/HDFS-14360
Project: Hadoop HDFS
Issue Type: Bug
Components: ec, erasure-coding
Reporter: Lin Zhang
I built my hadoop with isa-l supported. When I try to so some convert job,
exception happens.
{code:java}
//代码占位符
{code}
[2019-03-12T11:39:03.183+08:00] [INFO]
[1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger)] :
1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger) : #
[2019-03-12T11:39:03.184+08:00] [INFO]
[1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger)] :
1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger) : # A fatal error has been
detected by the Java Runtime Environment: [2019-03-12T11:39:03.184+08:00]
[INFO] [1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger)] :
1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger) : #
[2019-03-12T11:39:03.184+08:00] [INFO]
[1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger)] :
1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger) : # SIGSEGV (0xb) at
pc=0x00007fc42e182683, pid=17110, tid=0x00007fc40ce9f700
[2019-03-12T11:39:03.184+08:00] [INFO]
[1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger)] :
1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger) : #
[2019-03-12T11:39:03.184+08:00] [INFO]
[1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger)] :
1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger) : # JRE version: Java(TM)
SE Runtime Environment (8.0_121-b13) (build 1.8.0_121-b13)
[2019-03-12T11:39:03.184+08:00] [INFO]
[1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger)] :
1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger) : # Java VM: Java
HotSpot(TM) 64-Bit Server VM (25.121-b13 mixed mode linux-amd64 compressed
oops) [2019-03-12T11:39:03.184+08:00] [INFO]
[1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger)] :
1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger) : # Problematic frame:
[2019-03-12T11:39:03.184+08:00] [INFO]
[1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger)] :
1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger) : # V [libjvm.so+0x9bd683]
SafepointSynchronize::begin()+0x263 [2019-03-12T11:39:03.185+08:00] [INFO]
[1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger)] :
1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger) : #
[2019-03-12T11:39:03.185+08:00] [INFO]
[1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger)] :
1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger) : # Failed to write core
dump. Core dumps have been disabled. To enable core dumping, try "ulimit -c
unlimited" before starting Java again [2019-03-12T11:39:03.185+08:00] [INFO]
[1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger)] :
1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger) : #
[2019-03-12T11:39:03.185+08:00] [INFO]
[1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger)] :
1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger) : # An error report file
with more information is saved as: [2019-03-12T11:39:03.185+08:00] [INFO]
[1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger)] :
1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger) : #
/software/servers/hadoop-2.7.1/hs_err_pid17110.log
[2019-03-12T11:39:03.191+08:00] [INFO]
[1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger)] :
1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger) : #
[2019-03-12T11:39:03.191+08:00] [INFO]
[1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger)] :
1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger) : # If you would like to
submit a bug report, please visit: [2019-03-12T11:39:03.191+08:00] [INFO]
[1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger)] :
1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger) : #
http://bugreport.java.com/bugreport/crash.jsp [2019-03-12T11:39:03.191+08:00]
[INFO] [1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger)] :
1552362147634_CONVERT_CMD/test/zhanglin/1g(isLogger) : #
[2019-03-12T11:39:07.949+08:00] [ERROR] [pool-10-thread-1] : copy file
/test/zhanglin/1g to /test/ttlconverter/factory/test/zhanglin/1gfailed
[2019-03-12T11:39:07.949+08:00] [INFO] [DataXceiver for client
DFSClient_NONMAPREDUCE_1740978034_1 at /172.22.176.69:40662 [Receiving block
BP-442378117-172.16.150.142-1552360340470:blk_-9223372036854775792_1009]] :
Exception for
BP-442378117-172.16.150.142-1552360340470:blk_-9223372036854775792_1009
java.io.IOException: Premature EOF from inputStream at
org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:212) at
org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:213)
at
org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:134)
at
org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:109)
at
org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:529)
at
org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:972)
at
org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:891)
at
org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:171)
at
org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:105)
at
org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:290) at
java.lang.Thread.run(Thread.java:745)
[2019-03-12T11:39:07.951+08:00] [INFO] [DataXceiver for client
DFSClient_NONMAPREDUCE_1740978034_1 at /172.22.176.69:40660 [Sending block
BP-442378117-172.16.150.142-1552360340470:blk_1073741825_1001]] : Scheduling a
check for /data0/dfs
[2019-03-12T11:39:07.954+08:00] [INFO] [PacketResponder:
BP-442378117-172.16.150.142-1552360340470:blk_-9223372036854775792_1009,
type=LAST_IN_PIPELINE] : PacketResponder:
BP-442378117-172.16.150.142-1552360340470:blk_-9223372036854775792_1009,
type=LAST_IN_PIPELINE: Thread is interrupted.
[2019-03-12T11:39:07.954+08:00] [INFO] [PacketResponder:
BP-442378117-172.16.150.142-1552360340470:blk_-9223372036854775792_1009,
type=LAST_IN_PIPELINE] : PacketResponder:
BP-442378117-172.16.150.142-1552360340470:blk_-9223372036854775792_1009,
type=LAST_IN_PIPELINE terminating
[2019-03-12T11:39:07.954+08:00] [INFO] [DataXceiver for client
DFSClient_NONMAPREDUCE_1740978034_1 at /172.22.176.69:40662 [Receiving block
BP-442378117-172.16.150.142-1552360340470:blk_-9223372036854775792_1009]] :
opWriteBlock
BP-442378117-172.16.150.142-1552360340470:blk_-9223372036854775792_1009
received exception java.io.IOException: Premature EOF from inputStream
[2019-03-12T11:39:07.957+08:00] [ERROR] [DataXceiver for client
DFSClient_NONMAPREDUCE_1740978034_1 at /172.22.176.69:40662 [Receiving block
BP-442378117-172.16.150.142-1552360340470:blk_-9223372036854775792_1009]] :
A01-R02-I176-69-4CY8S12.JD.LOCAL:50010:DataXceiver error processing WRITE_BLOCK
operation src: /172.22.176.69:40662 dst: /172.22.176.69:50010
java.io.IOException: Premature EOF from inputStream
at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:212)
at
org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:213)
at
org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:134)
at
org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:109)
at
org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:529)
at
org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:972)
at
org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:891)
at
org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:171)
at
org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:105)
at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:290)
at java.lang.Thread.run(Thread.java:745)
I checked my native env, which showed below
{code:java}
//代码占位符
{code}
Native library checking:
hadoop: true /software/servers/hadoop-2.7.1/lib/native/libhadoop.so.1.0.0
zlib: true /lib64/libz.so.1
snappy: true /lib64/libsnappy.so.1
lz4: true revision:99
bzip2: true /lib64/libbz2.so.1
openssl: true /lib64/libcrypto.so
ISA-L: true /software/servers/hadoop-2.7.1/lib/native/libisal.so.2
--
This message was sent by Atlassian JIRA
(v7.6.3#76005)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]