C G,
Any specifics on how you reproduce any of these issues will be helpful.
I was able to copy a 5GB file without errors. copyFromLocal just copies
raw file content. Not sure of what '5,000,000 rows' means.
Raghu.
C G wrote:
Further experimentation, again single node configuration on a 4way 8G machine
w/0.14.0, trying to copyFromLocal 669M of data in 5,000,000 rows I see this in
the namenode log:
2007-08-24 00:50:45,902 WARN org.apache.hadoop.dfs.StateChange: DIR* NameSystem.completeFile: failed to complete /input/t.dat because dir.getFileBlocks() is non-null and pendingFile is null
2007-08-24 00:50:48,000 INFO org.apache.hadoop.ipc.Server: IPC Server handler 9
on 54310, call complete(/input/t.dat, DFSClient_-2013541261) from
XXX.XXX.XXX.XX:36470: error: java.io.IOException: Could not complete write to
file /input/t.dat by DFSClient_-2013541261
java.io.IOException: Could not complete write to file /input/t.dat by
DFSClient_-2013541261
at org.apache.hadoop.dfs.NameNode.complete(NameNode.java:359)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:585)
at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:340)
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:566)
Any thoughts or help appreciated...I'm planning to build out a large grid
running terabytes of data...assuming I can get it Hadoop to handle more than
500M :-(.
Thanks!
Raghu Angadi <[EMAIL PROTECTED]> wrote:
Regd the second problem :
It is surprising that this fails repeatedly around the same place. 0.14
does check the checksum at the datanode (0.13 did not do this check). I
will try to reproduce this.
Raghu.
C G wrote:
Hi All:
Second issue is a failure on copyFromLocal with lost connections. I'm trying to
copy a 5.8G, 88,784,045 million row file to HDFS. It makes progress for a
while, but at approx 2.1 gigs copied, it dies with a repeated series of errors.
There is 470G free on the file system. The error is repeated several times and
is:
$ bin/hadoop dfs -copyFromLocal sample.dat /input/sample.dat
07/08/23 15:58:10 WARN fs.DFSClient: Error while writing.
java.net.SocketException: Connection reset
at java.net.SocketOutputStream.socketWrite(SocketOutputStream.java:96)
at java.net.SocketOutputStream.write(SocketOutputStream.java:136)
at java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:65)
at java.io.BufferedOutputStream.write(BufferedOutputStream.java:109)
at java.io.DataOutputStream.write(DataOutputStream.java:90)
at org.apache.hadoop.dfs.DFSClient$DFSOutputStream.endBlock(DFSClient.java:1656)
at
org.apache.hadoop.dfs.DFSClient$DFSOutputStream.writeChunk(DFSClient.java:1610)
at
org.apache.hadoop.fs.FSOutputSummer.writeChecksumChunk(FSOutputSummer.java:140)
at org.apache.hadoop.fs.FSOutputSummer.write1(FSOutputSummer.java:100)
at org.apache.hadoop.fs.FSOutputSummer.write(FSOutputSummer.java:86)
at
org.apache.hadoop.fs.FSDataOutputStream$PositionCache.write(FSDataOutputStream.java:39)
at java.io.DataOutputStream.write(DataOutputStream.java:90)
at org.apache.hadoop.fs.FileUtil.copyContent(FileUtil.java:258)
at org.apache.hadoop.fs.FileUtil.copyContent(FileUtil.java:248)
at org.apache.hadoop.fs.FileUtil.copy(FileUtil.java:133)
at org.apache.hadoop.fs.FileSystem.copyFromLocalFile(FileSystem.java:776)
at org.apache.hadoop.fs.FileSystem.copyFromLocalFile(FileSystem.java:757)
at org.apache.hadoop.fs.FsShell.copyFromLocal(FsShell.java:116)
at org.apache.hadoop.fs.FsShell.run(FsShell.java:1229)
at org.apache.hadoop.util.ToolBase.doMain(ToolBase.java:187)
at org.apache.hadoop.fs.FsShell.main(FsShell.java:1342)
The following error also appears several times in the datanode logs:
2007-08-23 15:58:10,072 ERROR org.apache.hadoop.dfs.DataNode: DataXceiver:
java.io.IOException: Unexpected checksum mismatch while writing
blk_1461965301876815406 from /xxx.xxx.xxx.xx:50960
at org.apache.hadoop.dfs.DataNode$DataXceiver.writeBlock(DataNode.java:902)
at org.apache.hadoop.dfs.DataNode$DataXceiver.run(DataNode.java:727)
at java.lang.Thread.run(Thread.java:595)
Any help on these issues much appreciated.
---------------------------------
Luggage? GPS? Comic books?
Check out fitting gifts for grads at Yahoo! Search.
---------------------------------
Pinpoint customers who are looking for what you sell.
---------------------------------
Pinpoint customers who are looking for what you sell.