Hi, I am trying to run some of the binaries in libhdfs, in particular hdfs_write. My setup: - I am using hadoop-0.15.3 - I started namenode/datanode on the localhost - I compiled libhdfs for an x86_64 setting
I invoke hdfs_write as: hdfs_write foo.1 128000000 65536 When I run the hdfs_write test program, the program sees the following exception: java.io.IOException: Mismatch in writeChunk() args at org.apache.hadoop.dfs.DFSClient$DFSOutputStream.writeChunk(DFSClient.java:1588) at org.apache.hadoop.fs.FSOutputSummer.writeChecksumChunk(FSOutputSummer.java:140) at org.apache.hadoop.fs.FSOutputSummer.write1(FSOutputSummer.java:100) at org.apache.hadoop.fs.FSOutputSummer.write(FSOutputSummer.java:86) at org.apache.hadoop.fs.FSDataOutputStream$PositionCache.write(FSDataOutputStream.java:39) at java.io.DataOutputStream.write(libgcj.so.8rh) at java.io.FilterOutputStream.write(libgcj.so.8rh) Call to org.apache.hadoop.fs.FSDataOutputStream::write failed! When do "bin/hadoop dfs -ls", it shows the file foo.1 as being created. What gives? Sriram