Do you have any nodes down? Bad disks? Try to recover those nodes. What level of replication were you running with? btw this is really an HDFS issue, not a pig issue, so you'll likely get better support from the hdfs user list.
D On Thu, Oct 6, 2011 at 11:02 PM, kiranprasad <[email protected]>wrote: > > Hi Alex > > Thanks for your response. > > I ve checked with below mentioned command and I am getting > > [kiranprasad.g@pig4 hadoop-0.20.2]$ bin/hadoop fs -text > /data/arpumsisdn.txt | tail > 11/10/07 16:17:18 INFO hdfs.DFSClient: No node available for block: > blk_-8354424441116992221_1060 file=/data/arpumsisdn.txt > 11/10/07 16:17:18 INFO hdfs.DFSClient: Could not obtain block > blk_-8354424441116992221_1060 from any node: java.io.IOException: No live > nodes contain current block > 11/10/07 16:17:21 INFO hdfs.DFSClient: No node available for block: > blk_-8354424441116992221_1060 file=/data/arpumsisdn.txt > 11/10/07 16:17:21 INFO hdfs.DFSClient: Could not obtain block > blk_-8354424441116992221_1060 from any node: java.io.IOException: No live > nodes contain current block > 11/10/07 16:17:25 INFO hdfs.DFSClient: No node available for block: > blk_-8354424441116992221_1060 file=/data/arpumsisdn.txt > 11/10/07 16:17:25 INFO hdfs.DFSClient: Could not obtain block > blk_-8354424441116992221_1060 from any node: java.io.IOException: No live > nodes contain current block > 11/10/07 16:17:29 WARN hdfs.DFSClient: DFS Read: java.io.IOException: Could > not obtain block: blk_-8354424441116992221_1060 file=/data/arpumsisdn.txt > at org.apache.hadoop.hdfs.**DFSClient$DFSInputStream.** > chooseDataNode(DFSClient.java:**1812) > at org.apache.hadoop.hdfs.**DFSClient$DFSInputStream.** > blockSeekTo(DFSClient.java:**1638) > at org.apache.hadoop.hdfs.**DFSClient$DFSInputStream.read(** > DFSClient.java:1767) > at org.apache.hadoop.hdfs.**DFSClient$DFSInputStream.read(** > DFSClient.java:1695) > at java.io.DataInputStream.**readShort(DataInputStream.**java:295) > at org.apache.hadoop.fs.FsShell.**forMagic(FsShell.java:397) > at org.apache.hadoop.fs.FsShell.**access$200(FsShell.java:49) > at org.apache.hadoop.fs.FsShell$**2.process(FsShell.java:420) > at org.apache.hadoop.fs.FsShell$**DelayedExceptionThrowing.** > globAndProcess(FsShell.java:**1898) > at org.apache.hadoop.fs.FsShell.**text(FsShell.java:414) > at org.apache.hadoop.fs.FsShell.**doall(FsShell.java:1563) > at org.apache.hadoop.fs.FsShell.**run(FsShell.java:1763) > at org.apache.hadoop.util.**ToolRunner.run(ToolRunner.**java:65) > at org.apache.hadoop.util.**ToolRunner.run(ToolRunner.**java:79) > at org.apache.hadoop.fs.FsShell.**main(FsShell.java:1880) > > text: Could not obtain block: blk_-8354424441116992221_1060 > file=/data/arpumsisdn.txt > > > The Block is not available. How to recover the data block ? > > > -----Original Message----- From: Alex Rovner > Sent: Wednesday, October 05, 2011 5:55 PM > To: [email protected] > Subject: Re: ERROR 1066: Unable to open iterator for alias A. Backend error > : Could not obtain block: > > > You can also test quickly if thats the issue by running the following > command: > > hadoop fs -text /data/arpumsisdn.txt | tail > > On Wed, Oct 5, 2011 at 8:24 AM, Alex Rovner <[email protected]> wrote: > > Kiran, >> >> This looks like your HDFS is missing some blocks. Can you run fsck and see >> if you have missing blocks and if so for what files? >> >> http://hadoop.apache.org/**common/docs/r0.17.2/hdfs_user_** >> guide.html#Fsck<http://hadoop.apache.org/common/docs/r0.17.2/hdfs_user_guide.html#Fsck> >> >> Alex >> >> >> On Tue, Oct 4, 2011 at 7:53 AM, kiranprasad <[email protected]> >> **wrote: >> >> I am getting the below exception when trying to execute PIG latin script. >>> >>> Failed! >>> >>> Failed Jobs: >>> JobId Alias Feature Message Outputs >>> job_201110042009_0005 A MAP_ONLY Message: Job failed! >>> >>> hdfs://10.0.0.61/tmp/**temp1751671187/tmp-592386019<http://10.0.0.61/tmp/temp1751671187/tmp-592386019> >>> , >>> >>> Input(s): >>> Failed to read data from "/data/arpumsisdn.txt" >>> >>> Output(s): >>> Failed to produce result in "hdfs:// >>> 10.0.0.61/tmp/temp1751671187/**tmp-592386019<http://10.0.0.61/tmp/temp1751671187/tmp-592386019> >>> " >>> >>> Counters: >>> Total records written : 0 >>> Total bytes written : 0 >>> Spillable Memory Manager spill count : 0 >>> Total bags proactively spilled: 0 >>> Total records proactively spilled: 0 >>> >>> Job DAG: >>> job_201110042009_0005 >>> >>> >>> 2011-10-04 22:13:53,736 [main] INFO >>> >>> org.apache.pig.backend.hadoop.**executionengine.**mapReduceLayer.** >>> MapReduceLauncher >>> - Failed! >>> 2011-10-04 22:13:53,745 [main] ERROR org.apache.pig.tools.grunt.**Grunt >>> - >>> ERROR 1066: Unable to open iterator for alias A. Backend error : Could >>> not >>> obtain block: blk_-8354424441116992221_1060 file=/data/arpumsisdn.txt >>> Details at logfile: /home/kiranprasad.g/pig-0.8.1/** >>> pig_1317746514798.log >>> >>> >>> >>> Regards >>> Kiran.G >>> >> >> >> >> > >
