Re: Bad connect ack with firstBadLink

2012-05-07 Thread madhu phatak
Hi,
 Increasing the open file limit solved the issue. Thank you.

On Fri, May 4, 2012 at 9:39 PM, Mapred Learn  wrote:

> Check your number of blocks in the cluster.
>
> This also indicates that your datanodes are more full than they should be.
>
> Try deleting unnecessary blocks.
>
> On Fri, May 4, 2012 at 7:40 AM, Mohit Anchlia  >wrote:
>
> > Please see:
> >
> > http://hbase.apache.org/book.html#dfs.datanode.max.xcievers
> >
> > On Fri, May 4, 2012 at 5:46 AM, madhu phatak 
> wrote:
> >
> > > Hi,
> > > We are running a three node cluster . From two days whenever we copy
> file
> > > to hdfs , it is throwing  java.IO.Exception Bad connect ack with
> > > firstBadLink . I searched in net, but not able to resolve the issue.
> The
> > > following is the stack trace from datanode log
> > >
> > > 2012-05-04 18:08:08,868 INFO
> > > org.apache.hadoop.hdfs.server.datanode.DataNode: writeBlock
> > > blk_-7520371350112346377_50118 received exception
> > java.net.SocketException:
> > > Connection reset
> > > 2012-05-04 18:08:08,869 ERROR
> > > org.apache.hadoop.hdfs.server.datanode.DataNode: DatanodeRegistration(
> > > 172.23.208.17:50010,
> > > storageID=DS-1340171424-172.23.208.17-50010-1334672673051,
> > infoPort=50075,
> > > ipcPort=50020):DataXceiver
> > > java.net.SocketException: Connection reset
> > >at java.net.SocketInputStream.read(SocketInputStream.java:168)
> > >at
> java.io.BufferedInputStream.read1(BufferedInputStream.java:256)
> > >at
> java.io.BufferedInputStream.read(BufferedInputStream.java:317)
> > >at java.io.DataInputStream.read(DataInputStream.java:132)
> > >at
> > >
> > >
> >
> org.apache.hadoop.hdfs.server.datanode.BlockReceiver.readToBuf(BlockReceiver.java:262)
> > >at
> > >
> > >
> >
> org.apache.hadoop.hdfs.server.datanode.BlockReceiver.readNextPacket(BlockReceiver.java:309)
> > >at
> > >
> > >
> >
> org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:373)
> > >at
> > >
> > >
> >
> org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:525)
> > >at
> > >
> > >
> >
> org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:357)
> > >at
> > >
> > >
> >
> org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:103)
> > >at java.lang.Thread.run(Thread.java:662)
> > >
> > >
> > > It will be great if some one can point to the direction how to solve
> this
> > > problem.
> > >
> > > --
> > > https://github.com/zinnia-phatak-dev/Nectar
> > >
> >
>



-- 
https://github.com/zinnia-phatak-dev/Nectar


Re: Bad connect ack with firstBadLink

2012-05-04 Thread Mapred Learn
Check your number of blocks in the cluster.

This also indicates that your datanodes are more full than they should be.

Try deleting unnecessary blocks.

On Fri, May 4, 2012 at 7:40 AM, Mohit Anchlia wrote:

> Please see:
>
> http://hbase.apache.org/book.html#dfs.datanode.max.xcievers
>
> On Fri, May 4, 2012 at 5:46 AM, madhu phatak  wrote:
>
> > Hi,
> > We are running a three node cluster . From two days whenever we copy file
> > to hdfs , it is throwing  java.IO.Exception Bad connect ack with
> > firstBadLink . I searched in net, but not able to resolve the issue. The
> > following is the stack trace from datanode log
> >
> > 2012-05-04 18:08:08,868 INFO
> > org.apache.hadoop.hdfs.server.datanode.DataNode: writeBlock
> > blk_-7520371350112346377_50118 received exception
> java.net.SocketException:
> > Connection reset
> > 2012-05-04 18:08:08,869 ERROR
> > org.apache.hadoop.hdfs.server.datanode.DataNode: DatanodeRegistration(
> > 172.23.208.17:50010,
> > storageID=DS-1340171424-172.23.208.17-50010-1334672673051,
> infoPort=50075,
> > ipcPort=50020):DataXceiver
> > java.net.SocketException: Connection reset
> >at java.net.SocketInputStream.read(SocketInputStream.java:168)
> >at java.io.BufferedInputStream.read1(BufferedInputStream.java:256)
> >at java.io.BufferedInputStream.read(BufferedInputStream.java:317)
> >at java.io.DataInputStream.read(DataInputStream.java:132)
> >at
> >
> >
> org.apache.hadoop.hdfs.server.datanode.BlockReceiver.readToBuf(BlockReceiver.java:262)
> >at
> >
> >
> org.apache.hadoop.hdfs.server.datanode.BlockReceiver.readNextPacket(BlockReceiver.java:309)
> >at
> >
> >
> org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:373)
> >at
> >
> >
> org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:525)
> >at
> >
> >
> org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:357)
> >at
> >
> >
> org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:103)
> >at java.lang.Thread.run(Thread.java:662)
> >
> >
> > It will be great if some one can point to the direction how to solve this
> > problem.
> >
> > --
> > https://github.com/zinnia-phatak-dev/Nectar
> >
>


Re: Bad connect ack with firstBadLink

2012-05-04 Thread Mohit Anchlia
Please see:

http://hbase.apache.org/book.html#dfs.datanode.max.xcievers

On Fri, May 4, 2012 at 5:46 AM, madhu phatak  wrote:

> Hi,
> We are running a three node cluster . From two days whenever we copy file
> to hdfs , it is throwing  java.IO.Exception Bad connect ack with
> firstBadLink . I searched in net, but not able to resolve the issue. The
> following is the stack trace from datanode log
>
> 2012-05-04 18:08:08,868 INFO
> org.apache.hadoop.hdfs.server.datanode.DataNode: writeBlock
> blk_-7520371350112346377_50118 received exception java.net.SocketException:
> Connection reset
> 2012-05-04 18:08:08,869 ERROR
> org.apache.hadoop.hdfs.server.datanode.DataNode: DatanodeRegistration(
> 172.23.208.17:50010,
> storageID=DS-1340171424-172.23.208.17-50010-1334672673051, infoPort=50075,
> ipcPort=50020):DataXceiver
> java.net.SocketException: Connection reset
>at java.net.SocketInputStream.read(SocketInputStream.java:168)
>at java.io.BufferedInputStream.read1(BufferedInputStream.java:256)
>at java.io.BufferedInputStream.read(BufferedInputStream.java:317)
>at java.io.DataInputStream.read(DataInputStream.java:132)
>at
>
> org.apache.hadoop.hdfs.server.datanode.BlockReceiver.readToBuf(BlockReceiver.java:262)
>at
>
> org.apache.hadoop.hdfs.server.datanode.BlockReceiver.readNextPacket(BlockReceiver.java:309)
>at
>
> org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:373)
>at
>
> org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:525)
>at
>
> org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:357)
>at
>
> org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:103)
>at java.lang.Thread.run(Thread.java:662)
>
>
> It will be great if some one can point to the direction how to solve this
> problem.
>
> --
> https://github.com/zinnia-phatak-dev/Nectar
>


Bad connect ack with firstBadLink

2012-05-04 Thread madhu phatak
Hi,
We are running a three node cluster . From two days whenever we copy file
to hdfs , it is throwing  java.IO.Exception Bad connect ack with
firstBadLink . I searched in net, but not able to resolve the issue. The
following is the stack trace from datanode log

2012-05-04 18:08:08,868 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: writeBlock
blk_-7520371350112346377_50118 received exception java.net.SocketException:
Connection reset
2012-05-04 18:08:08,869 ERROR
org.apache.hadoop.hdfs.server.datanode.DataNode: DatanodeRegistration(
172.23.208.17:50010,
storageID=DS-1340171424-172.23.208.17-50010-1334672673051, infoPort=50075,
ipcPort=50020):DataXceiver
java.net.SocketException: Connection reset
at java.net.SocketInputStream.read(SocketInputStream.java:168)
at java.io.BufferedInputStream.read1(BufferedInputStream.java:256)
at java.io.BufferedInputStream.read(BufferedInputStream.java:317)
at java.io.DataInputStream.read(DataInputStream.java:132)
at
org.apache.hadoop.hdfs.server.datanode.BlockReceiver.readToBuf(BlockReceiver.java:262)
at
org.apache.hadoop.hdfs.server.datanode.BlockReceiver.readNextPacket(BlockReceiver.java:309)
at
org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:373)
at
org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:525)
at
org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:357)
at
org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:103)
at java.lang.Thread.run(Thread.java:662)


It will be great if some one can point to the direction how to solve this
problem.

-- 
https://github.com/zinnia-phatak-dev/Nectar


Re: java.io.IOException: Bad connect ack with firstBadLink

2010-11-09 Thread Oleg Ruchovets
What does it mean:   It looks like second attempt to process data after the
first time it failed?
All Task AttemptsTask AttemptsMachineStatusProgressStart TimeShuffle
FinishedSort FinishedFinish TimeErrorsTask LogsCountersActions
attempt_201011081008_0002_r_08_0/default-rack/<http://hadoop1.infolinks.local:8022>
FAILED0.00%9-Nov-2010 04:36:159-Nov-2010 05:09:08 (32mins, 52sec)9-Nov-2010
05:09:08 (0sec)9-Nov-2010 05:14:07 (37mins, 51sec)

java.io.IOException: Bad connect ack with firstBadLink 10.11.87.65:50010
at 
org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.createBlockOutputStream(DFSClient.java:2870)
at 
org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.nextBlockOutputStream(DFSClient.java:2793)
at 
org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.access$2000(DFSClient.java:2076)
at 
org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$DataStreamer.run(DFSClient.java:2262)

Last 
4KB<http://hadoop1.infolinks.local:8022/tasklog?taskid=attempt_201011081008_0002_r_08_0&start=-4097>
Last 
8KB<http://hadoop1.infolinks.local:8022/tasklog?taskid=attempt_201011081008_0002_r_08_0&start=-8193>
All<http://hadoop1.infolinks.local:8022/tasklog?taskid=attempt_201011081008_0002_r_08_0&all=true>
10



attempt_201011081008_0002_r_08_1/default-rack/<http://hadoop-transfer.infolinks.local:8022>
RUNNING24.98%9-Nov-2010 05:50:21




On Tue, Nov 9, 2010 at 12:58 PM, Oleg Ruchovets wrote:

> Hi , running hadoop map/reduce got such exception?
>
>   1) Why does it happen?
>
>   2) Job didn't failed and continue it's execution? Does this exception cause 
> losing data or map/reduce
>
> uses recovery mechanism?
>
>
>
> 2010-11-09 05:10:08,735 INFO org.apache.hadoop.hdfs.DFSClient: Exception in 
> createBlockOutputStream java.io.IOException: Bad connect ack with 
> firstBadLink 10.11.87.65:50010
> 2010-11-09 05:10:08,735 INFO org.apache.hadoop.hdfs.DFSClient: Abandoning 
> block blk_-9208177033562590356_775948
> 2010-11-09 05:10:08,739 INFO org.apache.hadoop.hdfs.DFSClient: Waiting to 
> find target node: 10.11.87.61:50010
> 2010-11-09 05:11:23,743 INFO org.apache.hadoop.hdfs.DFSClient: Exception in 
> createBlockOutputStream java.net.SocketTimeoutException: 69000 millis timeout 
> while waiting for channel to be ready for read. ch : 
> java.nio.channels.SocketChannel[connected local=/10.11.87.61:55309 
> remote=/10.11.87.61:50010]
> 2010-11-09 05:11:23,743 INFO org.apache.hadoop.hdfs.DFSClient: Abandoning 
> block blk_-10251707095594311_775950
> 2010-11-09 05:11:23,744 INFO org.apache.hadoop.hdfs.DFSClient: Waiting to 
> find target node: 10.11.87.61:50010
> 2010-11-09 05:12:29,815 INFO org.apache.hadoop.hdfs.DFSClient: Exception in 
> createBlockOutputStream java.io.IOException: Bad connect ack with 
> firstBadLink 10.11.87.65:50010
> 2010-11-09 05:12:29,816 INFO org.apache.hadoop.hdfs.DFSClient: Abandoning 
> block blk_3509928762116143133_775950
> 2010-11-09 05:12:29,818 INFO org.apache.hadoop.hdfs.DFSClient: Waiting to 
> find target node: 10.11.87.61:50010
> 2010-11-09 05:13:35,949 INFO org.apache.hadoop.hdfs.DFSClient: Exception in 
> createBlockOutputStream java.io.IOException: Bad connect ack with 
> firstBadLink 10.11.87.65:50010
> 2010-11-09 05:13:35,949 INFO org.apache.hadoop.hdfs.DFSClient: Abandoning 
> block blk_3138002906377068146_775950
> 2010-11-09 05:13:35,950 INFO org.apache.hadoop.hdfs.DFSClient: Waiting to 
> find target node: 10.11.87.61:50010
> 2010-11-09 05:13:51,757 WARN org.apache.hadoop.hdfs.DFSClient: DataStreamer 
> Exception: java.io.IOException: Unable to create new block.
>   at 
> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.nextBlockOutputStream(DFSClient.java:2812)
>   at 
> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.access$2000(DFSClient.java:2076)
>   at 
> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$DataStreamer.run(DFSClient.java:2262)
>
> 2010-11-09 05:13:51,757 WARN org.apache.hadoop.hdfs.DFSClient: Error Recovery 
> for block blk_3138002906377068146_775950 bad datanode[2] nodes == null
> 2010-11-09 05:13:51,758 WARN org.apache.hadoop.hdfs.DFSClient: Could not get 
> block locations. Source file 
> "/user/hadoop/requests_logs/merged/2010-11-08/_temporary/_attempt_201011081008_0002_r_08_0/part-r-8"
>  - Aborting...
> 2010-11-09 05:13:51,760 WARN org.apache.hadoop.mapred.TaskTracker: Error 
> running child
> java.io.IOException: Bad connect ack with firstBadLink 10.11.87.65:50010
>   at 
> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.createBlockOutputStream(DFSClient.java:2870)
>   at 
> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.nextBlockOutputStream(DFSClient.java:2793)
>   at 
> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.access$2000(DFSClient.java:2076)
>   at 
> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$DataStreamer.run(DFSClient.java:2262)
> 2010-11-09 05:13:54,888 INFO org.apache.hadoop.mapred.TaskRunner: Runnning 
> cleanup for the task
>
>
> Thanks
>
> Oleg.
>
> --
>
>


java.io.IOException: Bad connect ack with firstBadLink

2010-11-09 Thread Oleg Ruchovets
Hi , running hadoop map/reduce got such exception?

  1) Why does it happen?

  2) Job didn't failed and continue it's execution? Does this
exception cause losing data or map/reduce

uses recovery mechanism?



2010-11-09 05:10:08,735 INFO org.apache.hadoop.hdfs.DFSClient:
Exception in createBlockOutputStream java.io.IOException: Bad connect
ack with firstBadLink 10.11.87.65:50010
2010-11-09 05:10:08,735 INFO org.apache.hadoop.hdfs.DFSClient:
Abandoning block blk_-9208177033562590356_775948
2010-11-09 05:10:08,739 INFO org.apache.hadoop.hdfs.DFSClient: Waiting
to find target node: 10.11.87.61:50010
2010-11-09 05:11:23,743 INFO org.apache.hadoop.hdfs.DFSClient:
Exception in createBlockOutputStream java.net.SocketTimeoutException:
69000 millis timeout while waiting for channel to be ready for read.
ch : java.nio.channels.SocketChannel[connected
local=/10.11.87.61:55309 remote=/10.11.87.61:50010]
2010-11-09 05:11:23,743 INFO org.apache.hadoop.hdfs.DFSClient:
Abandoning block blk_-10251707095594311_775950
2010-11-09 05:11:23,744 INFO org.apache.hadoop.hdfs.DFSClient: Waiting
to find target node: 10.11.87.61:50010
2010-11-09 05:12:29,815 INFO org.apache.hadoop.hdfs.DFSClient:
Exception in createBlockOutputStream java.io.IOException: Bad connect
ack with firstBadLink 10.11.87.65:50010
2010-11-09 05:12:29,816 INFO org.apache.hadoop.hdfs.DFSClient:
Abandoning block blk_3509928762116143133_775950
2010-11-09 05:12:29,818 INFO org.apache.hadoop.hdfs.DFSClient: Waiting
to find target node: 10.11.87.61:50010
2010-11-09 05:13:35,949 INFO org.apache.hadoop.hdfs.DFSClient:
Exception in createBlockOutputStream java.io.IOException: Bad connect
ack with firstBadLink 10.11.87.65:50010
2010-11-09 05:13:35,949 INFO org.apache.hadoop.hdfs.DFSClient:
Abandoning block blk_3138002906377068146_775950
2010-11-09 05:13:35,950 INFO org.apache.hadoop.hdfs.DFSClient: Waiting
to find target node: 10.11.87.61:50010
2010-11-09 05:13:51,757 WARN org.apache.hadoop.hdfs.DFSClient:
DataStreamer Exception: java.io.IOException: Unable to create new
block.
at 
org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.nextBlockOutputStream(DFSClient.java:2812)
at 
org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.access$2000(DFSClient.java:2076)
at 
org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$DataStreamer.run(DFSClient.java:2262)

2010-11-09 05:13:51,757 WARN org.apache.hadoop.hdfs.DFSClient: Error
Recovery for block blk_3138002906377068146_775950 bad datanode[2]
nodes == null
2010-11-09 05:13:51,758 WARN org.apache.hadoop.hdfs.DFSClient: Could
not get block locations. Source file
"/user/hadoop/requests_logs/merged/2010-11-08/_temporary/_attempt_201011081008_0002_r_08_0/part-r-8"
- Aborting...
2010-11-09 05:13:51,760 WARN org.apache.hadoop.mapred.TaskTracker:
Error running child
java.io.IOException: Bad connect ack with firstBadLink 10.11.87.65:50010
at 
org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.createBlockOutputStream(DFSClient.java:2870)
at 
org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.nextBlockOutputStream(DFSClient.java:2793)
at 
org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.access$2000(DFSClient.java:2076)
at 
org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$DataStreamer.run(DFSClient.java:2262)
2010-11-09 05:13:54,888 INFO org.apache.hadoop.mapred.TaskRunner:
Runnning cleanup for the task


Thanks

Oleg.

--


RE: hdfs.DFSClient: Exception in createBlockOutputStream java.io.IOException: Bad connect ack with firstBadLink

2009-09-01 Thread umer arshad

I have resolved the issue:
What i did: 

1) '/etc/init.d/iptables stop' -->stopped firewall
2) SELINUX=disabled in '/etc/selinux/config' file.-->disabled selinux
I worked for me after these two changes.
thanks,
--umer

> From: m_umer_ars...@hotmail.com
> To: common-user@hadoop.apache.org
> Subject: hdfs.DFSClient: Exception in createBlockOutputStream 
> java.io.IOException: Bad connect ack with firstBadLink
> Date: Mon, 31 Aug 2009 23:35:31 +
> 
> 
> Hi,
> 
> I have set-up an 8-node private hadoop cluster having IP-addresses as follows:
> 
> 192.168.1.10 (master)
> 192.168.1.11 
> 192.168.1.12  
> 192.168.1.13
> 192.168.1.14
> 192.168.1.15
> 192.168.1.16
> 192.168.1.17
> 
> Address 192.168.1.10, master, is running NN+JT and all other nodes are slaves 
> i.e. running DN+TT. I am trying to put data on HDFS using command: hadoop dfs 
> -put 8GB_input 8GB_input
> 
> I have noticed that some blocks are not replicated/placed on nodes with IP 
> addresses 192.168.1.11, 192.168.1.15, and 192.168.1.16 and i get the 
> following error messages:
> 
> $ hadoop dfs -put 8GB_input 8GB_input
> 09/08/31 18:25:45 INFO hdfs.DFSClient: Exception in createBlockOutputStream 
> java.io.IOException: Bad connect ack with firstBadLink 192.168.1.11:50010
> 09/08/31 18:25:45 INFO hdfs.DFSClient: Abandoning block 
> blk_-8575812198227241296_1001
> 09/08/31 18:25:51 INFO hdfs.DFSClient: Exception in createBlockOutputStream 
> java.io.IOException: Bad connect ack with firstBadLink 192.168.1.16:50010
> 09/08/31 18:25:51 INFO hdfs.DFSClient: Abandoning block 
> blk_-2932256218448902464_1001
> 09/08/31 18:25:57 INFO hdfs.DFSClient: Exception in createBlockOutputStream 
> java.io.IOException: Bad connect ack with firstBadLink 192.168.1.11:50010
> 09/08/31 18:25:57 INFO hdfs.DFSClient: Abandoning block 
> blk_-1014449966480421244_1001
> 09/08/31 18:26:03 INFO hdfs.DFSClient: Exception in createBlockOutputStream 
> java.io.IOException: Bad connect ack with firstBadLink 192.168.1.16:50010
> 09/08/31 18:26:03 INFO hdfs.DFSClient: Abandoning block 
> blk_7193173823538206978_1001
> 09/08/31 18:26:09 WARN hdfs.DFSClient: DataStreamer Exception: 
> java.io.IOException: Unable to create new block.
> at 
> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.nextBlockOutputStream(DFSClient.java:2731)
> at 
> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.access$2000(DFSClient.java:1996)
> at 
> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$DataStreamer.run(DFSClient.java:2182)
> 
> 09/08/31 18:26:09 WARN hdfs.DFSClient: Error Recovery for block 
> blk_7193173823538206978_1001 bad datanode[2] nodes == null
> 09/08/31 18:26:09 WARN hdfs.DFSClient: Could not get block locations. Source 
> file "/user/umer/8GB_input" - Aborting...
> put: Bad connect ack with firstBadLink 192.168.1.16:50010
> -
> 
> Sometimes the input file is replicated successfully (excluding these three 
> nodes) and sometimes the copy process i.e. 'hdfs -put input input' terminates.
> 
> NOTE: My replication factor = 3. 
> 
> I am able to see that all machines are up-and-running using Web-UI 
> http://192.168.1.10:50070.
> 
> I will be greatful for any suggestion/comment in this regard. 
> 
> thanks,
> --umer
> 
> 
> 
> 
> _
> Share your memories online with anyone you want.
> http://www.microsoft.com/middleeast/windows/windowslive/products/photos-share.aspx?tab=1

_
More than messages–check out the rest of the Windows Live™.
http://www.microsoft.com/windows/windowslive/

hdfs.DFSClient: Exception in createBlockOutputStream java.io.IOException: Bad connect ack with firstBadLink

2009-08-31 Thread umer arshad

Hi,

I have set-up an 8-node private hadoop cluster having IP-addresses as follows:

192.168.1.10 (master)
192.168.1.11 
192.168.1.12  
192.168.1.13
192.168.1.14
192.168.1.15
192.168.1.16
192.168.1.17

Address 192.168.1.10, master, is running NN+JT and all other nodes are slaves 
i.e. running DN+TT. I am trying to put data on HDFS using command: hadoop dfs 
-put 8GB_input 8GB_input

I have noticed that some blocks are not replicated/placed on nodes with IP 
addresses 192.168.1.11, 192.168.1.15, and 192.168.1.16 and i get the following 
error messages:

$ hadoop dfs -put 8GB_input 8GB_input
09/08/31 18:25:45 INFO hdfs.DFSClient: Exception in createBlockOutputStream 
java.io.IOException: Bad connect ack with firstBadLink 192.168.1.11:50010
09/08/31 18:25:45 INFO hdfs.DFSClient: Abandoning block 
blk_-8575812198227241296_1001
09/08/31 18:25:51 INFO hdfs.DFSClient: Exception in createBlockOutputStream 
java.io.IOException: Bad connect ack with firstBadLink 192.168.1.16:50010
09/08/31 18:25:51 INFO hdfs.DFSClient: Abandoning block 
blk_-2932256218448902464_1001
09/08/31 18:25:57 INFO hdfs.DFSClient: Exception in createBlockOutputStream 
java.io.IOException: Bad connect ack with firstBadLink 192.168.1.11:50010
09/08/31 18:25:57 INFO hdfs.DFSClient: Abandoning block 
blk_-1014449966480421244_1001
09/08/31 18:26:03 INFO hdfs.DFSClient: Exception in createBlockOutputStream 
java.io.IOException: Bad connect ack with firstBadLink 192.168.1.16:50010
09/08/31 18:26:03 INFO hdfs.DFSClient: Abandoning block 
blk_7193173823538206978_1001
09/08/31 18:26:09 WARN hdfs.DFSClient: DataStreamer Exception: 
java.io.IOException: Unable to create new block.
at 
org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.nextBlockOutputStream(DFSClient.java:2731)
at 
org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.access$2000(DFSClient.java:1996)
at 
org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$DataStreamer.run(DFSClient.java:2182)

09/08/31 18:26:09 WARN hdfs.DFSClient: Error Recovery for block 
blk_7193173823538206978_1001 bad datanode[2] nodes == null
09/08/31 18:26:09 WARN hdfs.DFSClient: Could not get block locations. Source 
file "/user/umer/8GB_input" - Aborting...
put: Bad connect ack with firstBadLink 192.168.1.16:50010
-

Sometimes the input file is replicated successfully (excluding these three 
nodes) and sometimes the copy process i.e. 'hdfs -put input input' terminates.

NOTE: My replication factor = 3. 

I am able to see that all machines are up-and-running using Web-UI 
http://192.168.1.10:50070.

I will be greatful for any suggestion/comment in this regard. 

thanks,
--umer




_
Share your memories online with anyone you want.
http://www.microsoft.com/middleeast/windows/windowslive/products/photos-share.aspx?tab=1