I am getting this message if I try to process some data in HDFS but the 
necessary daemons are not started prior to my HDFS request (I am using Hadoop 
V0.20.2 so I am using /start-all.sh script). You're writting it works after 
some time so I guess the problem is in the communication between HDFS daemons 
in your Hadoop cluster.Is your HDFS formatted for all machines or have you add 
some nodes already into the formatted cluster?

_________________Regards,Lukas


From: uni...@foxmail.com
To: hdfs-user@hadoop.apache.org
Subject: When I connect to Hadoop it always 'retrying connect', something wrong 
with my configurations?
Date: Wed, 25 Apr 2012 16:04:58 +0800

I am running a Hadoop cluster of about 40 machines, and I got some problem with 
the HDFS.

When I try to connect to the HDFS, say using the 'hadoop fs -ls /' command, 
sometimes it needs to retry connect to HDFS:12/04/25 11:22:01 INFO ipc.Client: 
Retrying connect to server: master/10.10.10.51:8020. Already tried 0 time(s).
And then it can connect to HDFS and return the result. But the process costs 
time and I am looking for some way to solve it.

Is there something wrong with my Hadoop configurations? Or something wrong with 
the networking?
                                          

Reply via email to