Hi Mohammad,
Yes that's correct your rack awareness script takes the IP address of a
node and returns the rack name/id.
You then just have to ensure the script is executable and referenced (using
an absolute path) in the parameter topology.script.file.name in
core-site.xml.
Regards,
Hi Rahul,
The SNN does not act as a backup / standby NameNode in the event of failure.
The sole purpose of the Secondary NameNode (or as it’s otherwise / more
correctly known as the Checkpoint Node) is to perform checkpointing of the
current state of HDFS:
The SNN retrieves the
Hi Nagarjuna,
What's is in your /etc/hosts file? I think the line in logs where it says
DataNodeRegistration(0.0.0.0 [..], should be the hostname or IP of the
datanode (124.123.215.187 since you said it's a pseudo-distributed setup)
and not 0.0.0.0.
By the way are you using the dfs.hosts
Hi Janesh,
I think your SNN may be starting up with the wrong IP, I'm sure the machine
parameter should say 192.168.0.101?
http://namenode:50070/getimage?putimage=1
http://namenode:50070/getimage?putimage=1port=50090machine=0.0.0.0token=
-32:1989419481:0:136084943:1360849122845
Hi,
Could you first try running the example:
$ /usr/bin/hadoop jar /usr/lib/hadoop-0.20-mapreduce/hadoop-examples.jar
grep input output 'dfs[a-z.]+'
Do you receive the same error?
Not sure if it's related to a lack of RAM, but as the stack trace shows
errors with network timeout (I
Hi Manoj,
As you may be aware this means the reduces are unable to fetch intermediate
data from TaskTrackers that ran map tasks - you can try:
* increasing tasktracker.http.threads so there are more threads to handle
fetch requests from reduces.
* decreasing mapreduce.reduce.parallel.copies
Hi Jeba,
There are other considerations too, for example, if a single node holds 1 PB
of data and if it were to die this would cause a significant amount of
traffic as NameNode arranges for new replicas to be created.
Vijay
From: Bertrand Dechoux [mailto:decho...@gmail.com]
Sent: 30