ok, so all nodes are configured the same except for master/slave differences. They are all running hdfs all daemons seem to be running when I do a start-all.sh from the master. However the master Map/Reduce Administration page shows only two live nodes. The HDFS page shows 3.

Looking at the log files on the new slave node I see no outright errors but see this in the tasktracker log file. All machines have 8G memory. I think the important part below is TaskTracker's totalMemoryAllottedForTasks is -1. I've searched for others with this problem but haven't found something for my case, which is just trying to startup. No tasks have been run.

2012-05-24 11:20:46,786 INFO org.apache.hadoop.mapred.TaskTracker: Starting tracker tracker_occam3:localhost/127.0.0.1:45700 2012-05-24 11:20:46,792 INFO org.apache.hadoop.mapred.TaskTracker: Starting thread: Map-events fetcher for all reduce tasks on tracker_occam3:localhost/127.0.0.1:45700 2012-05-24 11:20:46,792 INFO org.apache.hadoop.mapred.TaskTracker: Using ResourceCalculatorPlugin : org.apache.hadoop.util.LinuxResourceCalculatorPlugin@5abd09e8 2012-05-24 11:20:46,795 WARN org.apache.hadoop.mapred.TaskTracker: TaskTracker's totalMemoryAllottedForTasks is -1. TaskMemoryManager is disabled. 2012-05-24 11:20:46,795 INFO org.apache.hadoop.mapred.IndexCache: IndexCache created with max memory = 10485760 2012-05-24 11:20:46,800 INFO org.apache.hadoop.mapred.TaskTracker: Shutting down: Map-events fetcher for all reduce tasks on tracker_occam3:localhost/127.0.0.1:45700 2012-05-24 11:20:46,800 INFO org.apache.hadoop.filecache.TrackerDistributedCacheManager: Cleanup...
java.lang.InterruptedException: sleep interrupted
    at java.lang.Thread.sleep(Native Method)
at org.apache.hadoop.filecache.TrackerDistributedCacheManager$CleanupThread.run(TrackerDistributedCacheManager.java:926) 2012-05-24 11:20:46,900 INFO org.apache.hadoop.ipc.Server: Stopping server on 45700 2012-05-24 11:20:46,901 INFO org.apache.hadoop.ipc.Server: IPC Server handler 3 on 45700: exiting 2012-05-24 11:20:46,901 INFO org.apache.hadoop.ipc.Server: IPC Server handler 1 on 45700: exiting 2012-05-24 11:20:46,902 INFO org.apache.hadoop.ipc.Server: IPC Server handler 2 on 45700: exiting 2012-05-24 11:20:46,902 INFO org.apache.hadoop.ipc.Server: Stopping IPC Server listener on 45700 2012-05-24 11:20:46,901 INFO org.apache.hadoop.ipc.Server: IPC Server handler 0 on 45700: exiting 2012-05-24 11:20:46,904 INFO org.apache.hadoop.ipc.metrics.RpcInstrumentation: shut down 2012-05-24 11:20:46,904 INFO org.apache.hadoop.mapred.TaskTracker: Shutting down StatusHttpServer 2012-05-24 11:20:46,904 INFO org.apache.hadoop.ipc.Server: IPC Server handler 7 on 45700: exiting 2012-05-24 11:20:46,903 INFO org.apache.hadoop.ipc.Server: IPC Server handler 6 on 45700: exiting 2012-05-24 11:20:46,903 INFO org.apache.hadoop.ipc.Server: IPC Server handler 4 on 45700: exiting 2012-05-24 11:20:46,904 INFO org.apache.hadoop.ipc.Server: IPC Server handler 5 on 45700: exiting 2012-05-24 11:20:46,904 INFO org.apache.hadoop.ipc.Server: Stopping IPC Server Responder 2012-05-24 11:20:46,909 INFO org.mortbay.log: Stopped SelectChannelConnector@0.0.0.0:50060



On 5/23/12 3:55 PM, James Warren wrote:
Hi Pat -

The setting for hadoop.tmp.dir is used both locally and on HDFS and
therefore should be consistent across your cluster.

http://stackoverflow.com/questions/2354525/what-should-be-hadoop-tmp-dir

cheers,
-James

On Wed, May 23, 2012 at 3:44 PM, Pat Ferrel<p...@occamsmachete.com>  wrote:

I have a two machine cluster and am adding a new machine. The new node has
a different location for hadoop.tmp.dir than the other two nodes and
refuses to start the datanode when started in the cluster. When I change
the location pointed to by hadoop.tmp.dir to be the same on all machines it
starts up fine on all machines.

Shouldn't I be able to have the master and slave1 set as:
<property>
<name>hadoop.tmp.dir</name>
<value>/app/hadoop/tmp</value>
<description>A base for other temporary directories.</description>
</property>

And slave2 set as:
<property>
<name>hadoop.tmp.dir</name>
<value>/media/d2/app/hadoop/**tmp</value>
<description>A base for other temporary directories.</description>
</property>

??? Slave2 runs standalone in single node mode just fine. Using 0.20.205.

Reply via email to