here's the reference I wanted to send you:
http://www.ibm.com/developerworks/eserver/library/es-ssh/index.html
go here
http://www.ibm.com/developerworks/eserver/library/es-ssh/index.html#figure12

hope it solves it

On 04/01/2008, Dhaya007 <[EMAIL PROTECTED]> wrote:
>
>
> Thanks For your good solution
> I ill check and update the same
>
>
>
>
> Khalil Honsali wrote:
> >
> > I also note that for non-root passwordless ssh,  you must chmod
> > authorized_keys file to 655,
> >
> > On 03/01/2008, Miles Osborne <[EMAIL PROTECTED]> wrote:
> >>
> >> You need to make sure that each slave node has a copy of the authorised
> >> keys
> >> you generated on the master node.
> >>
> >> Miles
> >>
> >> On 03/01/2008, Dhaya007 <[EMAIL PROTECTED]> wrote:
> >> >
> >> >
> >> > Thanks Arun,
> >> >
> >> > I am able to riun the datanode in slave (As per the solution given by
> >> You
> >> > (listinig port ))
> >> >
> >> > But still it asks the pasword while starting the dfs ans mapreduce
> >> >
> >> > First i generated rsa as password less as follws
> >> >
> >> > ssh-keygen -t rsa -P ""
> >> > cat $HOME/.ssh/id_rsa.pub >> $HOME/.ssh/authorized_keys
> >> > ssh master
> >> > ssh slave
> >> > I started the dfs in master as follows
> >> > /bin/start-dfs.sh
> >> > it asks the passowrd
> >> > Please help me to resolve the same (I dont know i am doing right in
> the
> >> > case
> >> > of ssh)
> >> >
> >> >
> >> >
> >> > Dhaya007 wrote:
> >> > >
> >> > >
> >> > >
> >> > > Arun C Murthy wrote:
> >> > >>
> >> > >> What version of Hadoop are you running?
> >> > >> Dhaya007:hadoop-0.15.1
> >> > >>
> >> > >> http://wiki.apache.org/lucene-hadoop/Help
> >> > >>
> >> > >> Dhaya007 wrote:
> >> > >>  > ..datanode-slave.log
> >> > >>> 2007-12-19 19:30:55,579 WARN org.apache.hadoop.dfs.DataNode:
> >> Invalid
> >> > >>> directory in dfs.data.dir: directory is not writable:
> >> > >>> /tmp/hadoop-hdpusr/dfs/data
> >> > >>> 2007-12-19 19:30:55,579 ERROR org.apache.hadoop.dfs.DataNode: All
> >> > >>> directories in dfs.data.dir are invalid.
> >> > >>
> >> > >> Did you check that directory?
> >> > >> Daya007:Yes, i have checked the folder in which there is no file
> >> saved.
> >> > >>
> >> > >> DataNode is complaining that it doesn't have any 'valid'
> directories
> >> to
> >> > >> store data in.
> >> > >>
> >> > >>> Tasktracker_slav.log
> >> > >>> 2008-01-02 15:10:34,419 ERROR
> org.apache.hadoop.mapred.TaskTracker:
> >> > Can
> >> > >>> not
> >> > >>> start task tracker because java.net.UnknownHostException: unknown
> >> > host:
> >> > >>> localhost
> >> > >>>     at org.apache.hadoop.ipc.Client$Connection.<init>(Client.java
> >> :136)
> >> > >>>     at org.apache.hadoop.ipc.Client.getConnection(Client.java
> :532)
> >> > >>>     at org.apache.hadoop.ipc.Client.call(Client.java:471)
> >> > >>>     at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:184)
> >> > >>>     at org.apache.hadoop.mapred.$Proxy0.getProtocolVersion
> (Unknown
> >> > Source)
> >> > >>>     at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:269)
> >> > >>>     at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:293)
> >> > >>>     at org.apache.hadoop.ipc.RPC.waitForProxy(RPC.java:246)
> >> > >>>     at
> >> > >>> org.apache.hadoop.mapred.TaskTracker.initialize(TaskTracker.java
> >> :427)
> >> > >>>     at org.apache.hadoop.mapred.TaskTracker.<init>(
> TaskTracker.java
> >> > :717)
> >> > >>>     at org.apache.hadoop.mapred.TaskTracker.main(TaskTracker.java
> >> > :1880)
> >> > >>>
> >> > >>
> >> > >> That probably means that the TaskTracker's hadoop-site.xml says
> that
> >> > >> 'localhost' is the JobTracker which isn't true...
> >> > >>
> >> > >> hadoop-site.xml is as follows
> >> > >> <?xml version="1.0"?>
> >> > >> <?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
> >> > >>
> >> > >> <!-- Put site-specific property overrides in this file. -->
> >> > >>
> >> > >> <configuration>
> >> > >> <property>
> >> > >>   <name>hadoop.tmp.dir</name>
> >> > >>   <value>/home/hdusr/hadoop-${user.name}</value>
> >> > >>   <description>A base for other temporary
> directories.</description>
> >> > >> </property>
> >> > >>
> >> > >> <property>
> >> > >>   <name>fs.default.name</name>
> >> > >>   <value>hdfs://master:54310</value>
> >> > >>   <description>The name of the default file system.  A URI whose
> >> > >>   scheme and authority determine the FileSystem
> implementation.  The
> >> > >>   uri's scheme determines the config property (fs.SCHEME.impl)
> >> naming
> >> > >>   the FileSystem implementation class.  The uri's authority is
> used
> >> to
> >> > >>   determine the host, port, etc. for a filesystem.</description>
> >> > >> </property>
> >> > >>
> >> > >> <property>
> >> > >>   <name>mapred.job.tracker</name>
> >> > >>   <value>master:54311</value>
> >> > >>   <description>The host and port that the MapReduce job tracker
> runs
> >> > >>   at.  If "local", then jobs are run in-process as a single map
> >> > >>   and reduce task.
> >> > >>   </description>
> >> > >> </property>
> >> > >>
> >> > >> <property>
> >> > >>   <name>dfs.replication</name>
> >> > >>   <value>2</value>
> >> > >>   <description>Default block replication.
> >> > >>   The actual number of replications can be specified when the file
> >> is
> >> > >> created.
> >> > >>   The default is used if replication is not specified in create
> >> time.
> >> > >>   </description>
> >> > >> </property>
> >> > >>
> >> > >> <property>
> >> > >>   <name>mapred.map.tasks</name>
> >> > >>   <value>20</value>
> >> > >>   <description>As a rule of thumb, use 10x the number of slaves (
> i.e
> >> .,
> >> > >> number of tasktrackers).
> >> > >>   </description>
> >> > >> </property>
> >> > >>
> >> > >> <property>
> >> > >>   <name>mapred.reduce.tasks</name>
> >> > >>   <value>4</value>
> >> > >>   <description>As a rule of thumb, use 2x the number of slave
> >> > processors
> >> > >> (i.e., number of tasktrackers).
> >> > >>   </description>
> >> > >> </property>
> >> > >> </configuration>
> >> > >>
> >> > >>  > namenode-master.log
> >> > >>  > 2008-01-02 14:44:02,636 INFO org.apache.hadoop.dfs.Storage:
> >> Storage
> >> > >>  > directory /tmp/hadoop-hdpusr/dfs/name does not exist.
> >> > >>  > 2008-01-02 14:44:02,638 INFO org.apache.hadoop.ipc.Server:
> >> Stopping
> >> > >> server
> >> > >>  > on 54310
> >> > >>  > 2008-01-02 14:44:02,653 ERROR org.apache.hadoop.dfs.NameNode:
> >> > >>  > org.apache.hadoop.dfs.InconsistentFSStateException: Directory
> >> > >>  > /tmp/hadoop-hdpusr/dfs/name is in an inconsistent state:
> storage
> >> > >> directory
> >> > >>  > does not exist or is not accessible.
> >> > >>
> >> > >> That means that, /tmp/hadoop-hdpusr/dfs/name doesn't exist or
> isn't
> >> > >> accessible.
> >> > >>
> >> > >> Dhaya007 I have checked the name folder but i wont find any folder
> >> in
> >> > the
> >> > >> specified dir
> >> > >> -*-*-
> >> > >>
> >> > >> Overall, this looks like an acute case of
> wrong-configuration-itis.
> >> > >> Please provid the corect configuration site example for multi node
> >> > >> cluster other than
> >> > >>
> >> >
> >>
> http://www.michael-noll.com/wiki/Running_Hadoop_On_Ubuntu_Linux_%28Single-Node_Cluster%29
> >> > >> because i followed the same
> >> > >>
> >> > >> Have you got the same hadoop-site.xml on all your nodes?
> >> > >> Dhaya007:Yes
> >> > >>
> >> > >> More info here:
> >> > >> http://lucene.apache.org/hadoop/docs/r0.15.1/cluster_setup.html
> >> > >> Dhaya007: I followed the same site you have mentioned but no
> >> solution
> >> > >>
> >> > >> Arun
> >> > >>
> >> > >>
> >> > >>> 2008-01-02 15:10:34,420 INFO org.apache.hadoop.mapred.TaskTracker
> :
> >> > >>> SHUTDOWN_MSG:
> >> > >>> /************************************************************
> >> > >>> SHUTDOWN_MSG: Shutting down TaskTracker at slave/172.16.0.58
> >> > >>> ************************************************************/
> >> > >>>
> >> > >>>
> >> > >>> And all the ports are running
> >> > >>> Some time it asks password and some time it wont while starting
> the
> >> > dfs
> >> > >>>
> >> > >>> Master logs
> >> > >>> 2008-01-02 14:44:02,677 INFO org.apache.hadoop.dfs.NameNode:
> >> > >>> SHUTDOWN_MSG:
> >> > >>> /************************************************************
> >> > >>> SHUTDOWN_MSG: Shutting down NameNode at master/172.16.0.25
> >> > >>> ************************************************************/
> >> > >>>
> >> > >>> Datanode-master.log
> >> > >>> 2008-01-02 16:26:32,380 INFO org.apache.hadoop.ipc.RPC: Server at
> >> > >>> localhost/127.0.0.1:54310 not available yet, Zzzzz...
> >> > >>> 2008-01-02 16:26:33,390 INFO org.apache.hadoop.ipc.Client:
> Retrying
> >> > >>> connect
> >> > >>> to server: localhost/127.0.0.1:54310. Already tried 1 time(s).
> >> > >>> 2008-01-02 16:26:34,400 INFO org.apache.hadoop.ipc.Client:
> Retrying
> >> > >>> connect
> >> > >>> to server: localhost/127.0.0.1:54310. Already tried 2 time(s).
> >> > >>> 2008-01-02 16:26:35,410 INFO org.apache.hadoop.ipc.Client:
> Retrying
> >> > >>> connect
> >> > >>> to server: localhost/127.0.0.1:54310. Already tried 3 time(s).
> >> > >>> 2008-01-02 16:26:36,420 INFO org.apache.hadoop.ipc.Client:
> Retrying
> >> > >>> connect
> >> > >>> to server: localhost/127.0.0.1:54310. Already tried 4 time(s).
> >> > >>> ***********************************************
> >> > >>> Jobtracker_master.log
> >> > >>> 2008-01-02 16:25:41,040 INFO org.apache.hadoop.ipc.Client:
> Retrying
> >> > >>> connect
> >> > >>> to server: localhost/127.0.0.1:54310. Already tried 10 time(s).
> >> > >>> 2008-01-02 16:25:42,050 INFO org.apache.hadoop.mapred.JobTracker:
> >> > >>> problem
> >> > >>> cleaning system directory: /tmp/hadoop-hdpusr/mapred/system
> >> > >>> java.net.ConnectException: Connection refused
> >> > >>>     at java.net.PlainSocketImpl.socketConnect(Native Method)
> >> > >>>     at java.net.PlainSocketImpl.doConnect(PlainSocketImpl.java
> :333)
> >> > >>>     at java.net.PlainSocketImpl.connectToAddress(
> >> PlainSocketImpl.java
> >> > :195)
> >> > >>>     at java.net.PlainSocketImpl.connect(PlainSocketImpl.java:182)
> >> > >>>     at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:366)
> >> > >>>     at java.net.Socket.connect(Socket.java:520)
> >> > >>>     at
> >> > >>> org.apache.hadoop.ipc.Client$Connection.setupIOstreams(
> Client.java
> >> > :152)
> >> > >>>     at org.apache.hadoop.ipc.Client.getConnection(Client.java
> :542)
> >> > >>>     at org.apache.hadoop.ipc.Client.call(Client.java:471)
> >> > >>>     at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:184)
> >> > >>>     at org.apache.hadoop.dfs.$Proxy0.getProtocolVersion(Unknown
> >> > Source)
> >> > >>>     at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:269)
> >> > >>>     at
> >> org.apache.hadoop.dfs.DFSClient.createNamenode(DFSClient.java
> >> > :147)
> >> > >>>     at org.apache.hadoop.dfs.DFSClient.<init>(DFSClient.java:161)
> >> > >>>     at
> >> > >>> org.apache.hadoop.dfs.DistributedFileSystem.initialize(
> >> > DistributedFileSystem.java:65)
> >> > >>>     at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:159)
> >> > >>>     at
> >> org.apache.hadoop.fs.FileSystem.getNamed(FileSystem.java:118)
> >> > >>>     at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:90)
> >> > >>>     at org.apache.hadoop.mapred.JobTracker.<init>(JobTracker.java
> >> :683)
> >> > >>>     at
> >> > >>> org.apache.hadoop.mapred.JobTracker.startTracker(JobTracker.java
> >> :120)
> >> > >>>     at org.apache.hadoop.mapred.JobTracker.main(JobTracker.java
> >> :2052)
> >> > >>> 2008-01-02 16:25:42,931 INFO org.apache.hadoop.ipc.Server: IPC
> >> Server
> >> > >>> handler 5 on 54311, call getFilesystemName() from 127.0.0.1:49283
> :
> >> > >>> error:
> >> > >>>
> >> org.apache.hadoop.mapred.JobTracker$IllegalStateException:FileSystem
> >> > >>> object
> >> > >>> not available yet
> >> > >>>
> >> org.apache.hadoop.mapred.JobTracker$IllegalStateException:FileSystem
> >> > >>> object
> >> > >>> not available yet
> >> > >>>     at
> >> > >>> org.apache.hadoop.mapred.JobTracker.getFilesystemName(
> >> JobTracker.java
> >> > :1475)
> >> > >>>     at sun.reflect.GeneratedMethodAccessor2.invoke(Unknown
> Source)
> >> > >>>     at
> >> > >>> sun.reflect.DelegatingMethodAccessorImpl.invoke(
> >> > DelegatingMethodAccessorImpl.java:25)
> >> > >>>     at java.lang.reflect.Method.invoke(Method.java:585)
> >> > >>>     at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:379)
> >> > >>>     at org.apache.hadoop.ipc.Server$Handler.run(Server.java:596)
> >> > >>> 2008-01-02 16:25:47,942 INFO org.apache.hadoop.ipc.Server: IPC
> >> Server
> >> > >>> handler 6 on 54311, call getFilesystemName() from 127.0.0.1:49293
> :
> >> > >>> error:
> >> > >>>
> >> org.apache.hadoop.mapred.JobTracker$IllegalStateException:FileSystem
> >> > >>> object
> >> > >>> not available yet
> >> > >>>
> >> org.apache.hadoop.mapred.JobTracker$IllegalStateException:FileSystem
> >> > >>> object
> >> > >>> not available yet
> >> > >>>     at
> >> > >>> org.apache.hadoop.mapred.JobTracker.getFilesystemName(
> >> JobTracker.java
> >> > :1475)
> >> > >>>     at sun.reflect.GeneratedMethodAccessor2.invoke(Unknown
> Source)
> >> > >>>     at
> >> > >>> sun.reflect.DelegatingMethodAccessorImpl.invoke(
> >> > DelegatingMethodAccessorImpl.java:25)
> >> > >>>     at java.lang.reflect.Method.invoke(Method.java:585)
> >> > >>>     at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:379)
> >> > >>>     at org.apache.hadoop.ipc.Server$Handler.run(Server.java:596)
> >> > >>> 2008-01-02 16:25:52,061 INFO org.apache.hadoop.ipc.Client:
> Retrying
> >> > >>> connect
> >> > >>> to server: localhost/127.0.0.1:54310. Already tried 1 time(s).
> >> > >>> 2008-01-02 16:25:52,951 INFO org.apache.hadoop.ipc.Server: IPC
> >> Server
> >> > >>> handler 7 on 54311, call getFilesystemName() from 127.0.0.1:49304
> :
> >> > >>> error:
> >> > >>>
> >> org.apache.hadoop.mapred.JobTracker$IllegalStateException:FileSystem
> >> > >>> object
> >> > >>> not available yet
> >> > >>>
> >> org.apache.hadoop.mapred.JobTracker$IllegalStateException:FileSystem
> >> > >>> object
> >> > >>> not available yet
> >> > >>>     at
> >> > >>> org.apache.hadoop.mapred.JobTracker.getFilesystemName(
> >> JobTracker.java
> >> > :1475)
> >> > >>>     at sun.reflect.GeneratedMethodAccessor2.invoke(Unknown
> Source)
> >> > >>>     at
> >> > >>> sun.reflect.DelegatingMethodAccessorImpl.invoke(
> >> > DelegatingMethodAccessorImpl.java:25)
> >> > >>>     at java.lang.reflect.Method.invoke(Method.java:585)
> >> > >>>     at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:379)
> >> > >>>     at org.apache.hadoop.ipc.Server$Handler.run(Server.java:596)
> >> > >>> 2008-01-02 16:25:53,070 INFO org.apache.hadoop.ipc.Client:
> Retrying
> >> > >>> connect
> >> > >>> to server: localhost/127.0.0.1:54310. Already tried 2 time(s).
> >> > >>> 2008-01-02 16:25:54,080 INFO org.apache.hadoop.ipc.Client:
> Retrying
> >> > >>> connect
> >> > >>> to server: localhost/127.0.0.1:54310. Already tried 3 time(s).
> >> > >>> 2008-01-02 16:25:55,090 INFO org.apache.hadoop.ipc.Client:
> Retrying
> >> > >>> connect
> >> > >>> to server: localhost/127.0.0.1:54310. Already tried 4 time(s).
> >> > >>> 2008-01-02 16:25:56,100 INFO org.apache.hadoop.ipc.Client:
> Retrying
> >> > >>> connect
> >> > >>> to server: localhost/127.0.0.1:54310. Already tried 5 time(s).
> >> > >>> 2008-01-02 16:25:56,281 INFO org.apache.hadoop.mapred.JobTracker:
> >> > >>> SHUTDOWN_MSG:
> >> > >>> /************************************************************
> >> > >>> SHUTDOWN_MSG: Shutting down JobTracker at master/172.16.0.25
> >> > >>> ************************************************************/
> >> > >>>
> >> > >>> Tasktracker_master.log
> >> > >>> 2008-01-02 16:26:14,080 INFO org.apache.hadoop.ipc.Client:
> Retrying
> >> > >>> connect
> >> > >>> to server: localhost/127.0.0.1:54311. Already tried 2 time(s).
> >> > >>> 2008-01-02 16:28:34,510 INFO org.apache.hadoop.mapred.TaskTracker
> :
> >> > >>> STARTUP_MSG:
> >> > >>> /************************************************************
> >> > >>> STARTUP_MSG: Starting TaskTracker
> >> > >>> STARTUP_MSG:   host = master/172.16.0.25
> >> > >>> STARTUP_MSG:   args = []
> >> > >>> ************************************************************/
> >> > >>> 2008-01-02 16:28:34,739 INFO org.mortbay.util.Credential:
> Checking
> >> > >>> Resource
> >> > >>> aliases
> >> > >>> 2008-01-02 16:28:34,827 INFO org.mortbay.http.HttpServer: Version
> >> > >>> Jetty/5.1.4
> >> > >>> 2008-01-02 16:28:35,281 INFO org.mortbay.util.Container: Started
> >> > >>> [EMAIL PROTECTED]
> >> > >>> 2008-01-02 16:28:35,332 INFO org.mortbay.util.Container: Started
> >> > >>> WebApplicationContext[/,/]
> >> > >>> 2008-01-02 16:28:35,332 INFO org.mortbay.util.Container: Started
> >> > >>> HttpContext[/logs,/logs]
> >> > >>> 2008-01-02 16:28:35,332 INFO org.mortbay.util.Container: Started
> >> > >>> HttpContext[/static,/static]
> >> > >>> 2008-01-02 16:28:35,336 INFO org.mortbay.http.SocketListener:
> >> Started
> >> > >>> SocketListener on 0.0.0.0:50060
> >> > >>> 2008-01-02 16:28:35,336 INFO org.mortbay.util.Container: Started
> >> > >>> [EMAIL PROTECTED]
> >> > >>> 2008-01-02 16:28:35,383 INFO
> >> org.apache.hadoop.metrics.jvm.JvmMetrics:
> >> > >>> Initializing JVM Metrics with processName=TaskTracker, sessionId=
> >> > >>> 2008-01-02 16:28:35,402 INFO org.apache.hadoop.mapred.TaskTracker
> :
> >> > >>> TaskTracker up at: /127.0.0.1:49599
> >> > >>> 2008-01-02 16:28:35,402 INFO org.apache.hadoop.mapred.TaskTracker
> :
> >> > >>> Starting
> >> > >>> tracker tracker_master:/127.0.0.1:49599
> >> > >>> 2008-01-02 16:28:35,406 INFO org.apache.hadoop.ipc.Server: IPC
> >> Server
> >> > >>> listener on 49599: starting
> >> > >>> 2008-01-02 16:28:35,406 INFO org.apache.hadoop.ipc.Server: IPC
> >> Server
> >> > >>> handler 0 on 49599: starting
> >> > >>> 2008-01-02 16:28:35,406 INFO org.apache.hadoop.ipc.Server: IPC
> >> Server
> >> > >>> handler 1 on 49599: starting
> >> > >>> 2008-01-02 16:28:35,490 INFO org.apache.hadoop.mapred.TaskTracker
> :
> >> > >>> Starting
> >> > >>> thread: Map-events fetcher for all reduce tasks on
> >> > >>> tracker_master:/127.0.0.1:49599
> >> > >>> 2008-01-02 16:28:35,500 INFO org.apache.hadoop.mapred.TaskTracker
> :
> >> > Lost
> >> > >>> connection to JobTracker
> [localhost/127.0.0.1:54311].  Retrying...
> >> > >>> org.apache.hadoop.ipc.RemoteException:
> >> > >>>
> >> org.apache.hadoop.mapred.JobTracker$IllegalStateException:FileSystem
> >> > >>> object
> >> > >>> not available yet
> >> > >>>     at
> >> > >>> org.apache.hadoop.mapred.JobTracker.getFilesystemName(
> >> JobTracker.java
> >> > :1475)
> >> > >>>     at sun.reflect.NativeMethodAccessorImpl.invoke0(Native
> Method)
> >> > >>>     at
> >> > >>> sun.reflect.NativeMethodAccessorImpl.invoke(
> >> > NativeMethodAccessorImpl.java:39)
> >> > >>>     at
> >> > >>> sun.reflect.DelegatingMethodAccessorImpl.invoke(
> >> > DelegatingMethodAccessorImpl.java:25)
> >> > >>>     at java.lang.reflect.Method.invoke(Method.java:585)
> >> > >>>     at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:379)
> >> > >>>     at org.apache.hadoop.ipc.Server$Handler.run(Server.java:596)
> >> > >>>
> >> > >>>     at org.apache.hadoop.ipc.Client.call(Client.java:482)
> >> > >>>     at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:184)
> >> > >>>     at org.apache.hadoop.mapred.$Proxy0.getFilesystemName(Unknown
> >> > Source)
> >> > >>>     at
> >> > >>> org.apache.hadoop.mapred.TaskTracker.offerService(
> TaskTracker.java
> >> > :773)
> >> > >>>     at org.apache.hadoop.mapred.TaskTracker.run(TaskTracker.java
> >> :1179)
> >> > >>>     at org.apache.hadoop.mapred.TaskTracker.main(TaskTracker.java
> >> > :1880)
> >> > >>> *******************************************
> >> > >>>
> >> > >>> Please help me to resolve the same.
> >> > >>>
> >> > >>>
> >> > >>> Khalil Honsali wrote:
> >> > >>>
> >> > >>>>Hi,
> >> > >>>>
> >> > >>>>I think you need to post more information, for example an excerpt
> >> of
> >> > the
> >> > >>>>failing datanode log. Also, please clarify the issue of
> >> connectivity:
> >> > >>>>- are you able to ssh passwordless (from master to slave, slave
> to
> >> > master,
> >> > >>>>slave to slave, master to master), you shouldn't be entering
> >> passwrd
> >> > >>>>everytime...
> >> > >>>>- are you able to telnet (not necessary but preferred)
> >> > >>>>- have you verified the ports as RUNNING on using netstat
> command?
> >> > >>>>
> >> > >>>>besides, the tasktracker starts ok but not the datanode?
> >> > >>>>
> >> > >>>>K. Honsali
> >> > >>>>
> >> > >>>>On 02/01/2008, Dhaya007 <[EMAIL PROTECTED]> wrote:
> >> > >>>>
> >> > >>>>>
> >> > >>>>>I am new to hadoop if any think wrong please correct me ....
> >> > >>>>>I Have configured a single/multi node cluster using following
> link
> >> > >>>>>
> >> > >>>>>
> >> >
> >>
> http://www.michael-noll.com/wiki/Running_Hadoop_On_Ubuntu_Linux_%28Single-Node_Cluster%29
> >> > >>>>>.
> >> > >>>>>I have followed the link but i am not able to start the haoop in
> >> > multi
> >> > >>>>>node
> >> > >>>>>environment
> >> > >>>>>The problems i am facing are as Follows:
> >> > >>>>>1.I have configured master and slave nodes with ssh less pharase
> >> if
> >> > try
> >> > >>>>>to
> >> > >>>>>run the start-dfs.sh it prompt the password for master:slave
> >> > machines.(I
> >> > >>>>>have copied the .ssh/id_rsa.pub key of master in to slaves
> >> > autherized_key
> >> > >>>>>file)
> >> > >>>>>
> >> > >>>>>2.After giving password datanode,namenode,jobtracker,tasktraker
> >> > started
> >> > >>>>>successfully in master but datanode is started in slave.
> >> > >>>>>
> >> > >>>>>
> >> > >>>>>3.Some time step 2 works and some time it says that permission
> >> > denied.
> >> > >>>>>
> >> > >>>>>4.I have checked the log file in the slave for datanode it says
> >> that
> >> > >>>>>incompatible node, then i have formated the slave, master and
> >> start
> >> > the
> >> > >>>>>dfs
> >> > >>>>>by start-dfs.sh still i am getting the error
> >> > >>>>>
> >> > >>>>>
> >> > >>>>>The host entry in etc/hosts are both master/slave
> >> > >>>>>master
> >> > >>>>>slave
> >> > >>>>>conf/masters
> >> > >>>>>master
> >> > >>>>>conf/slaves
> >> > >>>>>master
> >> > >>>>>slave
> >> > >>>>>
> >> > >>>>>The hadoop-site.xml  for both master/slave
> >> > >>>>><?xml version="1.0"?>
> >> > >>>>><?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
> >> > >>>>>
> >> > >>>>><!-- Put site-specific property overrides in this file. -->
> >> > >>>>>
> >> > >>>>><configuration>
> >> > >>>>><property>
> >> > >>>>>  <name>hadoop.tmp.dir</name>
> >> > >>>>>  <value>/home/hdusr/hadoop-${user.name}</value>
> >> > >>>>>  <description>A base for other temporary
> >> directories.</description>
> >> > >>>>></property>
> >> > >>>>>
> >> > >>>>><property>
> >> > >>>>>  <name>fs.default.name</name>
> >> > >>>>>  <value>hdfs://master:54310</value>
> >> > >>>>>  <description>The name of the default file system.  A URI whose
> >> > >>>>>  scheme and authority determine the FileSystem
> >> implementation.  The
> >> > >>>>>  uri's scheme determines the config property (fs.SCHEME.impl)
> >> naming
> >> > >>>>>  the FileSystem implementation class.  The uri's authority is
> >> used
> >> > to
> >> > >>>>>  determine the host, port, etc. for a filesystem.</description>
> >> > >>>>></property>
> >> > >>>>>
> >> > >>>>><property>
> >> > >>>>>  <name>mapred.job.tracker</name>
> >> > >>>>>  <value>master:54311</value>
> >> > >>>>>  <description>The host and port that the MapReduce job tracker
> >> runs
> >> > >>>>>  at.  If "local", then jobs are run in-process as a single map
> >> > >>>>>  and reduce task.
> >> > >>>>>  </description>
> >> > >>>>></property>
> >> > >>>>>
> >> > >>>>><property>
> >> > >>>>>  <name>dfs.replication</name>
> >> > >>>>>  <value>2</value>
> >> > >>>>>  <description>Default block replication.
> >> > >>>>>  The actual number of replications can be specified when the
> file
> >> is
> >> > >>>>>created.
> >> > >>>>>  The default is used if replication is not specified in create
> >> time.
> >> > >>>>>  </description>
> >> > >>>>></property>
> >> > >>>>>
> >> > >>>>><property>
> >> > >>>>>  <name>mapred.map.tasks</name>
> >> > >>>>>  <value>20</value>
> >> > >>>>>  <description>As a rule of thumb, use 10x the number of slaves
> (
> >> i.e
> >> > .,
> >> > >>>>>number of tasktrackers).
> >> > >>>>>  </description>
> >> > >>>>></property>
> >> > >>>>>
> >> > >>>>><property>
> >> > >>>>>  <name>mapred.reduce.tasks</name>
> >> > >>>>>  <value>4</value>
> >> > >>>>>  <description>As a rule of thumb, use 2x the number of slave
> >> > >>>>> processors
> >> > >>>>>(i.e., number of tasktrackers).
> >> > >>>>>  </description>
> >> > >>>>></property>
> >> > >>>>></configuration>
> >> > >>>>>
> >> > >>>>>Please help me to reslove the same. Or else provide any other
> >> > tutorial
> >> > >>>>>for
> >> > >>>>>multi node cluster setup.I am egarly waiting for the tutorials.
> >> > >>>>>
> >> > >>>>>
> >> > >>>>>Thanks
> >> > >>>>>
> >> > >>>>>--
> >> > >>>>>View this message in context:
> >> > >>>>>
> >> >
> >>
> http://www.nabble.com/Not-able-to-start-Data-Node-tp14573889p14573889.html
> >> > >>>>>Sent from the Hadoop Users mailing list archive at Nabble.com.
> >> > >>>>>
> >> > >>>>>
> >> > >>>>
> >> > >>>>
> >> > >>>
> >> > >>
> >> > >>
> >> > >>
> >> > >
> >> > >
> >> >
> >> > --
> >> > View this message in context:
> >> >
> >>
> http://www.nabble.com/Not-able-to-start-Data-Node-tp14573889p14594256.html
> >> > Sent from the Hadoop Users mailing list archive at Nabble.com.
> >> >
> >> >
> >>
> >
> >
> >
> > --
> > ---------------------------------------------------------
> > شهر مبارك كريم
> > كل عام و أنتم بخير
> > ---------------------------------------------------------
> > Honsali Khalil − 本査理 カリル
> > Academic>Japan>NIT>Grad. Sc. Eng.>Dept. CS>Matsuo&Tsumura Lab.
> > http://www.matlab.nitech.ac.jp/~k-hon/
> > +81 (zero-)eight-zero 5134 8119
> > [EMAIL PROTECTED] (instant reply mail)
> >
> >
>
> --
> View this message in context:
> http://www.nabble.com/Not-able-to-start-Data-Node-tp14573889p14613800.html
> Sent from the Hadoop Users mailing list archive at Nabble.com.
>
>


-- 
---------------------------------------------------------
شهر مبارك كريم
كل عام و أنتم بخير
---------------------------------------------------------
Honsali Khalil − 本査理 カリル
Academic>Japan>NIT>Grad. Sc. Eng.>Dept. CS>Matsuo&Tsumura Lab.
http://www.matlab.nitech.ac.jp/~k-hon/
+81 (zero-)eight-zero 5134 8119
[EMAIL PROTECTED] (instant reply mail)

Reply via email to