Thanks For your good solution 
I ill check and update the same 



Khalil Honsali wrote:
> 
> I also note that for non-root passwordless ssh,  you must chmod
> authorized_keys file to 655,
> 
> On 03/01/2008, Miles Osborne <[EMAIL PROTECTED]> wrote:
>>
>> You need to make sure that each slave node has a copy of the authorised
>> keys
>> you generated on the master node.
>>
>> Miles
>>
>> On 03/01/2008, Dhaya007 <[EMAIL PROTECTED]> wrote:
>> >
>> >
>> > Thanks Arun,
>> >
>> > I am able to riun the datanode in slave (As per the solution given by
>> You
>> > (listinig port ))
>> >
>> > But still it asks the pasword while starting the dfs ans mapreduce
>> >
>> > First i generated rsa as password less as follws
>> >
>> > ssh-keygen -t rsa -P ""
>> > cat $HOME/.ssh/id_rsa.pub >> $HOME/.ssh/authorized_keys
>> > ssh master
>> > ssh slave
>> > I started the dfs in master as follows
>> > /bin/start-dfs.sh
>> > it asks the passowrd
>> > Please help me to resolve the same (I dont know i am doing right in the
>> > case
>> > of ssh)
>> >
>> >
>> >
>> > Dhaya007 wrote:
>> > >
>> > >
>> > >
>> > > Arun C Murthy wrote:
>> > >>
>> > >> What version of Hadoop are you running?
>> > >> Dhaya007:hadoop-0.15.1
>> > >>
>> > >> http://wiki.apache.org/lucene-hadoop/Help
>> > >>
>> > >> Dhaya007 wrote:
>> > >>  > ..datanode-slave.log
>> > >>> 2007-12-19 19:30:55,579 WARN org.apache.hadoop.dfs.DataNode:
>> Invalid
>> > >>> directory in dfs.data.dir: directory is not writable:
>> > >>> /tmp/hadoop-hdpusr/dfs/data
>> > >>> 2007-12-19 19:30:55,579 ERROR org.apache.hadoop.dfs.DataNode: All
>> > >>> directories in dfs.data.dir are invalid.
>> > >>
>> > >> Did you check that directory?
>> > >> Daya007:Yes, i have checked the folder in which there is no file
>> saved.
>> > >>
>> > >> DataNode is complaining that it doesn't have any 'valid' directories
>> to
>> > >> store data in.
>> > >>
>> > >>> Tasktracker_slav.log
>> > >>> 2008-01-02 15:10:34,419 ERROR org.apache.hadoop.mapred.TaskTracker:
>> > Can
>> > >>> not
>> > >>> start task tracker because java.net.UnknownHostException: unknown
>> > host:
>> > >>> localhost
>> > >>>     at org.apache.hadoop.ipc.Client$Connection.<init>(Client.java
>> :136)
>> > >>>     at org.apache.hadoop.ipc.Client.getConnection(Client.java:532)
>> > >>>     at org.apache.hadoop.ipc.Client.call(Client.java:471)
>> > >>>     at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:184)
>> > >>>     at org.apache.hadoop.mapred.$Proxy0.getProtocolVersion(Unknown
>> > Source)
>> > >>>     at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:269)
>> > >>>     at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:293)
>> > >>>     at org.apache.hadoop.ipc.RPC.waitForProxy(RPC.java:246)
>> > >>>     at
>> > >>> org.apache.hadoop.mapred.TaskTracker.initialize(TaskTracker.java
>> :427)
>> > >>>     at org.apache.hadoop.mapred.TaskTracker.<init>(TaskTracker.java
>> > :717)
>> > >>>     at org.apache.hadoop.mapred.TaskTracker.main(TaskTracker.java
>> > :1880)
>> > >>>
>> > >>
>> > >> That probably means that the TaskTracker's hadoop-site.xml says that
>> > >> 'localhost' is the JobTracker which isn't true...
>> > >>
>> > >> hadoop-site.xml is as follows
>> > >> <?xml version="1.0"?>
>> > >> <?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
>> > >>
>> > >> <!-- Put site-specific property overrides in this file. -->
>> > >>
>> > >> <configuration>
>> > >> <property>
>> > >>   <name>hadoop.tmp.dir</name>
>> > >>   <value>/home/hdusr/hadoop-${user.name}</value>
>> > >>   <description>A base for other temporary directories.</description>
>> > >> </property>
>> > >>
>> > >> <property>
>> > >>   <name>fs.default.name</name>
>> > >>   <value>hdfs://master:54310</value>
>> > >>   <description>The name of the default file system.  A URI whose
>> > >>   scheme and authority determine the FileSystem implementation.  The
>> > >>   uri's scheme determines the config property (fs.SCHEME.impl)
>> naming
>> > >>   the FileSystem implementation class.  The uri's authority is used
>> to
>> > >>   determine the host, port, etc. for a filesystem.</description>
>> > >> </property>
>> > >>
>> > >> <property>
>> > >>   <name>mapred.job.tracker</name>
>> > >>   <value>master:54311</value>
>> > >>   <description>The host and port that the MapReduce job tracker runs
>> > >>   at.  If "local", then jobs are run in-process as a single map
>> > >>   and reduce task.
>> > >>   </description>
>> > >> </property>
>> > >>
>> > >> <property>
>> > >>   <name>dfs.replication</name>
>> > >>   <value>2</value>
>> > >>   <description>Default block replication.
>> > >>   The actual number of replications can be specified when the file
>> is
>> > >> created.
>> > >>   The default is used if replication is not specified in create
>> time.
>> > >>   </description>
>> > >> </property>
>> > >>
>> > >> <property>
>> > >>   <name>mapred.map.tasks</name>
>> > >>   <value>20</value>
>> > >>   <description>As a rule of thumb, use 10x the number of slaves (i.e
>> .,
>> > >> number of tasktrackers).
>> > >>   </description>
>> > >> </property>
>> > >>
>> > >> <property>
>> > >>   <name>mapred.reduce.tasks</name>
>> > >>   <value>4</value>
>> > >>   <description>As a rule of thumb, use 2x the number of slave
>> > processors
>> > >> (i.e., number of tasktrackers).
>> > >>   </description>
>> > >> </property>
>> > >> </configuration>
>> > >>
>> > >>  > namenode-master.log
>> > >>  > 2008-01-02 14:44:02,636 INFO org.apache.hadoop.dfs.Storage:
>> Storage
>> > >>  > directory /tmp/hadoop-hdpusr/dfs/name does not exist.
>> > >>  > 2008-01-02 14:44:02,638 INFO org.apache.hadoop.ipc.Server:
>> Stopping
>> > >> server
>> > >>  > on 54310
>> > >>  > 2008-01-02 14:44:02,653 ERROR org.apache.hadoop.dfs.NameNode:
>> > >>  > org.apache.hadoop.dfs.InconsistentFSStateException: Directory
>> > >>  > /tmp/hadoop-hdpusr/dfs/name is in an inconsistent state: storage
>> > >> directory
>> > >>  > does not exist or is not accessible.
>> > >>
>> > >> That means that, /tmp/hadoop-hdpusr/dfs/name doesn't exist or isn't
>> > >> accessible.
>> > >>
>> > >> Dhaya007 I have checked the name folder but i wont find any folder
>> in
>> > the
>> > >> specified dir
>> > >> -*-*-
>> > >>
>> > >> Overall, this looks like an acute case of wrong-configuration-itis.
>> > >> Please provid the corect configuration site example for multi node
>> > >> cluster other than
>> > >>
>> >
>> http://www.michael-noll.com/wiki/Running_Hadoop_On_Ubuntu_Linux_%28Single-Node_Cluster%29
>> > >> because i followed the same
>> > >>
>> > >> Have you got the same hadoop-site.xml on all your nodes?
>> > >> Dhaya007:Yes
>> > >>
>> > >> More info here:
>> > >> http://lucene.apache.org/hadoop/docs/r0.15.1/cluster_setup.html
>> > >> Dhaya007: I followed the same site you have mentioned but no
>> solution
>> > >>
>> > >> Arun
>> > >>
>> > >>
>> > >>> 2008-01-02 15:10:34,420 INFO org.apache.hadoop.mapred.TaskTracker:
>> > >>> SHUTDOWN_MSG:
>> > >>> /************************************************************
>> > >>> SHUTDOWN_MSG: Shutting down TaskTracker at slave/172.16.0.58
>> > >>> ************************************************************/
>> > >>>
>> > >>>
>> > >>> And all the ports are running
>> > >>> Some time it asks password and some time it wont while starting the
>> > dfs
>> > >>>
>> > >>> Master logs
>> > >>> 2008-01-02 14:44:02,677 INFO org.apache.hadoop.dfs.NameNode:
>> > >>> SHUTDOWN_MSG:
>> > >>> /************************************************************
>> > >>> SHUTDOWN_MSG: Shutting down NameNode at master/172.16.0.25
>> > >>> ************************************************************/
>> > >>>
>> > >>> Datanode-master.log
>> > >>> 2008-01-02 16:26:32,380 INFO org.apache.hadoop.ipc.RPC: Server at
>> > >>> localhost/127.0.0.1:54310 not available yet, Zzzzz...
>> > >>> 2008-01-02 16:26:33,390 INFO org.apache.hadoop.ipc.Client: Retrying
>> > >>> connect
>> > >>> to server: localhost/127.0.0.1:54310. Already tried 1 time(s).
>> > >>> 2008-01-02 16:26:34,400 INFO org.apache.hadoop.ipc.Client: Retrying
>> > >>> connect
>> > >>> to server: localhost/127.0.0.1:54310. Already tried 2 time(s).
>> > >>> 2008-01-02 16:26:35,410 INFO org.apache.hadoop.ipc.Client: Retrying
>> > >>> connect
>> > >>> to server: localhost/127.0.0.1:54310. Already tried 3 time(s).
>> > >>> 2008-01-02 16:26:36,420 INFO org.apache.hadoop.ipc.Client: Retrying
>> > >>> connect
>> > >>> to server: localhost/127.0.0.1:54310. Already tried 4 time(s).
>> > >>> ***********************************************
>> > >>> Jobtracker_master.log
>> > >>> 2008-01-02 16:25:41,040 INFO org.apache.hadoop.ipc.Client: Retrying
>> > >>> connect
>> > >>> to server: localhost/127.0.0.1:54310. Already tried 10 time(s).
>> > >>> 2008-01-02 16:25:42,050 INFO org.apache.hadoop.mapred.JobTracker:
>> > >>> problem
>> > >>> cleaning system directory: /tmp/hadoop-hdpusr/mapred/system
>> > >>> java.net.ConnectException: Connection refused
>> > >>>     at java.net.PlainSocketImpl.socketConnect(Native Method)
>> > >>>     at java.net.PlainSocketImpl.doConnect(PlainSocketImpl.java:333)
>> > >>>     at java.net.PlainSocketImpl.connectToAddress(
>> PlainSocketImpl.java
>> > :195)
>> > >>>     at java.net.PlainSocketImpl.connect(PlainSocketImpl.java:182)
>> > >>>     at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:366)
>> > >>>     at java.net.Socket.connect(Socket.java:520)
>> > >>>     at
>> > >>> org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java
>> > :152)
>> > >>>     at org.apache.hadoop.ipc.Client.getConnection(Client.java:542)
>> > >>>     at org.apache.hadoop.ipc.Client.call(Client.java:471)
>> > >>>     at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:184)
>> > >>>     at org.apache.hadoop.dfs.$Proxy0.getProtocolVersion(Unknown
>> > Source)
>> > >>>     at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:269)
>> > >>>     at
>> org.apache.hadoop.dfs.DFSClient.createNamenode(DFSClient.java
>> > :147)
>> > >>>     at org.apache.hadoop.dfs.DFSClient.<init>(DFSClient.java:161)
>> > >>>     at
>> > >>> org.apache.hadoop.dfs.DistributedFileSystem.initialize(
>> > DistributedFileSystem.java:65)
>> > >>>     at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:159)
>> > >>>     at
>> org.apache.hadoop.fs.FileSystem.getNamed(FileSystem.java:118)
>> > >>>     at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:90)
>> > >>>     at org.apache.hadoop.mapred.JobTracker.<init>(JobTracker.java
>> :683)
>> > >>>     at
>> > >>> org.apache.hadoop.mapred.JobTracker.startTracker(JobTracker.java
>> :120)
>> > >>>     at org.apache.hadoop.mapred.JobTracker.main(JobTracker.java
>> :2052)
>> > >>> 2008-01-02 16:25:42,931 INFO org.apache.hadoop.ipc.Server: IPC
>> Server
>> > >>> handler 5 on 54311, call getFilesystemName() from 127.0.0.1:49283:
>> > >>> error:
>> > >>>
>> org.apache.hadoop.mapred.JobTracker$IllegalStateException:FileSystem
>> > >>> object
>> > >>> not available yet
>> > >>>
>> org.apache.hadoop.mapred.JobTracker$IllegalStateException:FileSystem
>> > >>> object
>> > >>> not available yet
>> > >>>     at
>> > >>> org.apache.hadoop.mapred.JobTracker.getFilesystemName(
>> JobTracker.java
>> > :1475)
>> > >>>     at sun.reflect.GeneratedMethodAccessor2.invoke(Unknown Source)
>> > >>>     at
>> > >>> sun.reflect.DelegatingMethodAccessorImpl.invoke(
>> > DelegatingMethodAccessorImpl.java:25)
>> > >>>     at java.lang.reflect.Method.invoke(Method.java:585)
>> > >>>     at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:379)
>> > >>>     at org.apache.hadoop.ipc.Server$Handler.run(Server.java:596)
>> > >>> 2008-01-02 16:25:47,942 INFO org.apache.hadoop.ipc.Server: IPC
>> Server
>> > >>> handler 6 on 54311, call getFilesystemName() from 127.0.0.1:49293:
>> > >>> error:
>> > >>>
>> org.apache.hadoop.mapred.JobTracker$IllegalStateException:FileSystem
>> > >>> object
>> > >>> not available yet
>> > >>>
>> org.apache.hadoop.mapred.JobTracker$IllegalStateException:FileSystem
>> > >>> object
>> > >>> not available yet
>> > >>>     at
>> > >>> org.apache.hadoop.mapred.JobTracker.getFilesystemName(
>> JobTracker.java
>> > :1475)
>> > >>>     at sun.reflect.GeneratedMethodAccessor2.invoke(Unknown Source)
>> > >>>     at
>> > >>> sun.reflect.DelegatingMethodAccessorImpl.invoke(
>> > DelegatingMethodAccessorImpl.java:25)
>> > >>>     at java.lang.reflect.Method.invoke(Method.java:585)
>> > >>>     at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:379)
>> > >>>     at org.apache.hadoop.ipc.Server$Handler.run(Server.java:596)
>> > >>> 2008-01-02 16:25:52,061 INFO org.apache.hadoop.ipc.Client: Retrying
>> > >>> connect
>> > >>> to server: localhost/127.0.0.1:54310. Already tried 1 time(s).
>> > >>> 2008-01-02 16:25:52,951 INFO org.apache.hadoop.ipc.Server: IPC
>> Server
>> > >>> handler 7 on 54311, call getFilesystemName() from 127.0.0.1:49304:
>> > >>> error:
>> > >>>
>> org.apache.hadoop.mapred.JobTracker$IllegalStateException:FileSystem
>> > >>> object
>> > >>> not available yet
>> > >>>
>> org.apache.hadoop.mapred.JobTracker$IllegalStateException:FileSystem
>> > >>> object
>> > >>> not available yet
>> > >>>     at
>> > >>> org.apache.hadoop.mapred.JobTracker.getFilesystemName(
>> JobTracker.java
>> > :1475)
>> > >>>     at sun.reflect.GeneratedMethodAccessor2.invoke(Unknown Source)
>> > >>>     at
>> > >>> sun.reflect.DelegatingMethodAccessorImpl.invoke(
>> > DelegatingMethodAccessorImpl.java:25)
>> > >>>     at java.lang.reflect.Method.invoke(Method.java:585)
>> > >>>     at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:379)
>> > >>>     at org.apache.hadoop.ipc.Server$Handler.run(Server.java:596)
>> > >>> 2008-01-02 16:25:53,070 INFO org.apache.hadoop.ipc.Client: Retrying
>> > >>> connect
>> > >>> to server: localhost/127.0.0.1:54310. Already tried 2 time(s).
>> > >>> 2008-01-02 16:25:54,080 INFO org.apache.hadoop.ipc.Client: Retrying
>> > >>> connect
>> > >>> to server: localhost/127.0.0.1:54310. Already tried 3 time(s).
>> > >>> 2008-01-02 16:25:55,090 INFO org.apache.hadoop.ipc.Client: Retrying
>> > >>> connect
>> > >>> to server: localhost/127.0.0.1:54310. Already tried 4 time(s).
>> > >>> 2008-01-02 16:25:56,100 INFO org.apache.hadoop.ipc.Client: Retrying
>> > >>> connect
>> > >>> to server: localhost/127.0.0.1:54310. Already tried 5 time(s).
>> > >>> 2008-01-02 16:25:56,281 INFO org.apache.hadoop.mapred.JobTracker:
>> > >>> SHUTDOWN_MSG:
>> > >>> /************************************************************
>> > >>> SHUTDOWN_MSG: Shutting down JobTracker at master/172.16.0.25
>> > >>> ************************************************************/
>> > >>>
>> > >>> Tasktracker_master.log
>> > >>> 2008-01-02 16:26:14,080 INFO org.apache.hadoop.ipc.Client: Retrying
>> > >>> connect
>> > >>> to server: localhost/127.0.0.1:54311. Already tried 2 time(s).
>> > >>> 2008-01-02 16:28:34,510 INFO org.apache.hadoop.mapred.TaskTracker:
>> > >>> STARTUP_MSG:
>> > >>> /************************************************************
>> > >>> STARTUP_MSG: Starting TaskTracker
>> > >>> STARTUP_MSG:   host = master/172.16.0.25
>> > >>> STARTUP_MSG:   args = []
>> > >>> ************************************************************/
>> > >>> 2008-01-02 16:28:34,739 INFO org.mortbay.util.Credential: Checking
>> > >>> Resource
>> > >>> aliases
>> > >>> 2008-01-02 16:28:34,827 INFO org.mortbay.http.HttpServer: Version
>> > >>> Jetty/5.1.4
>> > >>> 2008-01-02 16:28:35,281 INFO org.mortbay.util.Container: Started
>> > >>> [EMAIL PROTECTED]
>> > >>> 2008-01-02 16:28:35,332 INFO org.mortbay.util.Container: Started
>> > >>> WebApplicationContext[/,/]
>> > >>> 2008-01-02 16:28:35,332 INFO org.mortbay.util.Container: Started
>> > >>> HttpContext[/logs,/logs]
>> > >>> 2008-01-02 16:28:35,332 INFO org.mortbay.util.Container: Started
>> > >>> HttpContext[/static,/static]
>> > >>> 2008-01-02 16:28:35,336 INFO org.mortbay.http.SocketListener:
>> Started
>> > >>> SocketListener on 0.0.0.0:50060
>> > >>> 2008-01-02 16:28:35,336 INFO org.mortbay.util.Container: Started
>> > >>> [EMAIL PROTECTED]
>> > >>> 2008-01-02 16:28:35,383 INFO
>> org.apache.hadoop.metrics.jvm.JvmMetrics:
>> > >>> Initializing JVM Metrics with processName=TaskTracker, sessionId=
>> > >>> 2008-01-02 16:28:35,402 INFO org.apache.hadoop.mapred.TaskTracker:
>> > >>> TaskTracker up at: /127.0.0.1:49599
>> > >>> 2008-01-02 16:28:35,402 INFO org.apache.hadoop.mapred.TaskTracker:
>> > >>> Starting
>> > >>> tracker tracker_master:/127.0.0.1:49599
>> > >>> 2008-01-02 16:28:35,406 INFO org.apache.hadoop.ipc.Server: IPC
>> Server
>> > >>> listener on 49599: starting
>> > >>> 2008-01-02 16:28:35,406 INFO org.apache.hadoop.ipc.Server: IPC
>> Server
>> > >>> handler 0 on 49599: starting
>> > >>> 2008-01-02 16:28:35,406 INFO org.apache.hadoop.ipc.Server: IPC
>> Server
>> > >>> handler 1 on 49599: starting
>> > >>> 2008-01-02 16:28:35,490 INFO org.apache.hadoop.mapred.TaskTracker:
>> > >>> Starting
>> > >>> thread: Map-events fetcher for all reduce tasks on
>> > >>> tracker_master:/127.0.0.1:49599
>> > >>> 2008-01-02 16:28:35,500 INFO org.apache.hadoop.mapred.TaskTracker:
>> > Lost
>> > >>> connection to JobTracker [localhost/127.0.0.1:54311].  Retrying...
>> > >>> org.apache.hadoop.ipc.RemoteException:
>> > >>>
>> org.apache.hadoop.mapred.JobTracker$IllegalStateException:FileSystem
>> > >>> object
>> > >>> not available yet
>> > >>>     at
>> > >>> org.apache.hadoop.mapred.JobTracker.getFilesystemName(
>> JobTracker.java
>> > :1475)
>> > >>>     at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>> > >>>     at
>> > >>> sun.reflect.NativeMethodAccessorImpl.invoke(
>> > NativeMethodAccessorImpl.java:39)
>> > >>>     at
>> > >>> sun.reflect.DelegatingMethodAccessorImpl.invoke(
>> > DelegatingMethodAccessorImpl.java:25)
>> > >>>     at java.lang.reflect.Method.invoke(Method.java:585)
>> > >>>     at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:379)
>> > >>>     at org.apache.hadoop.ipc.Server$Handler.run(Server.java:596)
>> > >>>
>> > >>>     at org.apache.hadoop.ipc.Client.call(Client.java:482)
>> > >>>     at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:184)
>> > >>>     at org.apache.hadoop.mapred.$Proxy0.getFilesystemName(Unknown
>> > Source)
>> > >>>     at
>> > >>> org.apache.hadoop.mapred.TaskTracker.offerService(TaskTracker.java
>> > :773)
>> > >>>     at org.apache.hadoop.mapred.TaskTracker.run(TaskTracker.java
>> :1179)
>> > >>>     at org.apache.hadoop.mapred.TaskTracker.main(TaskTracker.java
>> > :1880)
>> > >>> *******************************************
>> > >>>
>> > >>> Please help me to resolve the same.
>> > >>>
>> > >>>
>> > >>> Khalil Honsali wrote:
>> > >>>
>> > >>>>Hi,
>> > >>>>
>> > >>>>I think you need to post more information, for example an excerpt
>> of
>> > the
>> > >>>>failing datanode log. Also, please clarify the issue of
>> connectivity:
>> > >>>>- are you able to ssh passwordless (from master to slave, slave to
>> > master,
>> > >>>>slave to slave, master to master), you shouldn't be entering
>> passwrd
>> > >>>>everytime...
>> > >>>>- are you able to telnet (not necessary but preferred)
>> > >>>>- have you verified the ports as RUNNING on using netstat command?
>> > >>>>
>> > >>>>besides, the tasktracker starts ok but not the datanode?
>> > >>>>
>> > >>>>K. Honsali
>> > >>>>
>> > >>>>On 02/01/2008, Dhaya007 <[EMAIL PROTECTED]> wrote:
>> > >>>>
>> > >>>>>
>> > >>>>>I am new to hadoop if any think wrong please correct me ....
>> > >>>>>I Have configured a single/multi node cluster using following link
>> > >>>>>
>> > >>>>>
>> >
>> http://www.michael-noll.com/wiki/Running_Hadoop_On_Ubuntu_Linux_%28Single-Node_Cluster%29
>> > >>>>>.
>> > >>>>>I have followed the link but i am not able to start the haoop in
>> > multi
>> > >>>>>node
>> > >>>>>environment
>> > >>>>>The problems i am facing are as Follows:
>> > >>>>>1.I have configured master and slave nodes with ssh less pharase
>> if
>> > try
>> > >>>>>to
>> > >>>>>run the start-dfs.sh it prompt the password for master:slave
>> > machines.(I
>> > >>>>>have copied the .ssh/id_rsa.pub key of master in to slaves
>> > autherized_key
>> > >>>>>file)
>> > >>>>>
>> > >>>>>2.After giving password datanode,namenode,jobtracker,tasktraker
>> > started
>> > >>>>>successfully in master but datanode is started in slave.
>> > >>>>>
>> > >>>>>
>> > >>>>>3.Some time step 2 works and some time it says that permission
>> > denied.
>> > >>>>>
>> > >>>>>4.I have checked the log file in the slave for datanode it says
>> that
>> > >>>>>incompatible node, then i have formated the slave, master and
>> start
>> > the
>> > >>>>>dfs
>> > >>>>>by start-dfs.sh still i am getting the error
>> > >>>>>
>> > >>>>>
>> > >>>>>The host entry in etc/hosts are both master/slave
>> > >>>>>master
>> > >>>>>slave
>> > >>>>>conf/masters
>> > >>>>>master
>> > >>>>>conf/slaves
>> > >>>>>master
>> > >>>>>slave
>> > >>>>>
>> > >>>>>The hadoop-site.xml  for both master/slave
>> > >>>>><?xml version="1.0"?>
>> > >>>>><?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
>> > >>>>>
>> > >>>>><!-- Put site-specific property overrides in this file. -->
>> > >>>>>
>> > >>>>><configuration>
>> > >>>>><property>
>> > >>>>>  <name>hadoop.tmp.dir</name>
>> > >>>>>  <value>/home/hdusr/hadoop-${user.name}</value>
>> > >>>>>  <description>A base for other temporary
>> directories.</description>
>> > >>>>></property>
>> > >>>>>
>> > >>>>><property>
>> > >>>>>  <name>fs.default.name</name>
>> > >>>>>  <value>hdfs://master:54310</value>
>> > >>>>>  <description>The name of the default file system.  A URI whose
>> > >>>>>  scheme and authority determine the FileSystem
>> implementation.  The
>> > >>>>>  uri's scheme determines the config property (fs.SCHEME.impl)
>> naming
>> > >>>>>  the FileSystem implementation class.  The uri's authority is
>> used
>> > to
>> > >>>>>  determine the host, port, etc. for a filesystem.</description>
>> > >>>>></property>
>> > >>>>>
>> > >>>>><property>
>> > >>>>>  <name>mapred.job.tracker</name>
>> > >>>>>  <value>master:54311</value>
>> > >>>>>  <description>The host and port that the MapReduce job tracker
>> runs
>> > >>>>>  at.  If "local", then jobs are run in-process as a single map
>> > >>>>>  and reduce task.
>> > >>>>>  </description>
>> > >>>>></property>
>> > >>>>>
>> > >>>>><property>
>> > >>>>>  <name>dfs.replication</name>
>> > >>>>>  <value>2</value>
>> > >>>>>  <description>Default block replication.
>> > >>>>>  The actual number of replications can be specified when the file
>> is
>> > >>>>>created.
>> > >>>>>  The default is used if replication is not specified in create
>> time.
>> > >>>>>  </description>
>> > >>>>></property>
>> > >>>>>
>> > >>>>><property>
>> > >>>>>  <name>mapred.map.tasks</name>
>> > >>>>>  <value>20</value>
>> > >>>>>  <description>As a rule of thumb, use 10x the number of slaves (
>> i.e
>> > .,
>> > >>>>>number of tasktrackers).
>> > >>>>>  </description>
>> > >>>>></property>
>> > >>>>>
>> > >>>>><property>
>> > >>>>>  <name>mapred.reduce.tasks</name>
>> > >>>>>  <value>4</value>
>> > >>>>>  <description>As a rule of thumb, use 2x the number of slave
>> > >>>>> processors
>> > >>>>>(i.e., number of tasktrackers).
>> > >>>>>  </description>
>> > >>>>></property>
>> > >>>>></configuration>
>> > >>>>>
>> > >>>>>Please help me to reslove the same. Or else provide any other
>> > tutorial
>> > >>>>>for
>> > >>>>>multi node cluster setup.I am egarly waiting for the tutorials.
>> > >>>>>
>> > >>>>>
>> > >>>>>Thanks
>> > >>>>>
>> > >>>>>--
>> > >>>>>View this message in context:
>> > >>>>>
>> >
>> http://www.nabble.com/Not-able-to-start-Data-Node-tp14573889p14573889.html
>> > >>>>>Sent from the Hadoop Users mailing list archive at Nabble.com.
>> > >>>>>
>> > >>>>>
>> > >>>>
>> > >>>>
>> > >>>
>> > >>
>> > >>
>> > >>
>> > >
>> > >
>> >
>> > --
>> > View this message in context:
>> >
>> http://www.nabble.com/Not-able-to-start-Data-Node-tp14573889p14594256.html
>> > Sent from the Hadoop Users mailing list archive at Nabble.com.
>> >
>> >
>>
> 
> 
> 
> -- 
> ---------------------------------------------------------
> شهر مبارك كريم
> كل عام و أنتم بخير
> ---------------------------------------------------------
> Honsali Khalil − 本査理 カリル
> Academic>Japan>NIT>Grad. Sc. Eng.>Dept. CS>Matsuo&Tsumura Lab.
> http://www.matlab.nitech.ac.jp/~k-hon/
> +81 (zero-)eight-zero 5134 8119
> [EMAIL PROTECTED] (instant reply mail)
> 
> 

-- 
View this message in context: 
http://www.nabble.com/Not-able-to-start-Data-Node-tp14573889p14613800.html
Sent from the Hadoop Users mailing list archive at Nabble.com.

Reply via email to