Are you encountering specific problems? I don't think that hadoop's config files will evaluate environment variables. So $HADOOP_HOME won't be interpreted correctly.
For passwordless ssh, see http://rcsg-gsir.imsb-dsgi.nrc-cnrc.gc.ca/documents/internet/node31.html or just check the manpage for ssh-keygen. - Aaron On Wed, Jun 17, 2009 at 9:30 AM, Divij Durve <divij.t...@gmail.com> wrote: > Im trying to setup a cluster with 3 different machines running Fedora. I > cant get them to log into the localhost without the password but thats the > least of my worries at the moment. > > I am posting my config files and the master and slave files let me know if > anyone can spot a problem with the configs... > > > Hadoop-site.xml > <?xml version="1.0"?> > <?xml-stylesheet type="text/xsl" href="configuration.xsl"?> > > <!-- Put site-specific property overrides in this file. --> > > <configuration> > <property> > <name>dfs.data.dir</name> > <value>$HADOOP_HOME/dfs-data</value> > <final>true</final> > </property> > > <property> > <name>dfs.name.dir</name> > <value>$HADOOP_HOME/dfs-name</value> > <final>true</final> > </property> > > > <property> > <name>hadoop.tmp.dir</name> > <value>$HADOOP_HOME/hadoop-tmp</value> > <description>A base for other temporary directories.</description> > </property> > > > <property> > <name>fs.default.name</name> > <value>hdfs://gobi.<something>.<something>:54310</value> > <description>The name of the default file system. A URI whose > scheme and authority determine the FileSystem implementation. The > uri's scheme determines the config property (fs.SCHEME.impl) > naming > the FileSystem implementation class. The uri's authority is > used to > determine the host, port, etc. for a FileSystem.</description> > </property> > > <property> > <name>mapred.job.tracker</name> > <value>kalahari.<something>.<something>:54311</value> > <description>The host and port that the MapReduce job tracker runs > at. If "local", then jobs are run in-process as a single map > and reduce task. > </description> > </property> > > <property> > <name>mapred.system.dir</name> > <value>$HADOOP_HOME/mapred-system</value> > <final>true</final> > </property> > > <property> > <name>dfs.replication</name> > <value>1</value> > <description>Default block replication. > The actual number of replications can be specified when the file is > created. > The default is used if replication is not specified in create > time. > </description> > </property> > > > <property> > <name>mapred.local.dir</name> > <value>$HADOOP_HOME/mapred-local</value> > <name>dfs.replication</name> > <value>1</value> > </property> > > > </configuration> > > > Slave: > kongur.something.something > > master: > kalahari.something.something > > i execute the dfs-start.sh command from gobi.something.something. > > is there any other info that i should provide in order to help? Also Kongur > is where im running the data node the master file on kongur should have > localhost in it rite? thanks for the help > > Divij >