I'm not sure what to do about your specific problem, but here's how to look inside DFS: use the hadoop executable in bin/ to execute common filesystem operations (ls, cp, rm, &c). For instance:

bin/hadoop dfs -ls /

does an "ls /" on DFS.

Go to http://wiki.apache.org/lucene-hadoop/hadoop-0.1-dev/bin/ hadoop_dfs?highlight=%28dfs%29 for more detailed documentation. I found it pretty intuitive, but watch out for the one important quirk: "a path may be relative or absolute. An absolute path starts with a '/', a relative path does not, and always relates to /user/ <currentUser>. There is no notion of current working directory."

    ~ Esteban

On May 23, 2007, at 3:12 PM, Avinash Lakshman wrote:

Here is the problem I get when I run the Word Count sample with a file that is over 130MB. It looks like the map tasks never complete. From my usage I am assuming that there are 3 map tasks and 2 reduce tasks correct ? Also
where can I see the intermediate files that are generated ?

Thanks
Avinash

java org.apache.hadoop.examples.WordCount -m 3 -r 2
/home/alakshman/hadoop-0.12.3/in-dir/test2.dat /tmp/out-dir
07/05/23 15:06:21 INFO mapred.InputFormatBase: Total input paths to process
: 1
07/05/23 15:06:21 INFO mapred.JobClient: Running job: job_ltsbiw
[EMAIL PROTECTED] examples]$ java org.apache.hadoop.examples.WordCount -m 2
-r 3  /home/alakshman/hadoop-0.12.3/in-dir/test2.dat /tmp/out-dir
07/05/23 15:06:30 INFO mapred.InputFormatBase: Total input paths to process
: 1
07/05/23 15:06:30 INFO mapred.JobClient: Running job: job_if514q
07/05/23 15:06:31 INFO mapred.JobClient:  map 0% reduce 0%
07/05/23 15:06:32 INFO mapred.LocalJobRunner:
file:/home/alakshman/hadoop-0.12.3/in-dir/test2.dat:0+33554432
07/05/23 15:06:32 INFO mapred.JobClient:  map 5% reduce 0%
07/05/23 15:06:33 INFO mapred.LocalJobRunner:
file:/home/alakshman/hadoop-0.12.3/in-dir/test2.dat:0+33554432
07/05/23 15:06:33 INFO mapred.JobClient:  map 13% reduce 0%
07/05/23 15:06:34 INFO mapred.LocalJobRunner:
file:/home/alakshman/hadoop-0.12.3/in-dir/test2.dat:0+33554432
07/05/23 15:06:34 INFO mapred.JobClient:  map 22% reduce 0%
07/05/23 15:06:35 INFO mapred.LocalJobRunner:
file:/home/alakshman/hadoop-0.12.3/in-dir/test2.dat:0+33554432
07/05/23 15:06:35 INFO mapred.JobClient:  map 30% reduce 0%
07/05/23 15:06:36 INFO mapred.LocalJobRunner:
file:/home/alakshman/hadoop-0.12.3/in-dir/test2.dat:0+33554432
07/05/23 15:06:36 INFO mapred.JobClient:  map 39% reduce 0%
07/05/23 15:06:37 INFO mapred.LocalJobRunner:
file:/home/alakshman/hadoop-0.12.3/in-dir/test2.dat:0+33554432
07/05/23 15:06:37 INFO mapred.JobClient:  map 46% reduce 0%
07/05/23 15:06:38 INFO mapred.LocalJobRunner:
file:/home/alakshman/hadoop-0.12.3/in-dir/test2.dat:0+33554432
07/05/23 15:06:38 INFO mapred.JobClient:  map 55% reduce 0%
07/05/23 15:06:39 INFO mapred.LocalJobRunner:
file:/home/alakshman/hadoop-0.12.3/in-dir/test2.dat:0+33554432
07/05/23 15:06:39 INFO mapred.JobClient:  map 63% reduce 0%
07/05/23 15:06:40 INFO mapred.LocalJobRunner:
file:/home/alakshman/hadoop-0.12.3/in-dir/test2.dat:0+33554432
07/05/23 15:06:40 INFO mapred.JobClient:  map 73% reduce 0%
07/05/23 15:06:41 INFO mapred.LocalJobRunner:
file:/home/alakshman/hadoop-0.12.3/in-dir/test2.dat:0+33554432
07/05/23 15:06:41 INFO mapred.JobClient:  map 81% reduce 0%
07/05/23 15:06:42 INFO mapred.LocalJobRunner:
file:/home/alakshman/hadoop-0.12.3/in-dir/test2.dat:0+33554432
07/05/23 15:06:42 INFO mapred.JobClient:  map 89% reduce 0%
07/05/23 15:06:43 INFO mapred.LocalJobRunner:
file:/home/alakshman/hadoop-0.12.3/in-dir/test2.dat:0+33554432
07/05/23 15:06:43 INFO mapred.JobClient:  map 92% reduce 0%
07/05/23 15:06:44 INFO mapred.LocalJobRunner:
file:/home/alakshman/hadoop-0.12.3/in-dir/test2.dat:0+33554432
07/05/23 15:06:45 INFO mapred.LocalJobRunner:
file:/home/alakshman/hadoop-0.12.3/in-dir/test2.dat:0+33554432
07/05/23 15:06:46 INFO mapred.LocalJobRunner:
file:/home/alakshman/hadoop-0.12.3/in-dir/test2.dat:0+33554432
07/05/23 15:06:47 INFO mapred.LocalJobRunner:
file:/home/alakshman/hadoop-0.12.3/in-dir/test2.dat:0+33554432
07/05/23 15:06:48 INFO mapred.LocalJobRunner:
file:/home/alakshman/hadoop-0.12.3/in-dir/test2.dat:0+33554432
07/05/23 15:06:49 INFO mapred.LocalJobRunner:
file:/home/alakshman/hadoop-0.12.3/in-dir/test2.dat:0+33554432
07/05/23 15:06:50 INFO mapred.LocalJobRunner:
file:/home/alakshman/hadoop-0.12.3/in-dir/test2.dat:0+33554432
07/05/23 15:06:51 INFO mapred.LocalJobRunner:
file:/home/alakshman/hadoop-0.12.3/in-dir/test2.dat:0+33554432
07/05/23 15:06:52 INFO mapred.LocalJobRunner:
file:/home/alakshman/hadoop-0.12.3/in-dir/test2.dat:0+33554432
07/05/23 15:06:53 INFO mapred.LocalJobRunner:
file:/home/alakshman/hadoop-0.12.3/in-dir/test2.dat:0+33554432



On 5/23/07 2:53 PM, "Arun C Murthy" <[EMAIL PROTECTED]> wrote:

On Wed, May 23, 2007 at 02:50:20PM -0700, Avinash Lakshman wrote:
I want to setup Hadoop to run on a cluster of 3 nodes. Here is the setup of
my hadoop-site.xml file :

<configuration>
       <property>
               <name>fs.default.name</name>
               <value>hdfs:///</value>
       </property>

Fix fs.default.name's value to <host>:<port> similar to your mapre.job.tracker
and give it a try...

Arun


       <property>
               <name>dfs.secondary.info.port</name>
               <value>10000</value>
       </property>

       <property>
               <name>mapred.job.tracker</name>
               <value>dev030.sctm.facebook.com:9001</value>
       </property>

       <property>
               <name>dfs.replication</name>
               <value>1</value>
       </property>

</configuration>

However whenever I start my DFS I get the following error :

dev030.sctm.facebook.com: starting secondarynamenode, logging to
/home/alakshman/hadoop-0.12.3/bin/../logs/hadoop-alakshman- secondarynamenode
-dev030.sctm.facebook.com.out
dev030.sctm.facebook.com: Exception in thread "main"
java.lang.IllegalArgumentException: port out of range:-1
dev030.sctm.facebook.com:       at
java.net.InetSocketAddress.<init>(InetSocketAddress.java:118)
dev030.sctm.facebook.com:       at
org.apache.hadoop.dfs.DataNode.createSocketAddr(DataNode.java:106)
dev030.sctm.facebook.com:       at
org.apache.hadoop.dfs.SecondaryNameNode.<init> (SecondaryNameNode.java:93)
dev030.sctm.facebook.com:       at
org.apache.hadoop.dfs.SecondaryNameNode.main (SecondaryNameNode.java:474)

How do I fix this ? Please advice.

Thanks
A


Reply via email to