Remote Job Submission

2008-05-23 Thread Natarajan, Senthil
Hi, I was wondering is it possible to submit MapReduce job on remote Hadoop cluster. (i.e) Submitting the job from the machine which doesn't have Hadoop installed and submitting to different machine where Hadoop installed. Is it possible to do this? I guess at least data can be uploaded to HDFS

RE: Hadoop Permission Problem

2008-05-09 Thread Natarajan, Senthil
Subject: Re: Hadoop Permission Problem Hi Senthil, In the error message, it says that the permission for datastore is 755. Are you sure that you have changed it to 777? Nicholas - Original Message From: Natarajan, Senthil [EMAIL PROTECTED] To: core-user@hadoop.apache.org core-user

RE: Hadoop Permission Problem

2008-05-09 Thread Natarajan, Senthil
2 test test 4096 May 9 09:29 hadoop-test From the output format, the directories above seem not HDFS directories. Are you running map/red jobs over local file system (e.g. Linux)? Nicholas - Original Message From: Natarajan, Senthil [EMAIL PROTECTED] To: core-user

RE: Hadoop Permission Problem

2008-05-09 Thread Natarajan, Senthil
@hadoop.apache.org Subject: Re: Hadoop Permission Problem Hi Senthil, I cannot see why it does not work. Could you try again, do a fs -ls right after you see the error message? Nicholas - Original Message From: Natarajan, Senthil [EMAIL PROTECTED] To: core-user@hadoop.apache.org core-user

RE: Hadoop Permission Problem

2008-05-09 Thread Natarajan, Senthil
4096 May 7 18:02 datastore This one is your local directory. I think you might have mixed up the local and hdfs directories. Nicholas - Original Message From: Natarajan, Senthil [EMAIL PROTECTED] To: core-user@hadoop.apache.org core-user@hadoop.apache.org Sent: Friday, May 9

Hadoop Job Submission

2008-05-08 Thread Natarajan, Senthil
Hi, I have some rudimentary question. In order to use Hadoop (both HDFS and MapReduce) does each user whoever wants to run the job needs to start their jobtracker and datanode and submit the job. Or is it possible to start the jobtracker and datanode using the user hadoop and other users

RE: [Reduce task stalls] Problem Detailed Report

2008-05-08 Thread Natarajan, Senthil
May be due to firewall. Try after stopping the iptables. If it works add firewall rules to allow communication between master and slaves (better allow all nodes in the subnet) -Original Message- From: Amit Kumar Singh [mailto:[EMAIL PROTECTED] Sent: Thursday, May 08, 2008 4:50 PM To:

Hadoop Permission Problem

2008-05-07 Thread Natarajan, Senthil
Hi, My datanode and jobtracker are started by user hadoop. And user Test needs to submit the job. So if the user Test copies file to HDFS, there is a permission error. /usr/local/hadoop/bin/hadoop dfs -copyFromLocal /home/Test/somefile.txt myapps copyFromLocal:

RE: Reduce Output

2008-04-15 Thread Natarajan, Senthil
:[EMAIL PROTECTED] Sent: Monday, April 14, 2008 2:20 PM To: core-user@hadoop.apache.org Subject: Re: Reduce Output Try using Text, Text as the output type and use something like a StringBuffer or Formatter to construct a tab-separated list. On 4/14/08 11:13 AM, Natarajan, Senthil [EMAIL PROTECTED

Reduce Output

2008-04-14 Thread Natarajan, Senthil
Hi, I have the reduce output like this. 1.0.2.92206475 1.0.2.9 316475847 1.0.3.933846495 1.0.4.93316975 But I want to display like this... 1.0.2.92206 475 1.0.2.9 316 475 847 1.0.3.93

RE: Reduce Output

2008-04-14 Thread Natarajan, Senthil
the reduce output using conventional methods. This works very well if you can process your data in less time than hadoop takes to start your job. On 4/14/08 9:42 AM, Natarajan, Senthil [EMAIL PROTECTED] wrote: Hi, I have the reduce output like this. 1.0.2.92206475

RE: Reduce Output

2008-04-14 Thread Natarajan, Senthil
The format of the reduce output is the responsibility of the reducer. You can store the output any way you like. On 4/14/08 10:17 AM, Natarajan, Senthil [EMAIL PROTECTED] wrote: Thanks Ted. Actually I was trying to do the third option by myself before posting this question. Problem is I couldn't

RE: Reduce Sort

2008-04-09 Thread Natarajan, Senthil
so that you get descending order. Make sure you set the number of reducers to 1 so that you only get a single output file. If you have less than 10 million values, the conventional sort is likely to be faster simply because of hadoop's startup time. On 4/8/08 8:37 AM, Natarajan, Senthil [EMAIL

RE: Reduce Sort

2008-04-08 Thread Natarajan, Senthil
values, the conventional sort is likely to be faster simply because of hadoop's startup time. On 4/8/08 8:37 AM, Natarajan, Senthil [EMAIL PROTECTED] wrote: Hi, I am new to MapReduce. After slightly modifying the example wordcount, to count the IP Address. I have two files part-0 and part

RE: Reduce Hangs

2008-03-28 Thread Natarajan, Senthil
/hosts with mapping from hostname to ip address of all nodes in cluster. I hope my experience will be helpful for you. On 3/27/08, Natarajan, Senthil [EMAIL PROTECTED] wrote: Hi, I have small Hadoop cluster, one master and three slaves. When I try the example wordcount on one of our log file

Reduce Hangs

2008-03-27 Thread Natarajan, Senthil
Hi, I have small Hadoop cluster, one master and three slaves. When I try the example wordcount on one of our log file (size ~350 MB) Map runs fine but reduce always hangs (sometime around 19%,60% ...) after very long time it finishes. I am seeing this error Shuffle Error: Exceeded

RE: NFS mounted home, host RSA keys, localhost, strict sshds and bad mojo.

2008-03-21 Thread Natarajan, Senthil
I guess the following file might have localhost entry, change to hostname HADOOP_INSTALL/conf/masters HADOOP_INSTALL/conf/slaves -Original Message- From: Colin Freas [mailto:[EMAIL PROTECTED] Sent: Friday, March 21, 2008 12:25 PM To: core-user@hadoop.apache.org Subject: NFS mounted