Re: Requesting containers on a specific host

2013-07-05 Thread devara...@huawei.com
Hi Kishore, As per the exception given, Node Manager is getting excluded. It might be happening that you have configured the Node Manager in excluded file using this configuration in Resource Manager. Could you check this configuration in RM, is it configured with any file and that file

Re: Which InputFormat to use?

2013-07-05 Thread Azuryy Yu
Using InputFormat under mapreduce package. mapred package is very old package. but generally you can extend from FileInputFormat under o.a.h.mapreduce package. On Fri, Jul 5, 2013 at 1:23 PM, Devaraj k devara...@huawei.com wrote: Hi Ahmed, ** ** Hadoop 0.20.0 included

Re: Decomssion datanode - no response

2013-07-05 Thread Azuryy Yu
I filed this issue at : https://issues.apache.org/jira/browse/HDFS-4959 On Fri, Jul 5, 2013 at 1:06 PM, Azuryy Yu azury...@gmail.com wrote: Client hasn't any connection problem. On Fri, Jul 5, 2013 at 12:46 PM, Devaraj k devara...@huawei.com wrote: And also could you check whether the

Re: How to update a file which is in HDFS

2013-07-05 Thread Robin East
The API for 1.1.2 FileSystem seems to include append(). Robin On 5 Jul 2013, at 01:50, Mohammad Tariq donta...@gmail.com wrote: The current stable release doesn't support append, not even through the API. If you really want this you have to switch to hadoop 2.x. See this JIRA. Warm

Re: How to update a file which is in HDFS

2013-07-05 Thread Robin East
Ok just read the JIRA in detail (pays to read these things before posting). It says: Append is not supported in Hadoop 1.x. Please upgrade to 2.x if you need append. If you enabled dfs.support.append for HBase, you're OK, as durable sync (why HBase required dfs.support.append) is now enabled

Re: How to update a file which is in HDFS

2013-07-05 Thread Harsh J
The answer to the delta part is more that HDFS does not presently support random writes. You cannot alter a closed file for anything other than appending at the end, which I doubt will help you if you are also receiving updates (it isn't clear from your question what this added data really is).

Re: DistCp fails on open files

2013-07-05 Thread Harsh J
Hi, If you are looking to make it ignore those files, you can probably consider using the -i flag. On Thu, Jul 4, 2013 at 6:51 PM, Manuel de Ferran manuel.defer...@gmail.com wrote: Hi all, I'm trying to copy files from a source HDFS cluster to another. But I have numerous files open for

Re: How to update a file which is in HDFS

2013-07-05 Thread Harsh J
If it is 1k new records at the end of the file then you may extract them out and append the existing file in HDFS. I'd recommend using HDFS from Apache Hadoop 2.x for this purpose. On Fri, Jul 5, 2013 at 4:22 PM, Manickam P manicka...@outlook.com wrote: Hi, Let me explain the question clearly.

Re: Decomssion datanode - no response

2013-07-05 Thread varun kumar
Try to give IPaddressofDatanode:50010 On Fri, Jul 5, 2013 at 12:25 PM, Azuryy Yu azury...@gmail.com wrote: I filed this issue at : https://issues.apache.org/jira/browse/HDFS-4959 On Fri, Jul 5, 2013 at 1:06 PM, Azuryy Yu azury...@gmail.com wrote: Client hasn't any connection problem.

Re: Requesting containers on a specific host

2013-07-05 Thread Krishna Kishore Bonagiri
Hi Devaraj, Thanks for pointing me to this RC. I am trying this out, and getting this error for NM to get started. My RM is running fine, but NM is failing and saying that it is disallowed by RM and received a SHUTDOWN message. Please give me clue to resolve this. 2013-07-05 09:49:20,043

Accessing HDFS

2013-07-05 Thread John Lilley
I've seen mentioned that you can access HDFS via ClientProtocol, as in: ClientProtocol namenode = DFSClient.createNamenode(conf); LocatedBlocks lbs = namenode.getBlockLocations(path, start, length); But we use: fs = FileSystem.get(URI, conf); filestatus = fs.getFileStatus(path);

starting Hadoop, the new way

2013-07-05 Thread Jay Vyas
Hi : Is there a hadoop 2.0 tutorial for 1.0 people ? Im used to running start-all.sh , but it appears that the new MR2 version of hadoop is much more sophisticated. In any case, Im wondering what the standard way to start the new generation of hadoop/mr2 hadoop/mapreduce and hadoop/hdfs is and

Re: How to update a file which is in HDFS

2013-07-05 Thread Mohammad Tariq
@Robin East : Thank you for keeping me updated. I was on 1.0.3 when I had tried append last time and it was not working despite of the fact that API had it. I tried it with 1.1.2 and it seems to work fine. @Manickam : Apologies for the incorrect info. Latest stable release(1.1.2) supports

Re: How to update a file which is in HDFS

2013-07-05 Thread Harsh J
The append in 1.x is very broken. You'll run into very weird states and we officially do not support it (we even call out in the config as broken). I wouldn't recommend using it even if a simple test appears to work. On Sat, Jul 6, 2013 at 6:27 AM, Mohammad Tariq donta...@gmail.com wrote: @Robin

Re: How to update a file which is in HDFS

2013-07-05 Thread Mohammad Tariq
I totally agree harsh. It was just to avoid any misinterpretation :). I have seen quite a few discussions as well that talk about the issues. I would strongly recommend to switch from 1.x if append is desired. Warm Regards, Tariq cloudfront.blogspot.com On Sat, Jul 6, 2013 at 7:29 AM, Harsh J

Re: Accessing HDFS

2013-07-05 Thread Harsh J
These APIs (ClientProtocol, DFSClient) are not for Public access. Please do not use them in production. The only API we care not to change incompatibly are the FileContext and the FileSystem APIs. They provide much of what you want - if not, log a JIRA. On Fri, Jul 5, 2013 at 11:40 PM, John

Debugging Hadoop Environment

2013-07-05 Thread Daniel Krahn
Hi, I would like to ask a few questions on how to debug Hadoop. First I'll explain what I'm trying to do. I'm doing my graduation thesis. Are group is starting an implementation on a voluntary environment. For this it is very important to have all the results signed so that we can guarantee that

New to hadoop - which version to start with ?

2013-07-05 Thread sudhir543-...@yahoo.com
I am new to hadoop, just started reading 'hadoop the definitive guide'. I downloaded hadoop 1.1.2 and tried to run a sample Map reduce job using cygwin, but I got following error java.io.IOException: Failed to set permissions of path:

Re: New to hadoop - which version to start with ?

2013-07-05 Thread Azuryy Yu
hadoop fs -chmod -R 755 \tmp\hadoop-Sudhir\mapred\staging Then It should works. On Sat, Jul 6, 2013 at 1:27 PM, sudhir543-...@yahoo.com sudhir543-...@yahoo.com wrote: I am new to hadoop, just started reading 'hadoop the definitive guide'. I downloaded hadoop 1.1.2 and tried to run a sample