unsubscribe

2014-03-19 Thread A BlueCoder

Re: Need FileName with Content

2014-03-19 Thread Ranjini Rathinam
Hi, If we give the below code, === word.set("filename"+""+tokenizer.nextToken()); output.collect(word,one); == The output is wrong. because it shows the filename word occurance vinitha java 4 vinitha oracle 3 sony

RE: NullpointerException in OffLineImageViewer

2014-03-19 Thread Brahma Reddy Battula
Seems to be issue,,please file a jira.. https://issues.apache.org/jira From: c.agra...@outlook.com [c.agra...@outlook.com] on behalf of Chetan Agrawal [cagrawa...@gmail.com] Sent: Wednesday, March 19, 2014 12:50 PM To: hadoop users Subject: NullpointerExcep

RE: how to free up space of the old Data Node

2014-03-19 Thread Brahma Reddy Battula
Following Node is down..Please have look on datanode logs and try to make it up...Before going for further action..(like decreasing the replication factor..) Dead datanodes: Name: 172.18.126.99:50010 (nsda3dmsrpt02.internal.bigpond.com) From: Phan, Truong Q

RE: how to free up space of the old Data Node

2014-03-19 Thread Brahma Reddy Battula
Please check my inline comments which are in blue color... From: Phan, Truong Q [troung.p...@team.telstra.com] Sent: Thursday, March 20, 2014 10:28 AM To: user@hadoop.apache.org Subject: RE: how to free up space of the old Data Node Thanks for the reply. This

RE: how to free up space of the old Data Node

2014-03-19 Thread Phan, Truong Q
Hi Battula, I hope Battula is your first name. :P Here are the output of your suggested commands: [root@nsda3dmsrpt02] /usr/lib/hadoop-0.20-mapreduce# sudo -u hdfs hadoop fsck / DEPRECATED: Use of this script to execute hdfs command is deprecated. Instead use the hdfs command for it. Connecting

RE: how to free up space of the old Data Node

2014-03-19 Thread Vinayakumar B
You can change the replication factor using the following command hdfs dfs - setrep [-R] Once this is done, you can re-commission the datanode, then all the overreplicated blocks will be removed. If not removed, restart the datanode. Regards, Vinayakumar B From: Phan, Truong Q [mailto:troung.

Hadoop MapReduce Streaming - how to change the final output file name with the desired name rather than in partition like: part-0000*

2014-03-19 Thread Phan, Truong Q
Hi Could you please provide me an alternative link where it explains on how to change the final output file name with the desired name rather than in partition like: part-*? Can I have a sample Python's code to run MapReduce Streaming with a custome output file names? One helper from the A

RE: how to free up space of the old Data Node

2014-03-19 Thread Phan, Truong Q
Thanks for the reply. This Hadoop cluster is our POC and the node has less space compare to the other two nodes. How do I change the Replication Factore (RF) from 3 down to 2? Is this controlled by this parameter (dfs.datanode.handler.count)? Thanks and Regards, Truong Phan P+ 61 2 8576 577

NullpointerException in OffLineImageViewer

2014-03-19 Thread Chetan Agrawal
I want to access and study the hadoop cluster's metadata which is stored in fsimage file on namenode machine. i came to know that offLineImageViewer is used to do so. But when i am trying doing it i am getting an exception. /usr/hadoop/hadoop-1.2.1# bin/hadoop oiv -i fsimage -o fsimage.txt Warnin

NullPointerException in offLineImageViewer

2014-03-19 Thread Chetan Agrawal
I want to access and study the hadoop cluster's metadata which is stored in fsimage file on namenode machine. i came to know that offLineImageViewer is used to do so. But when i am trying doing it i am getting an exception. /usr/hadoop/hadoop-1.2.1# bin/hadoop oiv -i fsimage -o fsimage.txt Warnin

RE: how to free up space of the old Data Node

2014-03-19 Thread Brahma Reddy Battula
Please check my inline comments which are in blue color... From: Phan, Truong Q [troung.p...@team.telstra.com] Sent: Thursday, March 20, 2014 8:04 AM To: user@hadoop.apache.org Subject: how to free up space of the old Data Node Hi I have 3 nodes Hadoop cluster

how to free up space of the old Data Node

2014-03-19 Thread Phan, Truong Q
Hi I have 3 nodes Hadoop cluster in which I created 3 Data Nodes. However, I don't have enough space in one of the node to cater other projects' log. So I decommissioned this node from a Data node list but I could not re-claimed the space from it. Is there a way to get this node to release space

Re: Need FileName with Content

2014-03-19 Thread Stanley Shi
You want to do a word count for each file, but the code give you a word count for all the files, right? = word.set(tokenizer.nextToken()); output.collect(word, one); == change it to: word.set("filename"+""+tokenizer.nextToken()); output.collect(word,one); Regards, *Stanle

Re: The reduce copier failed

2014-03-19 Thread Harsh J
While it does mean a retry, if the job eventually fails (after finite retries all fail as well), then you have a problem to investigate. If the job eventually succeeded, then this may have been a transient issue. Worth investigating either way. On Thu, Mar 20, 2014 at 12:57 AM, Mahmood Naderan wr

Re: question about yarn webapp

2014-03-19 Thread Harsh J
Hello, This is a hadoop-common functionality. See the StacksServlet class code: https://github.com/apache/hadoop-common/blob/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/http/HttpServer2.java#L1044 On Wed, Mar 19, 2014 at 9:17 PM, 赵璨 wrote: >I can get Server Sta

The reduce copier failed

2014-03-19 Thread Mahmood Naderan
Hi In the middle of a map-reduce job I get map 20% reduce 6% ... The reduce copier failed map 20% reduce 0% map 20% reduce 1% map 20% reduce 2% map 20% reduce 3%   Does that imply a *retry* process? Or I have to be worried about that message? Regards, Mahmood

Re: Doubt

2014-03-19 Thread praveenesh kumar
Why not ? Its just a matter of installing 2 different packages. Depends on what do you want to use it for, you need to take care of few things, but as far as installation is concerned, it should be easily doable. Regards Prav On Wed, Mar 19, 2014 at 3:41 PM, sri harsha wrote: > Hi all, > is it

Re: Doubt

2014-03-19 Thread sri harsha
thank s jay and praveen, i want to use both separately don't want to use mongodb in the place of hbase On Wed, Mar 19, 2014 at 9:25 PM, Jay Vyas wrote: > Certainly it is , and quite common especially if you have some high > performance machines : they can run as mapreduce slaves and also doubl

Class loading in Hadoop and HBase

2014-03-19 Thread Amit Sela
Hi all, I'm running with Hadoop 1.0.4 and HBase 0.94.12 bundled (OSGi) versions I built. Most issues I encountered are related to class loaders. One of the patterns I noticed in both projects is: ClassLoader cl = Thread.currentThread().getContextClassLoader(); if(cl == null) { cl

Is Hadoop's TooRunner thread-safe?

2014-03-19 Thread Something Something
I would like to trigger a few Hadoop jobs simultaneously. I've created a pool of threads using Executors.newFixedThreadPool. Idea is that if the pool size is 2, my code will trigger 2 Hadoop jobs at the same exact time using 'ToolRunner.run'. In my testing, I noticed that these 2 threads keep st

Re: Is Hadoop's TooRunner thread-safe?

2014-03-19 Thread Something Something
Any thoughts on this? Confirm or Deny it's an issue.. may be? On Mon, Mar 17, 2014 at 11:43 AM, Something Something < mailinglist...@gmail.com> wrote: > I would like to trigger a few Hadoop jobs simultaneously. I've created a > pool of threads using Executors.newFixedThreadPool. Idea is that

JSR 203 NIO 2 for HDFS

2014-03-19 Thread Damien Carol
Hi, I'm working on minimal implementation of JSR 203 to provide access to HDFS (1.2.1) for a GUI tool needed in my company. Some features already works as create a directory, delete something, list files in directory. I would know if someone already worked on something like that. Maybe a F

Re: Doubt

2014-03-19 Thread Jay Vyas
Certainly it is , and quite common especially if you have some high performance machines : they can run as mapreduce slaves and also double as mongo hosts. The problem would of course be that when running mapreduce jobs you might have very slow network bandwidth at times, and if your front end ne

question about yarn webapp

2014-03-19 Thread 赵璨
I can get Server Stacks from web ui. But I don't know which code handle the function, how can the web app get the stacks information from jvm?

Doubt

2014-03-19 Thread sri harsha
Hi all, is it possible to install Mongodb on the same VM which consists hadoop? -- amiable harsha

RE: Webinar On Hadoop!

2014-03-19 Thread Clay McDonald
Really? What is Big Data? Really? I think you've spamed the wrong list From: Vivek Kumar [mailto:si...@soapt.com] Sent: Wednesday, March 19, 2014 8:09 AM To: user@hadoop.apache.org Subject: Webinar On Hadoop! Hi , Want to Know Big Data / Hadoop ? If yes , join us for Free Webinar by indus

Re: I am about to lose all my data please help

2014-03-19 Thread praveenesh kumar
Is this property correct ? fs.default.name -BLANKED Regards Prav On Wed, Mar 19, 2014 at 12:58 PM, Fatih Haltas wrote: > Thanks for you helps, but still could not solve my problem. > > > On Tue, Mar 18, 2014 at 10:13 AM, Stanley Shi wrote: > >> Ah yes, I overlooked this. Then ple

Re: I am about to lose all my data please help

2014-03-19 Thread Fatih Haltas
Thanks for you helps, but still could not solve my problem. On Tue, Mar 18, 2014 at 10:13 AM, Stanley Shi wrote: > Ah yes, I overlooked this. Then please check the file are there or not: > "ls /home/hadoop/project/hadoop-data/dfs/name"? > > Regards, > *Stanley Shi,* > > > > On Tue, Mar 18, 2014

Need FileName with Content

2014-03-19 Thread Ranjini Rathinam
Hi, I have folder named INPUT. Inside INPUT i have 5 resume are there. hduser@localhost:~/Ranjini$ hadoop fs -ls /user/hduser/INPUT Found 5 items -rw-r--r-- 1 hduser supergroup 5438 2014-03-18 15:20 /user/hduser/INPUT/Rakesh Chowdary_Microstrategy.txt -rw-r--r-- 1 hduser supergroup

Re: NodeHealthReport local-dirs turned bad

2014-03-19 Thread Margusja
tnx got it work. In my init script I used wrong user. It was permissions problem like Rohith said. Best regards, Margus (Margusja) Roo +372 51 48 780 http://margus.roo.ee http://ee.linkedin.com/in/margusroo skype: margusja ldapsearch -x -h ldap.sk.ee -b c=EE "(serialNumber=37303140314)" On 19/0

Webinar On Hadoop!

2014-03-19 Thread Vivek Kumar
Hi , Want to Know Big Data / Hadoop ? If yes , join us for Free Webinar by industry experts at below link. *FREE webinar on Hadoop, Hosted by : Manoj , Research Director* *Join us for a webinar on Mar 19, 2014 at 8:00 PM IST.* *Register now!* https://attendee.gotowebinar.com/register/5418099

RE: NodeHealthReport local-dirs turned bad

2014-03-19 Thread Rohith Sharma K S
Hi There is no relation to NameNode format Does NodeManger is started with default configuration? If no , any NodeManger health script is configured? Suspect can be 1. /hadoop does not have permission or 2. disk is full Thanks & Regards Rohith Sharma K S -Original Mess

NodeHealthReport local-dirs turned bad

2014-03-19 Thread Margusja
Hi I have one node in unhealthy status: Total Vmem allocated for Containers 4.20 GB Vmem enforcement enabledfalse Total Pmem allocated for Container 2 GB Pmem enforcement enabledfalse NodeHealthyStatus false LastNodeHealthTime Wed Mar 19 13:31:24 EE