Re: issue about run MR job use system user

2014-07-24 Thread ch huang
i reslove this by make alex directory under staging directory and set the owner to alex On Thu, Jul 24, 2014 at 10:11 PM, java8964 wrote: > Are you sure user 'Alex' belongs to 'hadoop' group? Why not your run > command 'id alex' to prove it? And 'Alex' belongs to 'hadoop' group can be > confirm

Re: issue about distcp " Source and target differ in block-size. Use -pb to preserve block-sizes during copy."

2014-07-24 Thread ch huang
2014-07-24 17:33:04,783 WARN org.apache.hadoop.security.UserGroupInformation: PriviledgedActionException as:hdfs (auth:SIMPLE) cause:org.apache.hadoop.ipc.StandbyException: Operation category READ is not supported in state standby 2014-07-24 17:33:05,742 WARN org.apache.hadoop.security.UserGroupInf

Re: issue about run MR job use system user

2014-07-24 Thread Stanley Shi
The "alex" should belongs to "hadoop" group on namenode; Regards, *Stanley Shi,* On Thu, Jul 24, 2014 at 10:11 PM, java8964 wrote: > Are you sure user 'Alex' belongs to 'hadoop' group? Why not your run > command 'id alex' to prove it? And 'Alex' belongs to 'hadoop' group can be > confirmed on

Re: issue about distcp " Source and target differ in block-size. Use -pb to preserve block-sizes during copy."

2014-07-24 Thread Stanley Shi
Would you please also past the corresponding namenode log? Regards, *Stanley Shi,* On Fri, Jul 25, 2014 at 9:15 AM, ch huang wrote: > hi,maillist: >i try to copy data from my old cluster to new cluster,i get > error ,how to handle this? > > 14/07/24 18:35:58 INFO mapreduce.Job: Ta

issue about distcp " Source and target differ in block-size. Use -pb to preserve block-sizes during copy."

2014-07-24 Thread ch huang
hi,maillist: i try to copy data from my old cluster to new cluster,i get error ,how to handle this? 14/07/24 18:35:58 INFO mapreduce.Job: Task Id : attempt_1406182801379_0004_m_00_1, Status : FAILED Error: java.io.IOException: File copy failed: webhdfs://CH22:50070/mytest/pipe_url_b

Re: Building custom block placement policy. What is srcPath?

2014-07-24 Thread Harsh J
Hi, Inline. On Fri, Jul 25, 2014 at 2:55 AM, Arjun Bakshi wrote: > Hi, > > Thanks for the reply. It cleared up a few things. > > I hadn't thought of situations of under-replication, but I'll give it some > thought now. It should be easier since, as you've mentioned, by that time > the namenode k

Re: Building custom block placement policy. What is srcPath?

2014-07-24 Thread Arjun Bakshi
Hi, Thanks for the reply. It cleared up a few things. I hadn't thought of situations of under-replication, but I'll give it some thought now. It should be easier since, as you've mentioned, by that time the namenode knows all the blocks that came from the same file as the under-replicated blo

Re: RegionServer many socket fds are in CLOSE_WAIT and not getting cleared

2014-07-24 Thread Arpit Agarwal
+Hbase User On Wed, Jul 23, 2014 at 11:04 PM, Shankar hiremath < shankar.hirem...@huawei.com> wrote: > Dear All, > > Observation: > I have HBase cluster with Kerberos enabled, when the Region Server > startups then we observed some 20-30 socket fds are in CLOSE_WAIT state, > After that when the

Re: how to reduce delay in HDFS restart

2014-07-24 Thread Arpit Agarwal
Which version of Hadoop? Yes saveNamespace as you described will checkpoint the FsImage and reset your edits log so it will reduce startup time. On Thu, Jul 24, 2014 at 8:25 AM, Anfernee Xu wrote: > Yes, I have secondary NM, but without HA, but after I killed NM and > secondary NM, the startup

Re: Building custom block placement policy. What is srcPath?

2014-07-24 Thread Harsh J
Hello, (Inline) On Thu, Jul 24, 2014 at 11:11 PM, Arjun Bakshi wrote: > Hi, > > I want to write a block placement policy that takes the size of the file > being placed into account. Something like what is done in CoHadoop or BEEMR > paper. I have the following questions: > > 1- What is srcPath i

Building custom block placement policy. What is srcPath?

2014-07-24 Thread Arjun Bakshi
Hi, I want to write a block placement policy that takes the size of the file being placed into account. Something like what is done in CoHadoop or BEEMR paper. I have the following questions: 1- What is srcPath in chooseTarget? Is it the path to the original un-chunked file, or it is a path

Re: Decommissioning a data node and problems bringing it back online

2014-07-24 Thread andrew touchet
Hello Wellington, That sounds wonderful! I appreciate everyone's help. Best Regards, Andrew Touchet On Thu, Jul 24, 2014 at 12:01 PM, Wellington Chevreuil < wellington.chevre...@gmail.com> wrote: > You should not face any data loss. The replicas were just moved away from > that node to other

Re: Decommissioning a data node and problems bringing it back online

2014-07-24 Thread andrew touchet
Hi Mirko, Thanks for the reply! "...it will not bring in exactly the same blocks like before" Is that what usually happens when adding nodes back in? Should I expect any data loss due to starting the data node process before running the balancing tool? Best Regards, Andrew Touchet On Thu, Ju

Re: Decommissioning a data node and problems bringing it back online

2014-07-24 Thread Wellington Chevreuil
You should not face any data loss. The replicas were just moved away from that node to other nodes in the cluster during decommission. Once you recommission the node and re-balance your cluster, HDFS will re-distribute replicas between the nodes evenly, and the recommissioned node will receive r

Re: Decommissioning a data node and problems bringing it back online

2014-07-24 Thread Mirko Kämpf
After you added the nodes back to your cluster you run the balancer tool, but it will not bring in exactly the same blocks like before. Cheers, Mirko 2014-07-24 17:34 GMT+01:00 andrew touchet : > Thanks for the reply, > > I am using Hadoop-0.20. We installed from Apache not cloundera, if that

Re: Decommissioning a data node and problems bringing it back online

2014-07-24 Thread andrew touchet
Thanks for the reply, I am using Hadoop-0.20. We installed from Apache not cloundera, if that makes a difference. Currently I really need to know how to get the data that was replicated during decommissioning back onto my two data nodes. On Thursday, July 24, 2014, Stanley Shi wrote: > whi

Re: HDP hadoop 2.4.1 fails to run mapreduce app

2014-07-24 Thread Zhijie Shen
Would you please change the log level to DEBUG to see what happens when creating the client protocol provider? On Thu, Jul 24, 2014 at 2:13 AM, MrAsanjar . wrote: > please help; > i have verified mapre-site.xml => mapreduce.framework.name=yarn > verified HADOOP_CLASS_PATH in hadoop-env.sh > > /

Re: how to reduce delay in HDFS restart

2014-07-24 Thread Anfernee Xu
Yes, I have secondary NM, but without HA, but after I killed NM and secondary NM, the startup time was still too long(4 hours), what else should I do? Will "hadoop dfsadmin -saveNamespace" resolve the issue? Thanks for your help. On Wed, Jul 23, 2014 at 10:58 PM, Stanley Shi wrote: > Do you h

RE: issue about run MR job use system user

2014-07-24 Thread java8964
Are you sure user 'Alex' belongs to 'hadoop' group? Why not your run command 'id alex' to prove it? And 'Alex' belongs to 'hadoop' group can be confirmed on the namenode? Yong Date: Thu, 24 Jul 2014 17:11:06 +0800 Subject: issue about run MR job use system user From: justlo...@gmail.com To: user

issue about run MR job use system user

2014-07-24 Thread ch huang
hi,maillist: i create a system user on a box of my hadoop cluster ,but when i run MR job use this user ,it get a problem, the /data directory is for mapreduce history server option, and i also add the user into hadoop group ,since the /data privilege is 775 ,so it can write by user in had