i reslove this by make alex directory under staging directory and set the
owner to alex
On Thu, Jul 24, 2014 at 10:11 PM, java8964 wrote:
> Are you sure user 'Alex' belongs to 'hadoop' group? Why not your run
> command 'id alex' to prove it? And 'Alex' belongs to 'hadoop' group can be
> confirm
2014-07-24 17:33:04,783 WARN
org.apache.hadoop.security.UserGroupInformation: PriviledgedActionException
as:hdfs (auth:SIMPLE) cause:org.apache.hadoop.ipc.StandbyException:
Operation category READ is not supported in state standby
2014-07-24 17:33:05,742 WARN
org.apache.hadoop.security.UserGroupInf
The "alex" should belongs to "hadoop" group on namenode;
Regards,
*Stanley Shi,*
On Thu, Jul 24, 2014 at 10:11 PM, java8964 wrote:
> Are you sure user 'Alex' belongs to 'hadoop' group? Why not your run
> command 'id alex' to prove it? And 'Alex' belongs to 'hadoop' group can be
> confirmed on
Would you please also past the corresponding namenode log?
Regards,
*Stanley Shi,*
On Fri, Jul 25, 2014 at 9:15 AM, ch huang wrote:
> hi,maillist:
>i try to copy data from my old cluster to new cluster,i get
> error ,how to handle this?
>
> 14/07/24 18:35:58 INFO mapreduce.Job: Ta
hi,maillist:
i try to copy data from my old cluster to new cluster,i get
error ,how to handle this?
14/07/24 18:35:58 INFO mapreduce.Job: Task Id :
attempt_1406182801379_0004_m_00_1, Status : FAILED
Error: java.io.IOException: File copy failed:
webhdfs://CH22:50070/mytest/pipe_url_b
Hi,
Inline.
On Fri, Jul 25, 2014 at 2:55 AM, Arjun Bakshi wrote:
> Hi,
>
> Thanks for the reply. It cleared up a few things.
>
> I hadn't thought of situations of under-replication, but I'll give it some
> thought now. It should be easier since, as you've mentioned, by that time
> the namenode k
Hi,
Thanks for the reply. It cleared up a few things.
I hadn't thought of situations of under-replication, but I'll give it
some thought now. It should be easier since, as you've mentioned, by
that time the namenode knows all the blocks that came from the same file
as the under-replicated blo
+Hbase User
On Wed, Jul 23, 2014 at 11:04 PM, Shankar hiremath <
shankar.hirem...@huawei.com> wrote:
> Dear All,
>
> Observation:
> I have HBase cluster with Kerberos enabled, when the Region Server
> startups then we observed some 20-30 socket fds are in CLOSE_WAIT state,
> After that when the
Which version of Hadoop?
Yes saveNamespace as you described will checkpoint the FsImage and reset
your edits log so it will reduce startup time.
On Thu, Jul 24, 2014 at 8:25 AM, Anfernee Xu wrote:
> Yes, I have secondary NM, but without HA, but after I killed NM and
> secondary NM, the startup
Hello,
(Inline)
On Thu, Jul 24, 2014 at 11:11 PM, Arjun Bakshi wrote:
> Hi,
>
> I want to write a block placement policy that takes the size of the file
> being placed into account. Something like what is done in CoHadoop or BEEMR
> paper. I have the following questions:
>
> 1- What is srcPath i
Hi,
I want to write a block placement policy that takes the size of the file
being placed into account. Something like what is done in CoHadoop or
BEEMR paper. I have the following questions:
1- What is srcPath in chooseTarget? Is it the path to the original
un-chunked file, or it is a path
Hello Wellington,
That sounds wonderful! I appreciate everyone's help.
Best Regards,
Andrew Touchet
On Thu, Jul 24, 2014 at 12:01 PM, Wellington Chevreuil <
wellington.chevre...@gmail.com> wrote:
> You should not face any data loss. The replicas were just moved away from
> that node to other
Hi Mirko,
Thanks for the reply!
"...it will not bring in exactly the same blocks like before"
Is that what usually happens when adding nodes back in? Should I expect any
data loss due to starting the data node process before running the
balancing tool?
Best Regards,
Andrew Touchet
On Thu, Ju
You should not face any data loss. The replicas were just moved away from that
node to other nodes in the cluster during decommission. Once you recommission
the node and re-balance your cluster, HDFS will re-distribute replicas between
the nodes evenly, and the recommissioned node will receive r
After you added the nodes back to your cluster you run the balancer tool,
but it will not bring in exactly the same blocks like before.
Cheers,
Mirko
2014-07-24 17:34 GMT+01:00 andrew touchet :
> Thanks for the reply,
>
> I am using Hadoop-0.20. We installed from Apache not cloundera, if that
Thanks for the reply,
I am using Hadoop-0.20. We installed from Apache not cloundera, if that
makes a difference.
Currently I really need to know how to get the data that was replicated
during decommissioning back onto my two data nodes.
On Thursday, July 24, 2014, Stanley Shi wrote:
> whi
Would you please change the log level to DEBUG to see what happens when
creating the client protocol provider?
On Thu, Jul 24, 2014 at 2:13 AM, MrAsanjar . wrote:
> please help;
> i have verified mapre-site.xml => mapreduce.framework.name=yarn
> verified HADOOP_CLASS_PATH in hadoop-env.sh
>
> /
Yes, I have secondary NM, but without HA, but after I killed NM and
secondary NM, the startup time was still too long(4 hours), what else
should I do? Will "hadoop dfsadmin -saveNamespace" resolve the issue?
Thanks for your help.
On Wed, Jul 23, 2014 at 10:58 PM, Stanley Shi wrote:
> Do you h
Are you sure user 'Alex' belongs to 'hadoop' group? Why not your run command
'id alex' to prove it? And 'Alex' belongs to 'hadoop' group can be confirmed on
the namenode?
Yong
Date: Thu, 24 Jul 2014 17:11:06 +0800
Subject: issue about run MR job use system user
From: justlo...@gmail.com
To: user
hi,maillist:
i create a system user on a box of my hadoop cluster ,but when i
run MR job use this user ,it get a problem, the /data directory is for
mapreduce history server option, and i also add the user into hadoop group
,since the /data privilege is 775 ,so it can write by user in had
20 matches
Mail list logo