Re: ResourceManager fails to start

2015-07-07 Thread Sanjeev Tripurari
Hi Alexandru

Can you share whats value in capacity scheduler for

yarn.scheduler.capacity.am.failure.scheduling.delay.ms


Regards
-Sanjeev


On Fri, Jun 26, 2015 at 6:40 PM, Alexandru Pacurar 
alexandru.pacu...@propertyshark.com wrote:

  Hello,



 I’m running Hadoop 2.6 and I have encountered a problem with the
 resourcemanager. After a restart the resourcemanager refuses to start with
 the following error:



 2015-06-26 08:54:10,342 INFO  attempt.RMAppAttemptImpl
 (RMAppAttemptImpl.java:recover(796)) - Recovering attempt:
 appattempt_1435159945366_0792_01 with final state: null

 2015-06-26 08:54:10,342 INFO  security.AMRMTokenSecretManager
 (AMRMTokenSecretManager.java:createAndGetAMRMToken(195)) - Create AMRMToken
 for ApplicationAttempt: appattempt_1435159945366_0792_01

 2015-06-26 08:54:10,342 INFO  security.AMRMTokenSecretManager
 (AMRMTokenSecretManager.java:createPassword(307)) - Creating password for
 appattempt_1435159945366_0792_01

 2015-06-26 08:54:10,343 INFO  resourcemanager.ApplicationMasterService
 (ApplicationMasterService.java:registerAppAttempt(670)) - Registering app
 attempt : appattempt_1435159945366_0792_01

 2015-06-26 08:54:10,344 ERROR resourcemanager.ResourceManager
 (ResourceManager.java:serviceStart(594)) - Failed to load/recover state

 java.lang.NullPointerException

 at
 org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.CapacityScheduler.addApplicationAttempt(CapacityScheduler.java:734)

 at
 org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.CapacityScheduler.handle(CapacityScheduler.java:1089)

 at
 org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.CapacityScheduler.handle(CapacityScheduler.java:114)

 at
 org.apache.hadoop.yarn.server.resourcemanager.rmapp.attempt.RMAppAttemptImpl$AttemptRecoveredTransition.transition(RMAppAttemptImpl.java:1038)

 at
 org.apache.hadoop.yarn.server.resourcemanager.rmapp.attempt.RMAppAttemptImpl$AttemptRecoveredTransition.transition(RMAppAttemptImpl.java:1002)

 at
 org.apache.hadoop.yarn.state.StateMachineFactory$MultipleInternalArc.doTransition(StateMachineFactory.java:385)

 at
 org.apache.hadoop.yarn.state.StateMachineFactory.doTransition(StateMachineFactory.java:302)

 at
 org.apache.hadoop.yarn.state.StateMachineFactory.access$300(StateMachineFactory.java:46)

 at
 org.apache.hadoop.yarn.state.StateMachineFactory$InternalStateMachine.doTransition(StateMachineFactory.java:448)

 at
 org.apache.hadoop.yarn.server.resourcemanager.rmapp.attempt.RMAppAttemptImpl.handle(RMAppAttemptImpl.java:755)

 at
 org.apache.hadoop.yarn.server.resourcemanager.rmapp.attempt.RMAppAttemptImpl.handle(RMAppAttemptImpl.java:106)

 at
 org.apache.hadoop.yarn.server.resourcemanager.rmapp.RMAppImpl.recoverAppAttempts(RMAppImpl.java:831)

 at
 org.apache.hadoop.yarn.server.resourcemanager.rmapp.RMAppImpl.access$1900(RMAppImpl.java:101)

 at
 org.apache.hadoop.yarn.server.resourcemanager.rmapp.RMAppImpl$RMAppRecoveredTransition.transition(RMAppImpl.java:846)

 at
 org.apache.hadoop.yarn.server.resourcemanager.rmapp.RMAppImpl$RMAppRecoveredTransition.transition(RMAppImpl.java:836)

 at
 org.apache.hadoop.yarn.state.StateMachineFactory$MultipleInternalArc.doTransition(StateMachineFactory.java:385)

 at
 org.apache.hadoop.yarn.state.StateMachineFactory.doTransition(StateMachineFactory.java:302)

 at
 org.apache.hadoop.yarn.state.StateMachineFactory.access$300(StateMachineFactory.java:46)

 at
 org.apache.hadoop.yarn.state.StateMachineFactory$InternalStateMachine.doTransition(StateMachineFactory.java:448)

 at
 org.apache.hadoop.yarn.server.resourcemanager.rmapp.RMAppImpl.handle(RMAppImpl.java:711)

 at
 org.apache.hadoop.yarn.server.resourcemanager.RMAppManager.recoverApplication(RMAppManager.java:312)

 at
 org.apache.hadoop.yarn.server.resourcemanager.RMAppManager.recover(RMAppManager.java:413)

 at
 org.apache.hadoop.yarn.server.resourcemanager.ResourceManager.recover(ResourceManager.java:1207)

 at
 org.apache.hadoop.yarn.server.resourcemanager.ResourceManager$RMActiveServices.serviceStart(ResourceManager.java:590)

 at
 org.apache.hadoop.service.AbstractService.start(AbstractService.java:193)

 at
 org.apache.hadoop.yarn.server.resourcemanager.ResourceManager.startActiveServices(ResourceManager.java:1014)

 at
 org.apache.hadoop.yarn.server.resourcemanager.ResourceManager$1.run(ResourceManager.java:1051)

 at
 org.apache.hadoop.yarn.server.resourcemanager.ResourceManager$1.run(ResourceManager.java:1047)

 at 

Grant privileges on / of hdfs file system

2015-07-07 Thread Pratik Gadiya
Hi,

Can anyone provide me steps to grant a particular user access equivalnent to 
that of hdfs user in hadoop.
The reason behind this is I want to have my custom user who can create anything 
on the entire hdfs file system (/)
I tried couple of links however, none of them were useful.
Is there any way by adding/modifying some property tags I can do that ?

With Regards,
Pratik

DISCLAIMER
==
This e-mail may contain privileged and confidential information which is the 
property of Persistent Systems Ltd. It is intended only for the use of the 
individual or entity to which it is addressed. If you are not the intended 
recipient, you are not authorized to read, retain, copy, print, distribute or 
use this message. If you have received this communication in error, please 
notify the sender and delete all copies of this message. Persistent Systems 
Ltd. does not accept any liability for virus infected mails.



Re: Subscribe

2015-07-07 Thread Ted Yu
Take a look at http://hadoop.apache.org/mailing_lists.html

Cheers

On Tue, Jul 7, 2015 at 1:09 AM, Akmal Abbasov akmal.abba...@icloud.com
wrote:





Re: get the hadoop working dir using the bash commands?

2015-07-07 Thread Chris Nauroth
Hello,

Are you looking for the working directory used by HDFS to resolve relative
paths used in commands like hdfs dfs -ls myRelativePath?  If so, then
the working directory is the current user's home directory.  HDFS defines
the home directory as a common prefix followed by the username.  The
prefix is controlled by configuration property dfs.user.home.dir.prefix,
and the default is /user.

property
  namedfs.user.home.dir.prefix/name
  value/user/value
  descriptionThe directory to prepend to user name to get the user's
  home direcotry.
  /description
/property

You mentioned wanting to access this in bash.  I think you can combine
hdfs getconf with the current user to get it.  For example:


 echo $(hdfs getconf -confKey dfs.user.home.dir.prefix)/$USER
/user/chris

If you're running a secured cluster with complex auth-to-local name
conversion rules, then simply using $USER might not be sufficient.  If you
find that's the case, then look at using the hadoop kerbname your
Kerberos principal command.  The hadoop kerbname command alias only
exists in trunk right now.  For 2.x builds, you can still get the same
effect by running the underlying class directly: hadoop
org.apache.hadoop.security.HadoopKerberosName your Kerberos principal.


I hope this helps.

--Chris Nauroth




On 7/7/15, 2:33 AM, xeonmailinglist xeonmailingl...@gmail.com wrote:

Is it possible to get the hadoop working dir using the bash commands?



Re: Large number of application attempts

2015-07-07 Thread Xuan Gong
Hey, Akmal:
Are you running a lots of apps ? Or could you, maybe, share some of
file names of those znodes, please ?


Thanks

Xuan Gong
On 7/7/15, 1:23 AM, Akmal Abbasov akmal.abba...@icloud.com wrote:

Hi all,
I have HBase cluster.
I run a cron job each hour to create a snapshot of tables in HBase and
export these snapshots to another HBase cluster for backup.
But I¹ve noticed that there are ~4000 znodes in
/hadoop-ha/testhbase1/rmstore/ZKRMStateRoot/RMAppRoot in Zookeeper.
What are these znodes?
Are they failed yarn applications?
How I need to handle them?

Thank you.
Akmal



Subscribe

2015-07-07 Thread Akmal Abbasov



Large number of application attempts

2015-07-07 Thread Akmal Abbasov
Hi all,
I have HBase cluster. 
I run a cron job each hour to create a snapshot of tables in HBase and export 
these snapshots to another HBase cluster for backup. 
But I’ve noticed that there are ~4000 znodes in 
/hadoop-ha/testhbase1/rmstore/ZKRMStateRoot/RMAppRoot in Zookeeper.
What are these znodes? 
Are they failed yarn applications?
How I need to handle them?

Thank you.
Akmal

Re: Grant privileges on / of hdfs file system

2015-07-07 Thread Chris Nauroth
Hello Pratik,

The hdfs user (more specifically, whatever user launched the NameNode 
process) is the HDFS super-user.  The super-user has full access to the file 
system and also administrative operations.  You can declare additional users to 
be super-users by setting property dfs.permissions.superusergroup in 
hdfs-site.xml.  The default value of this property is supergroup.

property
  namedfs.permissions.superusergroup/name
  valuesupergroup/value
  descriptionThe name of the group of super-users./description
/property

Any user you add to group supergroup (or whatever custom group you use if you 
decided to change dfs.permissions.superusergroup) will be treated as an HDFS 
super-user.

It's important to keep in mind that this grants both full file system access 
and full administrative access.  That means the user would be able to call 
sensitive operations like hdfs dfsadmin -safemode enter.  If this isn't 
appropriate, then you might explore using file system permissions and ACLs to 
implement your requirements on the file system only.

More details are in the documentation here:

http://hadoop.apache.org/docs/r2.7.0/hadoop-project-dist/hadoop-hdfs/HdfsPermissionsGuide.html

I hope this helps.

--Chris Nauroth

From: Pratik Gadiya 
pratik_gad...@persistent.commailto:pratik_gad...@persistent.com
Reply-To: user@hadoop.apache.orgmailto:user@hadoop.apache.org 
user@hadoop.apache.orgmailto:user@hadoop.apache.org
Date: Tuesday, July 7, 2015 at 9:38 AM
To: user@hadoop.apache.orgmailto:user@hadoop.apache.org 
user@hadoop.apache.orgmailto:user@hadoop.apache.org
Subject: Grant privileges on / of hdfs file system

Hi,

Can anyone provide me steps to grant a particular user access equivalnent to 
that of hdfs user in hadoop.
The reason behind this is I want to have my custom user who can create anything 
on the entire hdfs file system (/)
I tried couple of links however, none of them were useful.
Is there any way by adding/modifying some property tags I can do that ?

With Regards,
Pratik

DISCLAIMER == This e-mail may contain privileged and confidential 
information which is the property of Persistent Systems Ltd. It is intended 
only for the use of the individual or entity to which it is addressed. If you 
are not the intended recipient, you are not authorized to read, retain, copy, 
print, distribute or use this message. If you have received this communication 
in error, please notify the sender and delete all copies of this message. 
Persistent Systems Ltd. does not accept any liability for virus infected mails.


Re: Problem when configure the security in hadoop

2015-07-07 Thread Zhijie Shen
?Not sure about HDFS special setup, but in general, to use HTTPs, you should 
have your keystore/truststore generated and config ssl-client.xml and 
ssl-server.xml properly.


- Zhijie


From: Colin Ma colin...@gmail.com
Sent: Friday, July 03, 2015 1:37 AM
To: user@hadoop.apache.org
Subject: Problem when configure the security in hadoop

Hi,
 I do the security configuration for Hadoop these days, the Kerberos 
works fine, but there maybe has some problems on sasl configuration.
 The following is the related configuration in hdfs-site.xml:
property
  namedfs.http.policy/name
  valueHTTPS_ONLY/value
/property
property
  namedfs.data.transfer.protection/name
  valueauthentication/value
/property

There is no problem to execute the command like:   hdfs dfs -ls /
But when I execute the command:   hdfs dfs -copyToLocal /temp/test.txt .  The 
following exception will be thrown:

 015-07-03 14:02:54,715 INFO 
org.apache.hadoop.hdfs.server.datanode.DataBlockScanner: Added 
bpid=BP-271423801-192.168.20.28-1423724265164 to blockPoolScannerMap, new size=1
2015-07-03 14:03:39,963 ERROR org.apache.hadoop.hdfs.server.datanode.DataNode: 
server-511:50010:DataXceiver error processing unknown operation  src: 
/192.168.20.28:58422http://192.168.20.28:58422 dst: 
/192.168.20.28:50010http://192.168.20.28:50010
java.io.EOFException: Premature EOF: no length prefix available
 at 
org.apache.hadoop.hdfs.protocolPB.PBHelper.vintPrefixed(PBHelper.java:2203)
 at 
org.apache.hadoop.hdfs.protocol.datatransfer.sasl.DataTransferSaslUtil.readSaslMessageAndNegotiationCipherOptions(DataTransferSaslUtil.java:233)
 at 
org.apache.hadoop.hdfs.protocol.datatransfer.sasl.SaslDataTransferServer.doSaslHandshake(SaslDataTransferServer.java:369)
 at 
org.apache.hadoop.hdfs.protocol.datatransfer.sasl.SaslDataTransferServer.getSaslStreams(SaslDataTransferServer.java:297)
 at 
org.apache.hadoop.hdfs.protocol.datatransfer.sasl.SaslDataTransferServer.receive(SaslDataTransferServer.java:124)
 at 
org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:183)
 at java.lang.Thread.run(Thread.java:745)
2015-07-03 15:34:39,917 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: 
Sent 1 blockreports 145 blocks total. Took 1 msec to generate and 6 msecs for 
RPC and NN processing.  Got back commands 
org.apache.hadoop.hdfs.server.protocol.FinalizeCommand@1b3bce82mailto:org.apache.hadoop.hdfs.server.protocol.FinalizeCommand@1b3bce82

 Just take a look the method doSaslHandshake() of 
SaslDataTransferClient.java and SaslDataTransferServer.java, maybe 
SaslDataTransferClient send a empty response cause this exception, and I think 
some mistakes in the configuration caused this problem.
Is there anyone can help to check this problem?
Thanks for your help.

Best regards,

Colin Ma


return tuple of key value pair from string in java UDF

2015-07-07 Thread Divya Gehlot
Hi,
I have a pig script which return field as
 Key=ValKey=ValKey=ValKey=ValKey=Val whose schema  is defined as
data:chararray in my pig script
Can I use java UDF to return a tuple of Key Value Pair.

Would really appreciate if somebody help me point to some guide or example
of this kind of scenario.


Thanks


unsubscribe

2015-07-07 Thread Will Bell
unsubscribe

Best regards

Will.
 
William J Bell
Sales Director – www.wandisco.com
+44 7825 770477 - will.b...@wandisco.com
 


Believed technically impossible, WANdisco customers now exploit 100% uptime and 
exact real-time replication across widely distributed Hadoop, Subversion and 
Git.


On 7 Jul 2015, at 10:44, Divya Gehlot divya.htco...@gmail.com wrote:

Hi,
I have a pig script which return field as 
 Key=ValKey=ValKey=ValKey=ValKey=Val whose schema  is defined as 
data:chararray in my pig script 
Can I use java UDF to return a tuple of Key Value Pair. 

Would really appreciate if somebody help me point to some guide or example of 
this kind of scenario.


Thanks



-- 


Learn how WANdisco Fusion solves Hadoop data protection and scalability 
challenges http://www.wandisco.com/hadoop/wd-fusion

Listed on the London Stock Exchange: WAND 
http://www.bloomberg.com/quote/WAND:LN

THIS MESSAGE AND ANY ATTACHMENTS ARE CONFIDENTIAL, PROPRIETARY, AND MAY BE 
PRIVILEGED.  If this message was misdirected, WANdisco, Inc. and its 
subsidiaries, (WANdisco) does not waive any confidentiality or privilege. 
 If you are not the intended recipient, please notify us immediately and 
destroy the message without disclosing its contents to anyone.  Any 
distribution, use or copying of this e-mail or the information it contains 
by other than an intended recipient is unauthorized.  The views and 
opinions expressed in this e-mail message are the author's own and may not 
reflect the views and opinions of WANdisco, unless the author is authorized 
by WANdisco to express such views or opinions on its behalf.  All email 
sent to or from this address is subject to electronic storage and review by 
WANdisco.  Although WANdisco operates anti-virus programs, it does not 
accept responsibility for any damage whatsoever caused by viruses being 
passed.


get the hadoop working dir using the bash commands?

2015-07-07 Thread xeonmailinglist

Is it possible to get the hadoop working dir using the bash commands?