Julien, you should refer to HDP documentation when installing Ranger on HDP. Hortonworks provides HDP build in a different way, you can ask questions related to that in Hortonworks forums or through their support channel.
On Mon, Feb 9, 2015 at 6:14 AM, Julien Carme <[email protected]> wrote: > Thanks a lot for your help, I removed the libs and it solved the problem. > > Actually, I did not even know Ranger was (somehow) provided with HDP 2.2, > so I could not guess there could be a conflict. What is slightly puzzling > is that by default only ranger plugins are provided, neither the admin nor > usersync are, so I don't see how these installed plugins can be of any use. > > Regards, > > Julien > > > 2015-02-06 18:57 GMT+01:00 Ramesh Mani <[email protected]>: > >> Hi Julien, >> >> From the error message what I see is that the hdfs-plugin installed on >> the HDP2.2 is having a compatibility issue. You are installing a later >> version of plugin than then one is in the HDP2.2 . You should find the >> plugin in /usr/hdp/2.2.*/ranger-hdfs-plugin folder. >> >> Are you installing the ranger plugins from apache master branch of >> ranger? if you might need to remove files the old plugin jars (ranger*jar ) >> from /usr/hdp/2.2.*/hadoop/lib and replace it with jars that got installed >> when you did ./enable-hdfs-plugin.sh. >> Note:Name of the jar should be same as the one which is going to be >> removed. >> >> cd /usr/hdp/2.2.*/hadoop/lib >> ls -ltr ranger*jar. >> >> Please try this and let me know. >> Regards, >> Ramesh >> >> >> On Feb 6, 2015, at 7:15 AM, Julien Carme <[email protected]> wrote: >> >> Hello, >> >> I am trying to install Ranger on my new fresh HDP 2.2. >> Things went alright until I tried to start ranger-hdfs-plugin. >> >> When I run ./enable-hdfs-plugin.sh and restart hdfs, nothing seem to work >> correctly. I don't see anything in "Ranger Admin Web interface -> Audit -> >> Agents", I can't create HDFS repositories in the web interface, and a >> simple 'hadoop fs -ls /' throws a NullPointerException. >> >> >> In my namenode logs, I found what seems to be the explanation: >> >> Execution of 'su -s /bin/bash - hdfs -c 'export >> PATH=$PATH:/usr/hdp/current/hadoop-client/bin ; hdfs --config >> /etc/hadoop/conf dfsadmin -safemode get' | grep 'Safe mode is OFF'' >> returned 1. >> >> 2015-02-06 15:33:04,761 ERROR namenode.XaSecureFSPermissionChecker >> (XaSecureFSPermissionChecker.java:AuthorizeAccessForUser(140)) - Unable to >> create Authorizer >> java.lang.ClassCastException: org.apache.ranger.pdp.hdfs.RangerAuthorizer >> cannot be cast to com.xasecure.authorization.hadoop.HDFSAccessVerifier >> at >> com.xasecure.authorization.hadoop.HDFSAccessVerifierFactory.getInstance(HDFSAccessVerifierFactory.java:43) >> at >> org.apache.hadoop.hdfs.server.namenode.XaSecureFSPermissionChecker.AuthorizeAccessForUser(XaSecureFSPermissionChecker.java:137) >> ... >> ... >> same exception severla time and then: >> >> 2015-02-06 15:33:04,766 WARN ipc.Server (Server.java:run(2058)) - IPC >> Server handler 23 on 8020, call >> org.apache.hadoop.hdfs.protocol.ClientProtocol.getListing from >> 10.34.62.32:60489 Call#32650 Retry#0 >> java.lang.NullPointerException >> at >> org.apache.hadoop.hdfs.server.namenode.XaSecureFSPermissionChecker.writeLog(XaSecureFSPermissionChecker.java:214) >> at >> org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java:193) >> ... >> >> >> Could you tell me how to solve this problem? >> >> Regards, >> >> >> >> CONFIDENTIALITY NOTICE >> NOTICE: This message is intended for the use of the individual or entity >> to which it is addressed and may contain information that is confidential, >> privileged and exempt from disclosure under applicable law. If the reader >> of this message is not the intended recipient, you are hereby notified that >> any printing, copying, dissemination, distribution, disclosure or >> forwarding of this communication is strictly prohibited. If you have >> received this communication in error, please contact the sender immediately >> and delete it from your system. Thank You. > > >
