Hello, everyone, I've downoaded Oozie from trunk and compiled with this command: bin/mkdistro.sh -e -Phadoop-2 -DskipTests -Dhadoop.version=2.7.1
When trying to create sharelib directory I'm getting this exception: bin/oozie-setup.sh sharelib create -fs hdfs://m1.local/ setting CATALINA_OPTS="$CATALINA_OPTS -Xmx1024m" Error: File /user/bbvoop/share/lib/lib_20150908180437/hive2/jersey-json-1.9.jar could only be replicated to 0 nodes instead of minReplication (=1). There are 3 datanode(s) running and 3 node(s) are excluded in this operation. at org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget4NewBlock(BlockManager.java:1550) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getNewBlockTargets(FSNamesystem.java:3110) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:3034) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:723) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:492) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:616) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:969) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2049) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2045) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:415) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2043) I can see the same error in NameNode's logs. The HDFS is working. If I try to upload some file with "hdfs dfs -put" everything works. In datanode logs I see this info message: 2015-09-08 18:09:52,905 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Failed to read expected SASL data transfer protection handshake from client at /10.104.18.113:59340. Perhaps the client is running an older version of Hadoop which does not support SASL data transfer protection So I guess I either have an incompatibility between my hadoop and my oozie libraries or oozie does not know that my hadoop setup is using ssl (and probably some other hadoop's configuration is missing). But I don't know where to look further? Could anyone help me resolve this? Thanks in advance and best regards.
