[ https://issues.apache.org/jira/browse/FALCON-2046?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15359241#comment-15359241 ]
ASF GitHub Bot commented on FALCON-2046: ---------------------------------------- Github user asfgit closed the pull request at: https://github.com/apache/falcon/pull/208 > HDFS Replication failing in secure Mode > --------------------------------------- > > Key: FALCON-2046 > URL: https://issues.apache.org/jira/browse/FALCON-2046 > Project: Falcon > Issue Type: Bug > Components: replication > Affects Versions: 0.10 > Reporter: Murali Ramasami > Assignee: Sowmya Ramesh > Priority: Critical > Fix For: trunk, 0.10 > > > HDFS Replication failing in secure Mode with the Authentication required > error. > Scenario: > HDFS replication from single source to single target > Extension property file > {noformat} > [hrt_qa@nat-os-r6-upns-falcon-multicluster-14 hadoopqe]$ cat > /tmp/falcon-extension/HdfsExtensionTesta1b85962.properties > # > # Licensed to the Apache Software Foundation (ASF) under one > # or more contributor license agreements. See the NOTICE file > # distributed with this work for additional information > # regarding copyright ownership. The ASF licenses this file > # to you under the Apache License, Version 2.0 (the > # "License"); you may not use this file except in compliance > # with the License. You may obtain a copy of the License at > # > # http://www.apache.org/licenses/LICENSE-2.0 > # > # Unless required by applicable law or agreed to in writing, software > # distributed under the License is distributed on an "AS IS" BASIS, > # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. > # See the License for the specific language governing permissions and > # limitations under the License. > # > ##### NOTE: This is a TEMPLATE file which can be copied and edited > jobName = HdfsExtensionTesta1b85962 > jobClusterName = Aa39cd108-b6c6e9ff > jobValidityStart = 2016-06-21T04:27Z > jobValidityEnd = 2016-06-21T04:47Z > jobFrequency = days(1) > sourceCluster = Aa39cd108-b6c6e9ff > sourceDir = /tmp/falcon-regression/HdfsExtensionTest/HdfsDR/source > targetCluster = Aa39cd108-38a7a9cc > targetDir = /tmp/falcon-regression/HdfsExtensionTest/HdfsDR/target > jobAclOwner = hrt_qa > jobAclGroup = users > jobAclPermission = * > extensionName = hdfs-mirroring > jobProcessFrequency = minutes(5) > {noformat} > Please see the application log below: > {noformat} > ================================================================= > >>> Invoking Main class now >>> > Fetching child yarn jobs > tag id : oozie-62d207ec7d2c61db9dd3220d0fda7c22 > Child yarn jobs are found - > Main class : org.apache.falcon.replication.FeedReplicator > Arguments : > -Dmapred.job.queue.name=default > -Dmapred.job.priority=NORMAL > -maxMaps > 1 > -mapBandwidth > 100 > -sourcePaths > > webhdfs://nat-os-r6-upns-falcon-multicluster-14.openstacklocal:20070/tmp/falcon-regression/HdfsExtensionTest/HdfsDR/source > -targetPath > > hdfs://nat-os-r6-upns-falcon-multicluster-10.openstacklocal:8020/tmp/falcon-regression/HdfsExtensionTest/HdfsDR/target > -falconFeedStorageType > FILESYSTEM > -availabilityFlag > NA > -counterLogDir > > hdfs://nat-os-r6-upns-falcon-multicluster-14.openstacklocal:8020/tmp/fs/falcon/workflows/process/HdfsExtensionTesta1b85962/logs/job-2016-06-21-04-27/ > <<< Invocation of Main class completed <<< > Failing Oozie Launcher, Main class [org.apache.oozie.action.hadoop.JavaMain], > main() threw exception, org.apache.hadoop.security.AccessControlException: > Authentication required > org.apache.oozie.action.hadoop.JavaMainException: > org.apache.hadoop.security.AccessControlException: Authentication required > at org.apache.oozie.action.hadoop.JavaMain.run(JavaMain.java:59) > at > org.apache.oozie.action.hadoop.LauncherMain.run(LauncherMain.java:51) > at org.apache.oozie.action.hadoop.JavaMain.main(JavaMain.java:35) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:498) > at > org.apache.oozie.action.hadoop.LauncherMapper.map(LauncherMapper.java:242) > at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:54) > at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:453) > at org.apache.hadoop.mapred.MapTask.run(MapTask.java:343) > at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:168) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:422) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1724) > at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:162) > Caused by: org.apache.hadoop.security.AccessControlException: Authentication > required > at > org.apache.hadoop.hdfs.web.WebHdfsFileSystem.validateResponse(WebHdfsFileSystem.java:457) > at > org.apache.hadoop.hdfs.web.WebHdfsFileSystem.access$200(WebHdfsFileSystem.java:113) > at > org.apache.hadoop.hdfs.web.WebHdfsFileSystem$AbstractRunner.runWithRetry(WebHdfsFileSystem.java:738) > at > org.apache.hadoop.hdfs.web.WebHdfsFileSystem$AbstractRunner.access$100(WebHdfsFileSystem.java:582) > at > org.apache.hadoop.hdfs.web.WebHdfsFileSystem$AbstractRunner$1.run(WebHdfsFileSystem.java:612) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:422) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1724) > at > org.apache.hadoop.hdfs.web.WebHdfsFileSystem$AbstractRunner.run(WebHdfsFileSystem.java:608) > at > org.apache.hadoop.hdfs.web.WebHdfsFileSystem.getDelegationToken(WebHdfsFileSystem.java:1505) > at > org.apache.hadoop.hdfs.web.WebHdfsFileSystem.getDelegationToken(WebHdfsFileSystem.java:331) > at > org.apache.hadoop.hdfs.web.WebHdfsFileSystem.getAuthParameters(WebHdfsFileSystem.java:547) > at > org.apache.hadoop.hdfs.web.WebHdfsFileSystem.toUrl(WebHdfsFileSystem.java:568) > at > org.apache.hadoop.hdfs.web.WebHdfsFileSystem$AbstractFsPathRunner.getUrl(WebHdfsFileSystem.java:838) > at > org.apache.hadoop.hdfs.web.WebHdfsFileSystem$AbstractRunner.runWithRetry(WebHdfsFileSystem.java:733) > at > org.apache.hadoop.hdfs.web.WebHdfsFileSystem$AbstractRunner.access$100(WebHdfsFileSystem.java:582) > at > org.apache.hadoop.hdfs.web.WebHdfsFileSystem$AbstractRunner$1.run(WebHdfsFileSystem.java:612) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:422) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1724) > at > org.apache.hadoop.hdfs.web.WebHdfsFileSystem$AbstractRunner.run(WebHdfsFileSystem.java:608) > at > org.apache.hadoop.hdfs.web.WebHdfsFileSystem.getHdfsFileStatus(WebHdfsFileSystem.java:985) > at > org.apache.hadoop.hdfs.web.WebHdfsFileSystem.getFileStatus(WebHdfsFileSystem.java:1001) > at org.apache.hadoop.fs.Globber.getFileStatus(Globber.java:57) > at org.apache.hadoop.fs.Globber.glob(Globber.java:252) > at org.apache.hadoop.fs.FileSystem.globStatus(FileSystem.java:1666) > at > org.apache.hadoop.tools.GlobbedCopyListing.doBuildListing(GlobbedCopyListing.java:77) > at > org.apache.hadoop.tools.CopyListing.buildListing(CopyListing.java:86) > at > org.apache.hadoop.tools.DistCp.createInputFileListing(DistCp.java:390) > at org.apache.hadoop.tools.DistCp.createAndSubmitJob(DistCp.java:188) > at org.apache.hadoop.tools.DistCp.execute(DistCp.java:153) > at > org.apache.falcon.replication.FeedReplicator.run(FeedReplicator.java:97) > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:76) > at > org.apache.falcon.replication.FeedReplicator.main(FeedReplicator.java:62) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > at java.lang.reflect.Method.invoke(Method.java:498) > at org.apache.oozie.action.hadoop.JavaMain.run(JavaMain.java:56) > ... 15 more > Oozie Launcher failed, finishing Hadoop job gracefully > Oozie Launcher, uploading action data to HDFS sequence file: > hdfs://nat-os-r6-upns-falcon-multicluster-14.openstacklocal:8020/user/hrt_qa/oozie-oozi/0000089-160621033852265-oozie-oozi-W/dr-replication--java/action-data.seq > {noformat} -- This message was sent by Atlassian JIRA (v6.3.4#6332)