[ 
https://issues.apache.org/jira/browse/KYLIN-5066?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17402027#comment-17402027
 ] 

Alexander commented on KYLIN-5066:
----------------------------------

Hello,

patch in attachment fixes issue on Hadoop3, that came in Commit 
57cced6ba8b683a7db6239a0d042734c171aba44

In Kylin3 branch there is no such commit, and i don't undestand where to merge 
this...

 

> CubeMigrationCrossClusterCLI - distcp (java.io.FileNotFoundException): Path 
> is not a file 
> ------------------------------------------------------------------------------------------
>
>                 Key: KYLIN-5066
>                 URL: https://issues.apache.org/jira/browse/KYLIN-5066
>             Project: Kylin
>          Issue Type: Bug
>          Components: Client - CLI
>    Affects Versions: v3.1.2
>         Environment: HDP 3.1.4
>            Reporter: Alexander
>            Assignee: Alexander
>            Priority: Major
>             Fix For: v3.1.3
>
>         Attachments: 
> 0001-KYLIN-5066-CubeMigrationCrossClusterCLI-distcp-java..patch
>
>
> Try to migrate cube between clusters:
> {code:java}
> $KYLIN_HOME/bin/kylin.sh 
> org.apache.kylin.tool.migration.CubeMigrationCrossClusterCLI -kylinUriSrc 
> ADMIN:KYLIN@hostname1:7070 -kylinUriDst ADMIN:KYLIN@hostname12:7070 -cube 
> CUBE_NAME -execute true -schemaOnly false -overwrite true
> {code}
> Got exception:
> {code:java}
> 2021-08-20 00:07:04,533 ERROR 
> [Cube-LOY_TML_FAVCAT_V1_4-data-migration-pool-7] 
> migration.CubeMigrationCrossClusterCLI:679 : fail to migrate htable 
> KYLIN_7E0QA66PIM due to {} 2021-08-20 00:07:04,533 ERROR 
> [Cube-LOY_TML_FAVCAT_V1_4-data-migration-pool-7] 
> migration.CubeMigrationCrossClusterCLI:679 : fail to migrate htable 
> KYLIN_7E0QA66PIM due to {} java.io.FileNotFoundException: Path is not a file: 
> /apps/hbase/data/data/default/KYLIN_7E0QA66PIM at 
> org.apache.hadoop.hdfs.server.namenode.INodeFile.valueOf(INodeFile.java:90) 
> at 
> org.apache.hadoop.hdfs.server.namenode.INodeFile.valueOf(INodeFile.java:76) 
> at 
> org.apache.hadoop.hdfs.server.namenode.FSDirStatAndListingOp.getBlockLocations(FSDirStatAndListingOp.java:158)
>  at 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocations(FSNamesystem.java:1931)
>  at 
> org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.getBlockLocations(NameNodeRpcServer.java:738)
>  at 
> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.getBlockLocations(ClientNamenodeProtocolServerSideTranslatorPB.java:426)
>  at 
> org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
>  at 
> org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:524)
>  at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1025) at 
> org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:876) at 
> org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:822) at 
> java.security.AccessController.doPrivileged(Native Method) at 
> javax.security.auth.Subject.doAs(Subject.java:422) at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1730)
>  at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2682)
>  at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) at 
> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
>  at 
> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
>  at java.lang.reflect.Constructor.newInstance(Constructor.java:423) at 
> org.apache.hadoop.ipc.RemoteException.instantiateException(RemoteException.java:121)
>  at 
> org.apache.hadoop.ipc.RemoteException.unwrapRemoteException(RemoteException.java:88)
>  at 
> org.apache.hadoop.hdfs.DFSClient.callGetBlockLocations(DFSClient.java:864) at 
> org.apache.hadoop.hdfs.DFSClient.getLocatedBlocks(DFSClient.java:851) at 
> org.apache.hadoop.hdfs.DFSClient.getLocatedBlocks(DFSClient.java:840) at 
> org.apache.hadoop.hdfs.DFSClient.open(DFSClient.java:1004) at 
> org.apache.hadoop.hdfs.DistributedFileSystem$4.doCall(DistributedFileSystem.java:320)
>  at 
> org.apache.hadoop.hdfs.DistributedFileSystem$4.doCall(DistributedFileSystem.java:316)
>  at 
> org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
>  at 
> org.apache.hadoop.hdfs.DistributedFileSystem.open(DistributedFileSystem.java:328)
>  at org.apache.hadoop.fs.FileSystem.open(FileSystem.java:899) at 
> org.apache.hadoop.tools.FileBasedCopyListing.fetchFileList(FileBasedCopyListing.java:78)
>  at 
> org.apache.hadoop.tools.FileBasedCopyListing.doBuildListing(FileBasedCopyListing.java:69)
>  at org.apache.hadoop.tools.CopyListing.buildListing(CopyListing.java:86) at 
> org.apache.hadoop.tools.DistCp.createInputFileListing(DistCp.java:368) at 
> org.apache.hadoop.tools.DistCp.prepareFileListing(DistCp.java:96) at 
> org.apache.hadoop.tools.DistCp.createAndSubmitJob(DistCp.java:205) at 
> org.apache.hadoop.tools.DistCp.execute(DistCp.java:182) at 
> org.apache.kylin.tool.migration.CubeMigrationCrossClusterCLI.copyHDFSPath(CubeMigrationCrossClusterCLI.java:699)
>  at 
> org.apache.kylin.tool.migration.CubeMigrationCrossClusterCLI.copyHFileByDistCp(CubeMigrationCrossClusterCLI.java:689)
>  at 
> org.apache.kylin.tool.migration.CubeMigrationCrossClusterCLI.copyHTable(CubeMigrationCrossClusterCLI.java:661)
>  at 
> org.apache.kylin.tool.migration.CubeMigrationCrossClusterCLI.copyHTable(CubeMigrationCrossClusterCLI.java:613)
>  at 
> org.apache.kylin.tool.migration.CubeMigrationCrossClusterCLI.access$200(CubeMigrationCrossClusterCLI.java:87)
>  at 
> org.apache.kylin.tool.migration.CubeMigrationCrossClusterCLI$3.doRun(CubeMigrationCrossClusterCLI.java:504)
>  at 
> org.apache.kylin.tool.migration.CubeMigrationCrossClusterCLI$MyRunnable.run(CubeMigrationCrossClusterCLI.java:739)
>  at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) 
> at java.util.concurrent.FutureTask.run(FutureTask.java:266) at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
>  at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
>  at java.lang.Thread.run(Thread.java:745)Caused by: 
> org.apache.hadoop.ipc.RemoteException(java.io.FileNotFoundException): Path is 
> not a file: /apps/hbase/data/data/default/KYLIN_7E0QA66PIM at 
> org.apache.hadoop.hdfs.server.namenode.INodeFile.valueOf(INodeFile.java:90) 
> at 
> org.apache.hadoop.hdfs.server.namenode.INodeFile.valueOf(INodeFile.java:76) 
> at 
> org.apache.hadoop.hdfs.server.namenode.FSDirStatAndListingOp.getBlockLocations(FSDirStatAndListingOp.java:158)
>  at 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocations(FSNamesystem.java:1931)
>  at 
> org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.getBlockLocations(NameNodeRpcServer.java:738)
>  at 
> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.getBlockLocations(ClientNamenodeProtocolServerSideTranslatorPB.java:426)
>  at 
> org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
>  at 
> org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:524)
>  at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1025) at 
> org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:876) at 
> org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:822) at 
> java.security.AccessController.doPrivileged(Native Method) at 
> javax.security.auth.Subject.doAs(Subject.java:422) at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1730)
>  at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2682)
>  at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1498) at 
> org.apache.hadoop.ipc.Client.call(Client.java:1444) at 
> org.apache.hadoop.ipc.Client.call(Client.java:1354) at 
> org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:228)
>  at 
> org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:116)
>  at com.sun.proxy.$Proxy11.getBlockLocations(Unknown Source) at 
> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.getBlockLocations(ClientNamenodeProtocolTranslatorPB.java:317)
>  at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) 
> at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>  at java.lang.reflect.Method.invoke(Method.java:498) at 
> org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:422)
>  at 
> org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeMethod(RetryInvocationHandler.java:165)
>  at 
> org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invoke(RetryInvocationHandler.java:157)
>  at 
> org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeOnce(RetryInvocationHandler.java:95)
>  at 
> org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:359)
>  at com.sun.proxy.$Proxy12.getBlockLocations(Unknown Source) at 
> org.apache.hadoop.hdfs.DFSClient.callGetBlockLocations(DFSClient.java:862) 
> ... 27 more
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

Reply via email to