[
https://issues.apache.org/jira/browse/FLINK-24249?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Piotr Nowojski updated FLINK-24249:
-----------------------------------
Description:
flink on yarn will localize user keytab on local machine disk,
trigger checkpoint will fail when jobmanager mkdirs on hdfs when the disk
damage,but the flink job not fail,so I can't recover from checkpoint
the exception like this
{code:java}java.io.IOException: Failed on local exception: java.io.IOException:
Login failure for joey from keytab
/data06/yarn/nm/usercache/joey/appcache/application_1631093653028_0015/container_e134_1631093653028_0015_01_000001/krb5.keytab;
Host Details : local host is: "test/10.9.8.37"; destination host is:
"test-227":8020;
at org.apache.hadoop.net.NetUtils.wrapException(NetUtils.java:772)
~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0]
at org.apache.hadoop.ipc.Client.call(Client.java:1474)
~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0]
at org.apache.hadoop.ipc.Client.call(Client.java:1401)
~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0]
at
org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:232)
~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0]
at com.sun.proxy.$Proxy41.mkdirs(Unknown Source) ~[?:?]
at
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.mkdirs(ClientNamenodeProtocolTranslatorPB.java:539)
~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0]
at sun.reflect.GeneratedMethodAccessor63.invoke(Unknown Source) ~[?:?]
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
~[?:1.8.0_181]
at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0_181]
at
org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:187)
~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0]
at
org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:102)
~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0]
at com.sun.proxy.$Proxy42.mkdirs(Unknown Source) ~[?:?]
at org.apache.hadoop.hdfs.DFSClient.primitiveMkdir(DFSClient.java:2742)
~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0]
at org.apache.hadoop.hdfs.DFSClient.mkdirs(DFSClient.java:2713)
~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0]
at
org.apache.hadoop.hdfs.DistributedFileSystem$17.doCall(DistributedFileSystem.java:870)
~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0]
at
org.apache.hadoop.hdfs.DistributedFileSystem$17.doCall(DistributedFileSystem.java:866)
~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0]
at
org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0]
at
org.apache.hadoop.hdfs.DistributedFileSystem.mkdirsInternal(DistributedFileSystem.java:866)
~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0]
at
org.apache.hadoop.hdfs.DistributedFileSystem.mkdirs(DistributedFileSystem.java:859)
~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0]
at org.apache.hadoop.fs.FileSystem.mkdirs(FileSystem.java:1819)
~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0]
at
org.apache.flink.runtime.fs.hdfs.HadoopFileSystem.mkdirs(HadoopFileSystem.java:183)
~[flink-dist_2.11-1.13.2.jar:1.13.2]
at
org.apache.flink.runtime.state.filesystem.FsCheckpointStorageAccess.initializeLocationForCheckpoint(FsCheckpointStorageAccess.java:129)
~[flink-dist_2.11-1.13.2.jar:1.13.2]
at
org.apache.flink.runtime.checkpoint.CheckpointCoordinator.initializeCheckpoint(CheckpointCoordinator.java:689)
~[flink-dist_2.11-1.13.2.jar:1.13.2]
at
org.apache.flink.runtime.checkpoint.CheckpointCoordinator.lambda$startTriggeringCheckpoint$2(CheckpointCoordinator.java:543)
~[flink-dist_2.11-1.13.2.jar:1.13.2]
at java.util.concurrent.CompletableFuture.uniApply(CompletableFuture.java:602)
[?:1.8.0_181]
at
java.util.concurrent.CompletableFuture$UniApply.tryFire(CompletableFuture.java:577)
[?:1.8.0_181]
at
java.util.concurrent.CompletableFuture$Completion.run(CompletableFuture.java:442)
[?:1.8.0_181]
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
[?:1.8.0_181]
at java.util.concurrent.FutureTask.run(FutureTask.java:266) [?:1.8.0_181]
at
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180)
[?:1.8.0_181]
at
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293)
[?:1.8.0_181]
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
[?:1.8.0_181]
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
[?:1.8.0_181]
at java.lang.Thread.run(Thread.java:748) [?:1.8.0_181]Caused by:
java.io.IOException: Login failure for joey from keytab
/data06/yarn/nm/usercache/joey/appcache/application_1631093653028_0015/container_e134_1631093653028_0015_01_000001/krb5.keytab
at
org.apache.hadoop.security.UserGroupInformation.reloginFromKeytab(UserGroupInformation.java:1086)
~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0]
at org.apache.hadoop.ipc.Client$Connection$1.run(Client.java:659)
~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0]
at java.security.AccessController.doPrivileged(Native Method) ~[?:1.8.0_181]
at javax.security.auth.Subject.doAs(Subject.java:422) ~[?:1.8.0_181]
at
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1692)
~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0]
at
org.apache.hadoop.ipc.Client$Connection.handleSaslConnectionFailure(Client.java:645)
~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0]
at org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:732)
~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0]
at org.apache.hadoop.ipc.Client$Connection.access$2800(Client.java:370)
~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0]
at org.apache.hadoop.ipc.Client.getConnection(Client.java:1523)
~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0]
at org.apache.hadoop.ipc.Client.call(Client.java:1440)
~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0]
... 32 moreCaused by: javax.security.auth.login.LoginException: Unable to
obtain password from user
at
com.sun.security.auth.module.Krb5LoginModule.promptForPass(Krb5LoginModule.java:897)
~[?:1.8.0_181]
at
com.sun.security.auth.module.Krb5LoginModule.attemptAuthentication(Krb5LoginModule.java:760)
~[?:1.8.0_181]
at com.sun.security.auth.module.Krb5LoginModule.login(Krb5LoginModule.java:617)
~[?:1.8.0_181]
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:1.8.0_181]
at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
~[?:1.8.0_181]
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
~[?:1.8.0_181]
at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0_181]
at javax.security.auth.login.LoginContext.invoke(LoginContext.java:755)
~[?:1.8.0_181]
at javax.security.auth.login.LoginContext.access$000(LoginContext.java:195)
~[?:1.8.0_181]
at javax.security.auth.login.LoginContext$4.run(LoginContext.java:682)
~[?:1.8.0_181]
at javax.security.auth.login.LoginContext$4.run(LoginContext.java:680)
~[?:1.8.0_181]
at java.security.AccessController.doPrivileged(Native Method) ~[?:1.8.0_181]
at javax.security.auth.login.LoginContext.invokePriv(LoginContext.java:680)
~[?:1.8.0_181]
at javax.security.auth.login.LoginContext.login(LoginContext.java:587)
~[?:1.8.0_181]
at
org.apache.hadoop.security.UserGroupInformation.reloginFromKeytab(UserGroupInformation.java:1078)
~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0]
at org.apache.hadoop.ipc.Client$Connection$1.run(Client.java:659)
~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0]
at java.security.AccessController.doPrivileged(Native Method) ~[?:1.8.0_181]
at javax.security.auth.Subject.doAs(Subject.java:422) ~[?:1.8.0_181]
at
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1692)
~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0]
at
org.apache.hadoop.ipc.Client$Connection.handleSaslConnectionFailure(Client.java:645)
~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0]
at org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:732)
~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0]
at org.apache.hadoop.ipc.Client$Connection.access$2800(Client.java:370)
~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0]
at org.apache.hadoop.ipc.Client.getConnection(Client.java:1523)
~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0]
at org.apache.hadoop.ipc.Client.call(Client.java:1440)
~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0]
... 32 more
{code}
was:
flink on yarn will localize user keytab on local machine disk,
trigger checkpoint will fail when jobmanager mkdirs on hdfs when the disk
damage,but the flink job not fail,so I can't recover from checkpoint
the exception like this
{code:java}
java.io.IOException: Failed on local exception: java.io.IOException: Login
failure for joey from keytab
/data06/yarn/nm/usercache/joey/appcache/application_1631093653028_0015/container_e134_1631093653028_0015_01_000001/krb5.keytab;
Host Details : local host is: "test/10.9.8.37"; destination host is:
"test-227":8020; at
org.apache.hadoop.net.NetUtils.wrapException(NetUtils.java:772)
~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0] at
org.apache.hadoop.ipc.Client.call(Client.java:1474)
~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0] at
org.apache.hadoop.ipc.Client.call(Client.java:1401)
~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0] at
org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:232)
~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0] at
com.sun.proxy.$Proxy41.mkdirs(Unknown Source) ~[?:?] at
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.mkdirs(ClientNamenodeProtocolTranslatorPB.java:539)
~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0] at
sun.reflect.GeneratedMethodAccessor63.invoke(Unknown Source) ~[?:?] at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
~[?:1.8.0_181] at java.lang.reflect.Method.invoke(Method.java:498)
~[?:1.8.0_181] at
org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:187)
~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0] at
org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:102)
~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0] at
com.sun.proxy.$Proxy42.mkdirs(Unknown Source) ~[?:?] at
org.apache.hadoop.hdfs.DFSClient.primitiveMkdir(DFSClient.java:2742)
~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0] at
org.apache.hadoop.hdfs.DFSClient.mkdirs(DFSClient.java:2713)
~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0] at
org.apache.hadoop.hdfs.DistributedFileSystem$17.doCall(DistributedFileSystem.java:870)
~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0] at
org.apache.hadoop.hdfs.DistributedFileSystem$17.doCall(DistributedFileSystem.java:866)
~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0] at
org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0] at
org.apache.hadoop.hdfs.DistributedFileSystem.mkdirsInternal(DistributedFileSystem.java:866)
~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0] at
org.apache.hadoop.hdfs.DistributedFileSystem.mkdirs(DistributedFileSystem.java:859)
~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0] at
org.apache.hadoop.fs.FileSystem.mkdirs(FileSystem.java:1819)
~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0] at
org.apache.flink.runtime.fs.hdfs.HadoopFileSystem.mkdirs(HadoopFileSystem.java:183)
~[flink-dist_2.11-1.13.2.jar:1.13.2] at
org.apache.flink.runtime.state.filesystem.FsCheckpointStorageAccess.initializeLocationForCheckpoint(FsCheckpointStorageAccess.java:129)
~[flink-dist_2.11-1.13.2.jar:1.13.2] at
org.apache.flink.runtime.checkpoint.CheckpointCoordinator.initializeCheckpoint(CheckpointCoordinator.java:689)
~[flink-dist_2.11-1.13.2.jar:1.13.2] at
org.apache.flink.runtime.checkpoint.CheckpointCoordinator.lambda$startTriggeringCheckpoint$2(CheckpointCoordinator.java:543)
~[flink-dist_2.11-1.13.2.jar:1.13.2] at
java.util.concurrent.CompletableFuture.uniApply(CompletableFuture.java:602)
[?:1.8.0_181] at
java.util.concurrent.CompletableFuture$UniApply.tryFire(CompletableFuture.java:577)
[?:1.8.0_181] at
java.util.concurrent.CompletableFuture$Completion.run(CompletableFuture.java:442)
[?:1.8.0_181] at
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
[?:1.8.0_181] at java.util.concurrent.FutureTask.run(FutureTask.java:266)
[?:1.8.0_181] at
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180)
[?:1.8.0_181] at
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293)
[?:1.8.0_181] at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
[?:1.8.0_181] at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
[?:1.8.0_181] at java.lang.Thread.run(Thread.java:748) [?:1.8.0_181]Caused
by: java.io.IOException: Login failure for joey from keytab
/data06/yarn/nm/usercache/joey/appcache/application_1631093653028_0015/container_e134_1631093653028_0015_01_000001/krb5.keytab
at
org.apache.hadoop.security.UserGroupInformation.reloginFromKeytab(UserGroupInformation.java:1086)
~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0] at
org.apache.hadoop.ipc.Client$Connection$1.run(Client.java:659)
~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0] at
java.security.AccessController.doPrivileged(Native Method) ~[?:1.8.0_181] at
javax.security.auth.Subject.doAs(Subject.java:422) ~[?:1.8.0_181] at
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1692)
~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0] at
org.apache.hadoop.ipc.Client$Connection.handleSaslConnectionFailure(Client.java:645)
~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0] at
org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:732)
~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0] at
org.apache.hadoop.ipc.Client$Connection.access$2800(Client.java:370)
~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0] at
org.apache.hadoop.ipc.Client.getConnection(Client.java:1523)
~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0] at
org.apache.hadoop.ipc.Client.call(Client.java:1440)
~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0] ... 32 moreCaused
by: javax.security.auth.login.LoginException: Unable to obtain password from
user
at
com.sun.security.auth.module.Krb5LoginModule.promptForPass(Krb5LoginModule.java:897)
~[?:1.8.0_181] at
com.sun.security.auth.module.Krb5LoginModule.attemptAuthentication(Krb5LoginModule.java:760)
~[?:1.8.0_181] at
com.sun.security.auth.module.Krb5LoginModule.login(Krb5LoginModule.java:617)
~[?:1.8.0_181] at sun.reflect.NativeMethodAccessorImpl.invoke0(Native
Method) ~[?:1.8.0_181] at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
~[?:1.8.0_181] at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
~[?:1.8.0_181] at java.lang.reflect.Method.invoke(Method.java:498)
~[?:1.8.0_181] at
javax.security.auth.login.LoginContext.invoke(LoginContext.java:755)
~[?:1.8.0_181] at
javax.security.auth.login.LoginContext.access$000(LoginContext.java:195)
~[?:1.8.0_181] at
javax.security.auth.login.LoginContext$4.run(LoginContext.java:682)
~[?:1.8.0_181] at
javax.security.auth.login.LoginContext$4.run(LoginContext.java:680)
~[?:1.8.0_181] at java.security.AccessController.doPrivileged(Native Method)
~[?:1.8.0_181] at
javax.security.auth.login.LoginContext.invokePriv(LoginContext.java:680)
~[?:1.8.0_181] at
javax.security.auth.login.LoginContext.login(LoginContext.java:587)
~[?:1.8.0_181] at
org.apache.hadoop.security.UserGroupInformation.reloginFromKeytab(UserGroupInformation.java:1078)
~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0] at
org.apache.hadoop.ipc.Client$Connection$1.run(Client.java:659)
~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0] at
java.security.AccessController.doPrivileged(Native Method) ~[?:1.8.0_181] at
javax.security.auth.Subject.doAs(Subject.java:422) ~[?:1.8.0_181] at
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1692)
~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0] at
org.apache.hadoop.ipc.Client$Connection.handleSaslConnectionFailure(Client.java:645)
~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0] at
org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:732)
~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0] at
org.apache.hadoop.ipc.Client$Connection.access$2800(Client.java:370)
~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0] at
org.apache.hadoop.ipc.Client.getConnection(Client.java:1523)
~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0] at
org.apache.hadoop.ipc.Client.call(Client.java:1440)
~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0] ... 32 more
{code}
> login from keytab fail when disk damage
> ---------------------------------------
>
> Key: FLINK-24249
> URL: https://issues.apache.org/jira/browse/FLINK-24249
> Project: Flink
> Issue Type: Bug
> Components: Runtime / Checkpointing
> Affects Versions: 1.13.2
> Reporter: YuAngZhang
> Priority: Major
>
> flink on yarn will localize user keytab on local machine disk,
> trigger checkpoint will fail when jobmanager mkdirs on hdfs when the disk
> damage,but the flink job not fail,so I can't recover from checkpoint
>
> the exception like this
> {code:java}java.io.IOException: Failed on local exception:
> java.io.IOException: Login failure for joey from keytab
> /data06/yarn/nm/usercache/joey/appcache/application_1631093653028_0015/container_e134_1631093653028_0015_01_000001/krb5.keytab;
> Host Details : local host is: "test/10.9.8.37"; destination host is:
> "test-227":8020;
> at org.apache.hadoop.net.NetUtils.wrapException(NetUtils.java:772)
> ~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0]
> at org.apache.hadoop.ipc.Client.call(Client.java:1474)
> ~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0]
> at org.apache.hadoop.ipc.Client.call(Client.java:1401)
> ~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0]
> at
> org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:232)
> ~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0]
> at com.sun.proxy.$Proxy41.mkdirs(Unknown Source) ~[?:?]
> at
> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.mkdirs(ClientNamenodeProtocolTranslatorPB.java:539)
> ~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0]
> at sun.reflect.GeneratedMethodAccessor63.invoke(Unknown Source) ~[?:?]
> at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> ~[?:1.8.0_181]
> at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0_181]
> at
> org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:187)
> ~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0]
> at
> org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:102)
> ~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0]
> at com.sun.proxy.$Proxy42.mkdirs(Unknown Source) ~[?:?]
> at org.apache.hadoop.hdfs.DFSClient.primitiveMkdir(DFSClient.java:2742)
> ~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0]
> at org.apache.hadoop.hdfs.DFSClient.mkdirs(DFSClient.java:2713)
> ~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0]
> at
> org.apache.hadoop.hdfs.DistributedFileSystem$17.doCall(DistributedFileSystem.java:870)
> ~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0]
> at
> org.apache.hadoop.hdfs.DistributedFileSystem$17.doCall(DistributedFileSystem.java:866)
> ~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0]
> at
> org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
> ~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0]
> at
> org.apache.hadoop.hdfs.DistributedFileSystem.mkdirsInternal(DistributedFileSystem.java:866)
> ~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0]
> at
> org.apache.hadoop.hdfs.DistributedFileSystem.mkdirs(DistributedFileSystem.java:859)
> ~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0]
> at org.apache.hadoop.fs.FileSystem.mkdirs(FileSystem.java:1819)
> ~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0]
> at
> org.apache.flink.runtime.fs.hdfs.HadoopFileSystem.mkdirs(HadoopFileSystem.java:183)
> ~[flink-dist_2.11-1.13.2.jar:1.13.2]
> at
> org.apache.flink.runtime.state.filesystem.FsCheckpointStorageAccess.initializeLocationForCheckpoint(FsCheckpointStorageAccess.java:129)
> ~[flink-dist_2.11-1.13.2.jar:1.13.2]
> at
> org.apache.flink.runtime.checkpoint.CheckpointCoordinator.initializeCheckpoint(CheckpointCoordinator.java:689)
> ~[flink-dist_2.11-1.13.2.jar:1.13.2]
> at
> org.apache.flink.runtime.checkpoint.CheckpointCoordinator.lambda$startTriggeringCheckpoint$2(CheckpointCoordinator.java:543)
> ~[flink-dist_2.11-1.13.2.jar:1.13.2]
> at
> java.util.concurrent.CompletableFuture.uniApply(CompletableFuture.java:602)
> [?:1.8.0_181]
> at
> java.util.concurrent.CompletableFuture$UniApply.tryFire(CompletableFuture.java:577)
> [?:1.8.0_181]
> at
> java.util.concurrent.CompletableFuture$Completion.run(CompletableFuture.java:442)
> [?:1.8.0_181]
> at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
> [?:1.8.0_181]
> at java.util.concurrent.FutureTask.run(FutureTask.java:266) [?:1.8.0_181]
> at
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180)
> [?:1.8.0_181]
> at
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293)
> [?:1.8.0_181]
> at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
> [?:1.8.0_181]
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> [?:1.8.0_181]
> at java.lang.Thread.run(Thread.java:748) [?:1.8.0_181]Caused by:
> java.io.IOException: Login failure for joey from keytab
> /data06/yarn/nm/usercache/joey/appcache/application_1631093653028_0015/container_e134_1631093653028_0015_01_000001/krb5.keytab
> at
> org.apache.hadoop.security.UserGroupInformation.reloginFromKeytab(UserGroupInformation.java:1086)
> ~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0]
> at org.apache.hadoop.ipc.Client$Connection$1.run(Client.java:659)
> ~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0]
> at java.security.AccessController.doPrivileged(Native Method) ~[?:1.8.0_181]
> at javax.security.auth.Subject.doAs(Subject.java:422) ~[?:1.8.0_181]
> at
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1692)
> ~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0]
> at
> org.apache.hadoop.ipc.Client$Connection.handleSaslConnectionFailure(Client.java:645)
> ~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0]
> at org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:732)
> ~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0]
> at org.apache.hadoop.ipc.Client$Connection.access$2800(Client.java:370)
> ~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0]
> at org.apache.hadoop.ipc.Client.getConnection(Client.java:1523)
> ~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0]
> at org.apache.hadoop.ipc.Client.call(Client.java:1440)
> ~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0]
> ... 32 moreCaused by: javax.security.auth.login.LoginException: Unable to
> obtain password from user
> at
> com.sun.security.auth.module.Krb5LoginModule.promptForPass(Krb5LoginModule.java:897)
> ~[?:1.8.0_181]
> at
> com.sun.security.auth.module.Krb5LoginModule.attemptAuthentication(Krb5LoginModule.java:760)
> ~[?:1.8.0_181]
> at
> com.sun.security.auth.module.Krb5LoginModule.login(Krb5LoginModule.java:617)
> ~[?:1.8.0_181]
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:1.8.0_181]
> at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> ~[?:1.8.0_181]
> at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> ~[?:1.8.0_181]
> at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0_181]
> at javax.security.auth.login.LoginContext.invoke(LoginContext.java:755)
> ~[?:1.8.0_181]
> at javax.security.auth.login.LoginContext.access$000(LoginContext.java:195)
> ~[?:1.8.0_181]
> at javax.security.auth.login.LoginContext$4.run(LoginContext.java:682)
> ~[?:1.8.0_181]
> at javax.security.auth.login.LoginContext$4.run(LoginContext.java:680)
> ~[?:1.8.0_181]
> at java.security.AccessController.doPrivileged(Native Method) ~[?:1.8.0_181]
> at javax.security.auth.login.LoginContext.invokePriv(LoginContext.java:680)
> ~[?:1.8.0_181]
> at javax.security.auth.login.LoginContext.login(LoginContext.java:587)
> ~[?:1.8.0_181]
> at
> org.apache.hadoop.security.UserGroupInformation.reloginFromKeytab(UserGroupInformation.java:1078)
> ~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0]
> at org.apache.hadoop.ipc.Client$Connection$1.run(Client.java:659)
> ~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0]
> at java.security.AccessController.doPrivileged(Native Method) ~[?:1.8.0_181]
> at javax.security.auth.Subject.doAs(Subject.java:422) ~[?:1.8.0_181]
> at
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1692)
> ~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0]
> at
> org.apache.hadoop.ipc.Client$Connection.handleSaslConnectionFailure(Client.java:645)
> ~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0]
> at org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:732)
> ~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0]
> at org.apache.hadoop.ipc.Client$Connection.access$2800(Client.java:370)
> ~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0]
> at org.apache.hadoop.ipc.Client.getConnection(Client.java:1523)
> ~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0]
> at org.apache.hadoop.ipc.Client.call(Client.java:1440)
> ~[flink-shaded-hadoop-2-uber-2.6.5-10.0.jar:2.6.5-10.0]
> ... 32 more
> {code}
--
This message was sent by Atlassian Jira
(v8.3.4#803005)