Nico, thank you for your reply./I looked at the commit you cherry-picked and
nothing in there explains theerror you got./==> The commit I cherry-picked
makes setting of 'zookeeper.sasl.disable' work correctly. I changed
flink-dist_2.11-1.2.0.jar according to it.So now zookeeper.sasl problem is
gone.Yes, the error log I posted in the original message is completely
different one./Can you verify that nothing of your flink 1.3 tests
remains/==> Below is what I just reproduced. I have 4 nodes cluster with
non-secure.After run yarn-session.sh, JM process be created in flink-03 node
but TM process not.Standalone works well.Any clue would be really
appreciate. Thanks.[bistel@flink-01 ~]$ jps1888 ResourceManager2000
NodeManager2433 NameNode2546 DataNode2754 SecondaryNameNode2891 Jps1724
QuorumPeerMain[bistel@flink-02 ~]$ jps2018 Jps1721 NodeManager1881
DataNode1515 QuorumPeerMain[bistel@flink-03 ~]$ jps1521 QuorumPeerMain1975
Jps1724 NodeManager1885 DataNode[bistel@flink-04 ~]$ jps2090 Jps1515
QuorumPeerMain1789 NodeManager1950 DataNode[bistel@flink-01 ~]$
/usr/local/flink-1.2.0/bin/yarn-session.sh -n 42017-09-07 09:49:35,467 INFO 
org.apache.flink.configuration.GlobalConfiguration            - Loading
configuration property: jobmanager.rpc.address, flink-012017-09-07
09:49:35,468 INFO  org.apache.flink.configuration.GlobalConfiguration           
- Loading configuration property: jobmanager.rpc.port, 61232017-09-07
09:49:35,468 INFO  org.apache.flink.configuration.GlobalConfiguration           
- Loading configuration property: jobmanager.heap.mb, 40962017-09-07
09:49:35,468 INFO  org.apache.flink.configuration.GlobalConfiguration           
- Loading configuration property: taskmanager.heap.mb, 81922017-09-07
09:49:35,468 INFO  org.apache.flink.configuration.GlobalConfiguration           
- Loading configuration property: taskmanager.numberOfTaskSlots, 42017-09-07
09:49:35,469 INFO  org.apache.flink.configuration.GlobalConfiguration           
- Loading configuration property: taskmanager.memory.preallocate,
false2017-09-07 09:49:35,469 INFO 
org.apache.flink.configuration.GlobalConfiguration            - Loading
configuration property: parallelism.default, 42017-09-07 09:49:35,469 INFO 
org.apache.flink.configuration.GlobalConfiguration            - Loading
configuration property: jobmanager.web.port, 80812017-09-07 09:49:35,469
INFO  org.apache.flink.configuration.GlobalConfiguration            -
Loading configuration property: fs.hdfs.hadoopconf,
/usr/local/hadoop/etc/hadoop/2017-09-07 09:49:35,470 INFO 
org.apache.flink.configuration.GlobalConfiguration            - Loading
configuration property: high-availability, zookeeper2017-09-07 09:49:35,470
INFO  org.apache.flink.configuration.GlobalConfiguration            -
Loading configuration property: high-availability.zookeeper.quorum,
flink-01:2181,flink-02:2181,flink-03:2181,flink-04:21812017-09-07
09:49:35,470 INFO  org.apache.flink.configuration.GlobalConfiguration           
- Loading configuration property: high-availability.zookeeper.path.root,
/flink2017-09-07 09:49:35,470 INFO 
org.apache.flink.configuration.GlobalConfiguration            - Loading
configuration property: high-availability.zookeeper.path.namespace,
/cluster_one2017-09-07 09:49:35,470 INFO 
org.apache.flink.configuration.GlobalConfiguration            - Loading
configuration property: high-availability.zookeeper.storageDir,
hdfs:///flink/recovery2017-09-07 09:49:35,470 INFO 
org.apache.flink.configuration.GlobalConfiguration            - Loading
configuration property: yarn.application-attempts, 102017-09-07 09:49:35,470
INFO  org.apache.flink.configuration.GlobalConfiguration            -
Loading configuration property: yarn.containers.vcores, 202017-09-07
09:49:35,471 INFO  org.apache.flink.configuration.GlobalConfiguration           
- Loading configuration property:
yarn.application-master.env.LD_LIBRARY_PATH,
/opt/tibco/TIBRV/8.0/lib2017-09-07 09:49:35,471 INFO 
org.apache.flink.configuration.GlobalConfiguration            - Loading
configuration property: yarn.taskmanager.env.LD_LIBRARY_PATH,
/opt/tibco/TIBRV/8.0/lib2017-09-07 09:49:35,471 INFO 
org.apache.flink.configuration.GlobalConfiguration            - Loading
configuration property: zookeeper.sasl.disable, true2017-09-07 09:49:35,662
INFO  org.apache.flink.configuration.GlobalConfiguration            -
Loading configuration property: jobmanager.rpc.address, flink-012017-09-07
09:49:35,662 INFO  org.apache.flink.configuration.GlobalConfiguration           
- Loading configuration property: jobmanager.rpc.port, 61232017-09-07
09:49:35,662 INFO  org.apache.flink.configuration.GlobalConfiguration           
- Loading configuration property: jobmanager.heap.mb, 40962017-09-07
09:49:35,663 INFO  org.apache.flink.configuration.GlobalConfiguration           
- Loading configuration property: taskmanager.heap.mb, 81922017-09-07
09:49:35,663 INFO  org.apache.flink.configuration.GlobalConfiguration           
- Loading configuration property: taskmanager.numberOfTaskSlots, 42017-09-07
09:49:35,663 INFO  org.apache.flink.configuration.GlobalConfiguration           
- Loading configuration property: taskmanager.memory.preallocate,
false2017-09-07 09:49:35,663 INFO 
org.apache.flink.configuration.GlobalConfiguration            - Loading
configuration property: parallelism.default, 42017-09-07 09:49:35,663 INFO 
org.apache.flink.configuration.GlobalConfiguration            - Loading
configuration property: jobmanager.web.port, 80812017-09-07 09:49:35,663
INFO  org.apache.flink.configuration.GlobalConfiguration            -
Loading configuration property: fs.hdfs.hadoopconf,
/usr/local/hadoop/etc/hadoop/2017-09-07 09:49:35,664 INFO 
org.apache.flink.configuration.GlobalConfiguration            - Loading
configuration property: high-availability, zookeeper2017-09-07 09:49:35,664
INFO  org.apache.flink.configuration.GlobalConfiguration            -
Loading configuration property: high-availability.zookeeper.quorum,
flink-01:2181,flink-02:2181,flink-03:2181,flink-04:21812017-09-07
09:49:35,664 INFO  org.apache.flink.configuration.GlobalConfiguration           
- Loading configuration property: high-availability.zookeeper.path.root,
/flink2017-09-07 09:49:35,664 INFO 
org.apache.flink.configuration.GlobalConfiguration            - Loading
configuration property: high-availability.zookeeper.path.namespace,
/cluster_one2017-09-07 09:49:35,664 INFO 
org.apache.flink.configuration.GlobalConfiguration            - Loading
configuration property: high-availability.zookeeper.storageDir,
hdfs:///flink/recovery2017-09-07 09:49:35,664 INFO 
org.apache.flink.configuration.GlobalConfiguration            - Loading
configuration property: yarn.application-attempts, 102017-09-07 09:49:35,664
INFO  org.apache.flink.configuration.GlobalConfiguration            -
Loading configuration property: yarn.containers.vcores, 202017-09-07
09:49:35,664 INFO  org.apache.flink.configuration.GlobalConfiguration           
- Loading configuration property:
yarn.application-master.env.LD_LIBRARY_PATH,
/opt/tibco/TIBRV/8.0/lib2017-09-07 09:49:35,665 INFO 
org.apache.flink.configuration.GlobalConfiguration            - Loading
configuration property: yarn.taskmanager.env.LD_LIBRARY_PATH,
/opt/tibco/TIBRV/8.0/lib2017-09-07 09:49:35,665 INFO 
org.apache.flink.configuration.GlobalConfiguration            - Loading
configuration property: zookeeper.sasl.disable, true2017-09-07 09:49:36,519
WARN  org.apache.hadoop.util.NativeCodeLoader                       - Unable
to load native-hadoop library for your platform... using builtin-java
classes where applicable2017-09-07 09:49:36,779 INFO 
org.apache.flink.runtime.security.modules.HadoopModule        - Hadoop user
set to bistel (auth:SIMPLE)2017-09-07 09:49:37,084 INFO 
org.apache.flink.configuration.GlobalConfiguration            - Loading
configuration property: jobmanager.rpc.address, flink-012017-09-07
09:49:37,084 INFO  org.apache.flink.configuration.GlobalConfiguration           
- Loading configuration property: jobmanager.rpc.port, 61232017-09-07
09:49:37,084 INFO  org.apache.flink.configuration.GlobalConfiguration           
- Loading configuration property: jobmanager.heap.mb, 40962017-09-07
09:49:37,084 INFO  org.apache.flink.configuration.GlobalConfiguration           
- Loading configuration property: taskmanager.heap.mb, 81922017-09-07
09:49:37,084 INFO  org.apache.flink.configuration.GlobalConfiguration           
- Loading configuration property: taskmanager.numberOfTaskSlots, 42017-09-07
09:49:37,084 INFO  org.apache.flink.configuration.GlobalConfiguration           
- Loading configuration property: taskmanager.memory.preallocate,
false2017-09-07 09:49:37,085 INFO 
org.apache.flink.configuration.GlobalConfiguration            - Loading
configuration property: parallelism.default, 42017-09-07 09:49:37,085 INFO 
org.apache.flink.configuration.GlobalConfiguration            - Loading
configuration property: jobmanager.web.port, 80812017-09-07 09:49:37,085
INFO  org.apache.flink.configuration.GlobalConfiguration            -
Loading configuration property: fs.hdfs.hadoopconf,
/usr/local/hadoop/etc/hadoop/2017-09-07 09:49:37,085 INFO 
org.apache.flink.configuration.GlobalConfiguration            - Loading
configuration property: high-availability, zookeeper2017-09-07 09:49:37,085
INFO  org.apache.flink.configuration.GlobalConfiguration            -
Loading configuration property: high-availability.zookeeper.quorum,
flink-01:2181,flink-02:2181,flink-03:2181,flink-04:21812017-09-07
09:49:37,085 INFO  org.apache.flink.configuration.GlobalConfiguration           
- Loading configuration property: high-availability.zookeeper.path.root,
/flink2017-09-07 09:49:37,085 INFO 
org.apache.flink.configuration.GlobalConfiguration            - Loading
configuration property: high-availability.zookeeper.path.namespace,
/cluster_one2017-09-07 09:49:37,085 INFO 
org.apache.flink.configuration.GlobalConfiguration            - Loading
configuration property: high-availability.zookeeper.storageDir,
hdfs:///flink/recovery2017-09-07 09:49:37,086 INFO 
org.apache.flink.configuration.GlobalConfiguration            - Loading
configuration property: yarn.application-attempts, 102017-09-07 09:49:37,086
INFO  org.apache.flink.configuration.GlobalConfiguration            -
Loading configuration property: yarn.containers.vcores, 202017-09-07
09:49:37,086 INFO  org.apache.flink.configuration.GlobalConfiguration           
- Loading configuration property:
yarn.application-master.env.LD_LIBRARY_PATH,
/opt/tibco/TIBRV/8.0/lib2017-09-07 09:49:37,086 INFO 
org.apache.flink.configuration.GlobalConfiguration            - Loading
configuration property: yarn.taskmanager.env.LD_LIBRARY_PATH,
/opt/tibco/TIBRV/8.0/lib2017-09-07 09:49:37,086 INFO 
org.apache.flink.configuration.GlobalConfiguration            - Loading
configuration property: zookeeper.sasl.disable, true2017-09-07 09:49:37,103
INFO  org.apache.flink.yarn.YarnClusterDescriptor                   - Using
values:2017-09-07 09:49:37,103 INFO 
org.apache.flink.yarn.YarnClusterDescriptor                   -  
TaskManager count = 42017-09-07 09:49:37,103 INFO 
org.apache.flink.yarn.YarnClusterDescriptor                   -   JobManager
memory = 10242017-09-07 09:49:37,103 INFO 
org.apache.flink.yarn.YarnClusterDescriptor                   -  
TaskManager memory = 10242017-09-07 09:49:37,118 INFO 
org.apache.hadoop.yarn.client.RMProxy                         - Connecting
to ResourceManager at flink-01/10.1.0.4:80322017-09-07 09:49:39,084 INFO 
org.apache.flink.yarn.Utils                                   - Copying from
file:/usr/local/flink-1.2.0/lib to
hdfs://flink-01:9000/user/bistel/.flink/application_1504745288687_0001/lib2017-09-07
09:49:43,419 INFO  org.apache.flink.yarn.Utils                                  
- Copying from file:/usr/local/flink-1.2.0/conf/log4j.properties to
hdfs://flink-01:9000/user/bistel/.flink/application_1504745288687_0001/log4j.properties2017-09-07
09:49:43,552 INFO  org.apache.flink.yarn.Utils                                  
- Copying from file:/usr/local/flink-1.2.0/lib/flink-dist_2.11-1.2.0.jar to
hdfs://flink-01:9000/user/bistel/.flink/application_1504745288687_0001/flink-dist_2.11-1.2.0.jar2017-09-07
09:49:43,816 INFO  org.apache.flink.yarn.Utils                                  
- Copying from /usr/local/flink-1.2.0/conf/flink-conf.yaml to
hdfs://flink-01:9000/user/bistel/.flink/application_1504745288687_0001/flink-conf.yaml2017-09-07
09:49:43,903 INFO  org.apache.flink.yarn.YarnClusterDescriptor                  
- Submitting application master application_1504745288687_00012017-09-07
09:49:44,011 INFO  org.apache.hadoop.yarn.client.api.impl.YarnClientImpl        
- Submitted application application_1504745288687_00012017-09-07
09:49:44,011 INFO  org.apache.flink.yarn.YarnClusterDescriptor                  
- Waiting for the cluster to be allocated2017-09-07 09:49:44,030 INFO 
org.apache.flink.yarn.YarnClusterDescriptor                   - Deploying
cluster, current state ACCEPTED2017-09-07 09:49:50,326 INFO 
org.apache.flink.yarn.YarnClusterDescriptor                   - YARN
application has been deployed successfully.Exception in thread "main"
java.lang.RuntimeException: Failed to retrieve JobManager address        at
org.apache.flink.client.program.ClusterClient.getJobManagerAddress(ClusterClient.java:248)
       
at
org.apache.flink.yarn.cli.FlinkYarnSessionCli.run(FlinkYarnSessionCli.java:627) 
      
at
org.apache.flink.yarn.cli.FlinkYarnSessionCli$1.call(FlinkYarnSessionCli.java:476)
       
at
org.apache.flink.yarn.cli.FlinkYarnSessionCli$1.call(FlinkYarnSessionCli.java:473)
       
at
org.apache.flink.runtime.security.HadoopSecurityContext$1.run(HadoopSecurityContext.java:43)
       
at java.security.AccessController.doPrivileged(Native Method)        at
javax.security.auth.Subject.doAs(Subject.java:422)        at
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1656)
       
at
org.apache.flink.runtime.security.HadoopSecurityContext.runSecured(HadoopSecurityContext.java:40)
       
at
org.apache.flink.yarn.cli.FlinkYarnSessionCli.main(FlinkYarnSessionCli.java:473)Caused
by: org.apache.flink.runtime.leaderretrieval.LeaderRetrievalException: Could
not retrieve the leader address and leader session ID.        at
org.apache.flink.runtime.util.LeaderRetrievalUtils.retrieveLeaderConnectionInfo(LeaderRetrievalUtils.java:175)
       
at
org.apache.flink.client.program.ClusterClient.getJobManagerAddress(ClusterClient.java:242)
       
... 9 moreCaused by: java.util.concurrent.TimeoutException: Futures timed
out after [60000 milliseconds]        at
scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)       
at scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)       
at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:190)       
at
scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
       
at scala.concurrent.Await$.result(package.scala:190)        at
scala.concurrent.Await.result(package.scala)        at
org.apache.flink.runtime.util.LeaderRetrievalUtils.retrieveLeaderConnectionInfo(LeaderRetrievalUtils.java:173)
       
... 10 more2017-09-07 09:50:51,519 INFO 
org.apache.flink.yarn.YarnClusterClient                       - Shutting
down YarnClusterClient from the client shutdown hook2017-09-07 09:50:51,519
INFO  org.apache.flink.yarn.YarnClusterClient                       -
Sending shutdown request to the Application Master2017-09-07 09:50:51,549
INFO  org.apache.flink.yarn.YarnClusterClient                       - Start
application client.2017-09-07 09:50:51,549 INFO 
org.apache.flink.yarn.YarnClusterClient                       - Starting
client actor system.2017-09-07 09:50:51,807 INFO 
akka.event.slf4j.Slf4jLogger                                  - Slf4jLogger
started2017-09-07 09:50:51,836 INFO  Remoting                                   
                  
- Starting remoting2017-09-07 09:50:51,936 INFO  Remoting                       
                              
- Remoting started; listening on addresses
:[akka.tcp://flink@flink-01:45463]2017-09-07 09:50:51,954 INFO 
org.apache.flink.yarn.ApplicationClient                       - Sending
StopCluster request to JobManager.2017-09-07 09:50:52,967 INFO 
org.apache.flink.yarn.ApplicationClient                       - Sending
StopCluster request to JobManager.2017-09-07 09:50:53,986 INFO 
org.apache.flink.yarn.ApplicationClient                       - Sending
StopCluster request to JobManager.2017-09-07 09:50:55,007 INFO 
org.apache.flink.yarn.ApplicationClient                       - Sending
StopCluster request to JobManager.2017-09-07 09:50:56,026 INFO 
org.apache.flink.yarn.ApplicationClient                       - Sending
StopCluster request to JobManager.2017-09-07 09:50:57,047 INFO 
org.apache.flink.yarn.ApplicationClient                       - Sending
StopCluster request to JobManager.2017-09-07 09:50:58,067 INFO 
org.apache.flink.yarn.ApplicationClient                       - Sending
StopCluster request to JobManager.2017-09-07 09:50:59,087 INFO 
org.apache.flink.yarn.ApplicationClient                       - Sending
StopCluster request to JobManager.2017-09-07 09:51:00,107 INFO 
org.apache.flink.yarn.ApplicationClient                       - Sending
StopCluster request to JobManager.2017-09-07 09:51:01,127 INFO 
org.apache.flink.yarn.ApplicationClient                       - Sending
StopCluster request to JobManager.2017-09-07 09:51:01,944 WARN 
org.apache.flink.yarn.YarnClusterClient                       - Error while
stopping YARN cluster.java.util.concurrent.TimeoutException: Futures timed
out after [10000 milliseconds]        at
scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)       
at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:153)       
at scala.concurrent.Await$$anonfun$ready$1.apply(package.scala:169)       
at scala.concurrent.Await$$anonfun$ready$1.apply(package.scala:169)       
at
scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
       
at scala.concurrent.Await$.ready(package.scala:169)        at
scala.concurrent.Await.ready(package.scala)        at
org.apache.flink.yarn.YarnClusterClient.shutdownCluster(YarnClusterClient.java:372)
       
at
org.apache.flink.yarn.YarnClusterClient.finalizeCluster(YarnClusterClient.java:342)
       
at
org.apache.flink.client.program.ClusterClient.shutdown(ClusterClient.java:208)  
     
at
org.apache.flink.yarn.YarnClusterClient$ClientShutdownHook.run(YarnClusterClient.java:446)2017-09-07
09:51:01,946 INFO  org.apache.flink.yarn.YarnClusterClient                      
- Deleted Yarn properties file at /tmp/.yarn-properties-bistel2017-09-07
09:51:01,946 INFO  org.apache.flink.yarn.YarnClusterClient                      
- Deleting files in
hdfs://flink-01:9000/user/bistel/.flink/application_1504745288687_00012017-09-07
09:51:02,146 INFO  org.apache.flink.yarn.ApplicationClient                      
- Sending StopCluster request to JobManager.2017-09-07 09:51:02,490 INFO 
org.apache.flink.yarn.YarnClusterClient                       - Application
application_1504745288687_0001 finished with state RUNNING and final state
UNDEFINED at 02017-09-07 09:51:02,490 INFO 
org.apache.flink.yarn.YarnClusterClient                       - YARN Client
is shutting down2017-09-07 09:51:02,598 INFO 
org.apache.flink.yarn.ApplicationClient                       - Stopped
Application client.2017-09-07 09:51:02,599 INFO 
org.apache.flink.yarn.ApplicationClient                       - Disconnect
from JobManager null.2017-09-07 09:51:02,633 INFO 
akka.remote.RemoteActorRefProvider$RemotingTerminator         - Shutting
down remote daemon.2017-09-07 09:51:02,635 INFO 
akka.remote.RemoteActorRefProvider$RemotingTerminator         - Remote
daemon shut down; proceeding with flushing remote transports.2017-09-07
09:51:02,651 INFO  akka.remote.RemoteActorRefProvider$RemotingTerminator        
- Remoting shut down.[bistel@flink-01 ~]$ jps1888 ResourceManager2000
NodeManager2433 NameNode2546 DataNode2754 SecondaryNameNode3143 Jps1724
QuorumPeerMain[bistel@flink-02 ~]$ jps2018 Jps1721 NodeManager1881
DataNode1515 QuorumPeerMain[bistel@flink-03 ~]$ jps1521 QuorumPeerMain2054
YarnApplicationMasterRunner1724 NodeManager1885 DataNode2142
Jps[bistel@flink-04 ~]$ jps2090 Jps1515 QuorumPeerMain1789 NodeManager1950
DataNode
Nico Kruber wrote
> I looked at the commit you cherry-picked and nothing in there explains the
> error you got. This rather sounds like something might be mixed up between
> (remaining artefacts of) flink 1.3 and 1.2.Can you verify that nothing of
> your flink 1.3 tests remains, e.g. running JobManager or TaskManager
> instances? Also that you're not accidentally running the yarn-session.sh
> script of 1.3?NicoOn Wednesday, 6 September 2017 06:36:42 CEST Sunny Yun
> wrote:> Hi,> > Using flink 1.2.0, I faced to issue>
> https://issues.apache.org/jira/browse/FLINK-6117>
> https://issues.apache.org/jira/browse/FLINK-6117.> This issue is fixed at
> version 1.3.0. But I have some reason to trying to> find out work around.>
> > I did,> 1. change source according to>
> https://github.com/apache/flink/commit/eef85e095a8a0e4c4553631b74ba7b9f173ce>
> bf0 2. replace $FLINK_HOME/lib/flink-dist_2.11-1.2.0.jar> 3. set
> flink-conf.yaml "zookeeper.sasl.disable: true"> 4. run yarn-session.sh> >
> > Original problem-Authentication failed- seems to be passed.> But I got
> this error,> > Exception in thread "main" java.lang.RuntimeException:
> Failed to retrieve> JobManager address>         at>
> org.apache.flink.client.program.ClusterClient.getJobManagerAddress(ClusterCl>
> ient.java:248) at>
> org.apache.flink.yarn.cli.FlinkYarnSessionCli.run(FlinkYarnSessionCli.java:6>
> 27) at>
> org.apache.flink.yarn.cli.FlinkYarnSessionCli$1.call(FlinkYarnSessionCli.jav>
> a:476) at>
> org.apache.flink.yarn.cli.FlinkYarnSessionCli$1.call(FlinkYarnSessionCli.jav>
> a:473) at>
> org.apache.flink.runtime.security.HadoopSecurityContext$1.run(HadoopSecurity>
> Context.java:43) at java.security.AccessController.doPrivileged(Native>
> Method) at javax.security.auth.Subject.doAs(Subject.java:422)>         at>
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.ja>
> va:1656) at>
> org.apache.flink.runtime.security.HadoopSecurityContext.runSecured(HadoopSec>
> urityContext.java:40) at>
> org.apache.flink.yarn.cli.FlinkYarnSessionCli.main(FlinkYarnSessionCli.java:>
> 473) Caused by:>
> org.apache.flink.runtime.leaderretrieval.LeaderRetrievalException: Could>
> not retrieve the leader address and leader session ID.>         at>
> org.apache.flink.runtime.util.LeaderRetrievalUtils.retrieveLeaderConnectionI>
> nfo(LeaderRetrievalUtils.java:175) at>
> org.apache.flink.client.program.ClusterClient.getJobManagerAddress(ClusterCl>
> ient.java:242) ... 9 more> Caused by:
> java.util.concurrent.TimeoutException: Futures timed out after> [60000
> milliseconds]>         at>
> scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)>        
> at>
> scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)>       
>  
> at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:190)> at>
> scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scal>
> a:53) at scala.concurrent.Await$.result(package.scala:190)>         at
> scala.concurrent.Await.result(package.scala)>         at>
> org.apache.flink.runtime.util.LeaderRetrievalUtils.retrieveLeaderConnectionI>
> nfo(LeaderRetrievalUtils.java:173) ... 10 more> > > I believe related
> setting(flink, hadoop, zookeeper) is correct. Because> yarn-session works
> smoothly with flink 1.3.2 in same environment.> > Does anyone have any
> inspiration for this error message?> > Thanks.> > ᐧsignature.asc (201
> bytes)
> <http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/attachment/15426/0/signature.asc>





--
Sent from: http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/

Reply via email to