Hi, Why in this new message there is a different host? Previously code was trying to connect to “master:8020” and now it is “startdt”? If you were able to change this host somehow between runs, I guess you should be also able to set it to correct one.
Piotrek > On 23 Oct 2017, at 09:11, 邓俊华 <dengjun...@startdt.com> wrote: > > Hi, > > Thanks for your replay! I have been block in this for several days. > And I have double checked that there are > hdfs-site.xml,core-site.xml,yarn-site.xml in YARN_CONF_DIR. But it is still > can't read hdfs namenode logical url. > > 2017-10-23 14:35:17,750 DEBUG > org.apache.flink.api.java.hadoop.mapred.utils.HadoopUtils - Cannot find > hdfs-default configuration file > 2017-10-23 14:35:17,750 DEBUG > org.apache.flink.api.java.hadoop.mapred.utils.HadoopUtils - Cannot find > hdfs-site configuration file > 2017-10-23 14:35:17,751 DEBUG > org.apache.flink.api.java.hadoop.mapred.utils.HadoopUtils - Adding > /data/hadoop-2.7.3/etc/hadoop/core-site.xml to hadoop configuration > 2017-10-23 14:35:17,751 DEBUG > org.apache.flink.api.java.hadoop.mapred.utils.HadoopUtils - Adding > /data/hadoop-2.7.3/etc/hadoop/hdfs-site.xml to hadoop configuration > > 2017-10-23 14:35:19,887 DEBUG org.apache.hadoop.hdfs.BlockReaderLocal > - dfs.domain.socket.path = > 2017-10-23 14:35:19,952 ERROR > org.apache.flink.yarn.YarnApplicationMasterRunner - YARN > Application Master initialization failed > java.lang.IllegalArgumentException: java.net.UnknownHostException: startdt > at > org.apache.hadoop.security.SecurityUtil.buildTokenService(SecurityUtil.java:378) > at > org.apache.hadoop.hdfs.NameNodeProxies.createNonHAProxy(NameNodeProxies.java:310) > at > org.apache.hadoop.hdfs.NameNodeProxies.createProxy(NameNodeProxies.java:176) > at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:678) > at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:619) > at > org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:149) > ------------------------------------------------------------------ > 发件人:Piotr Nowojski <pi...@data-artisans.com> > 发送时间:2017年10月20日(星期五) 21:39 > 收件人:邓俊华 <dengjun...@startdt.com> > 抄 送:user <user@flink.apache.org> > 主 题:Re: flink can't read hdfs namenode logical url > > Hi, > > Please double check the content of config files in YARN_CONF_DIR and > HADOOP_CONF_DIR (the first one has a priority over the latter one) and that > they are pointing to correct files. > > Also check logs (WARN and INFO) for any relevant entries. > > Piotrek > > On 20 Oct 2017, at 06:07, 邓俊华 <dengjun...@startdt.com > <mailto:dengjun...@startdt.com>> wrote: > > hi, > > I start yarn-ssession.sh on yarn, but it can't read hdfs logical url. It > always connect to hdfs://master:8020 <hdfs://master:8020>, it should be 9000, > my hdfs defaultfs is hdfs://master <hdfs://master>. > I have config the YARN_CONF_DIR and HADOOP_CONF_DIR, it didn't work. > Is it a bug? i use flink-1.3.0-bin-hadoop27-scala_2.10 > > 2017-10-20 11:00:05,395 DEBUG org.apache.hadoop.ipc.Client > - IPC Client (1035144464) connection to > startdt/173.16.5.215:8020 from admin: closed2017-10-20 11:00:05,398 ERROR > org.apache.flink.yarn.YarnApplicationMasterRunner - YARN > Application Master initialization failedjava.net.ConnectException: Call From > spark3/173.16.5.216 to master:8020 failed on connection exception: > java.net.ConnectException: Connection refused; For more details see: > http://wiki.apache.org/hadoop/ConnectionRefused > <http://wiki.apache.org/hadoop/ConnectionRefused> at > sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) > at > sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57) > at > sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) > at java.lang.reflect.Constructor.newInstance(Constructor.java:526) at > org.apache.hadoop.net.NetUtils.wrapWithMessage(NetUtils.java:792) at > org.apache.hadoop.net.NetUtils.wrapException(NetUtils.java:732) at > org.apache.hadoop.ipc.Client.call(Client.java:1479) at > org.apache.hadoop.ipc.Client.call(Client.java:1412) at > org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:229) > at com.sun.proxy.$Proxy9.getFileInfo(Unknown Source) at > org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.getFileInfo(ClientNamenodeProtocolTranslatorPB.java:771) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:606) at > org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:191) > at > org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:102) > at com.sun.proxy.$Proxy10.getFileInfo(Unknown Source) at > org.apache.hadoop.hdfs.DFSClient.getFileInfo(DFSClient.java:2108) at > org.apache.hadoop.hdfs.DistributedFileSystem$22.doCall(DistributedFileSystem.java:1305) > at > org.apache.hadoop.hdfs.DistributedFileSystem$22.doCall(DistributedFileSystem.java:1301) > at > org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) > > > >