Hi!

1) Do you use properly configured Zookeeper in your cluster?

2) I suspect (at least, it helped me greatly in CDH) that you should
install HDFS NFS service on the host with Zeppelin

Without NFS & (may be) ZooKeeper, your Hive tries to interpret "ha-cluster"
as DNS name in your network

On Wed, Jun 1, 2016 at 12:38 PM, Meilong Huang <ims...@gmail.com> wrote:

> Hi,
>
> To access hive tables with spark sql, I copied hive-seite.xml to zeppelin
> configuration folder. It works fine with non-ha cluster. But when I'm
> accessing hive tables in HA cluster, there is an unknownHost exception, do
> you have any idea? Many thanks.
>
> 2016-06-01 11:15:51,000 INFO  [dispatcher-event-loop-2]
> storage.BlockManagerInfo (Logging.scala:logInfo(58)) - Added
> broadcast_2_piece0 in memory on localhost:57979 (size: 42.3 KB, free: 511.5
> MB)
> 2016-06-01 11:15:51,003 INFO  [pool-2-thread-3] spark.SparkContext
> (Logging.scala:logInfo(58)) - Created broadcast 2 from take at
> NativeMethodAccessorImpl.java:-2
> 2016-06-01 11:15:51,922 ERROR [pool-2-thread-3] scheduler.Job
> (Job.java:run(182)) - Job failed
> org.apache.zeppelin.interpreter.InterpreterException:
> java.lang.reflect.InvocationTargetException
>         at
> org.apache.zeppelin.spark.ZeppelinContext.showDF(ZeppelinContext.java:301)
>         at
> org.apache.zeppelin.spark.SparkSqlInterpreter.interpret(SparkSqlInterpreter.java:144)
>         at
> org.apache.zeppelin.interpreter.ClassloaderInterpreter.interpret(ClassloaderInterpreter.java:57)
>         at
> org.apache.zeppelin.interpreter.LazyOpenInterpreter.interpret(LazyOpenInterpreter.java:93)
>         at
> org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:300)
>         at org.apache.zeppelin.scheduler.Job.run(Job.java:169)
>         at
> org.apache.zeppelin.scheduler.FIFOScheduler$1.run(FIFOScheduler.java:134)
>         at
> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
>         at java.util.concurrent.FutureTask.run(FutureTask.java:262)
>         at
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:178)
>         at
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:292)
>         at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>         at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>         at java.lang.Thread.run(Thread.java:745)
> Caused by: java.lang.reflect.InvocationTargetException
>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>         at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>         at java.lang.reflect.Method.invoke(Method.java:606)
>         at
> org.apache.zeppelin.spark.ZeppelinContext.showDF(ZeppelinContext.java:297)
>         ... 13 more
> Caused by: java.net.UnknownHostException: Invalid host name: local host
> is: (unknown); destination host is: "ha-cluster":8020;
> java.net.UnknownHostException; For more details see:
> http://wiki.apache.org/hadoop/UnknownHost
>         at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native
> Method)
>         at
> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)
>         at
> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
>         at java.lang.reflect.Constructor.newInstance(Constructor.java:526)
>         at
> org.apache.hadoop.net.NetUtils.wrapWithMessage(NetUtils.java:791)
>         at org.apache.hadoop.net.NetUtils.wrapException(NetUtils.java:743)
>         at org.apache.hadoop.ipc.Client$Connection.<init>(Client.java:402)
>         at org.apache.hadoop.ipc.Client.getConnection(Client.java:1511)
>         at org.apache.hadoop.ipc.Client.call(Client.java:1438)
>         at org.apache.hadoop.ipc.Client.call(Client.java:1399)
>         at
> org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:232)
>         at com.sun.proxy.$Proxy29.getFileInfo(Unknown Source)
>         at
> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.getFileInfo(ClientNamenodeProtocolTranslatorPB.java:752)
>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>         at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>         at java.lang.reflect.Method.invoke(Method.java:606)
>         at
> org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:187)
>         at
> org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:102)
>         at com.sun.proxy.$Proxy30.getFileInfo(Unknown Source)
>         at
> org.apache.hadoop.hdfs.DFSClient.getFileInfo(DFSClient.java:1988)
>         at
> org.apache.hadoop.hdfs.DistributedFileSystem$18.doCall(DistributedFileSystem.java:1118)
>         at
> org.apache.hadoop.hdfs.DistributedFileSystem$18.doCall(DistributedFileSystem.java:1114)
>         at
> org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
>
>
>
> hadoop conf for HA cluster:
>
> core-site.xml
>
> <configuration>
>   <property>
>      <name>master_hostname</name>
>      <value>MasterHostName</value>
>   </property>
>   <property>
>      <name>fs.defaultFS</name>
>      <value>hdfs://ha-cluster</value>
>   </property>
>
>   <property>
>      <name>hadoop.home</name>
>      <value>/usr/lib/hadoop-current</value>
>   </property>
>
>   <property>
>     <name>fs.trash.interval</name>
>     <value>60</value>
>   </property>
>   <property>
>     <name>fs.trash.checkpoint.interval</name>
>     <value>30</value>
>   </property>
>
>   <property>
>     <name>hadoop.tmp.dir</name>
>     <value>/mnt/disk1/hadoop/tmp</value>
>     <description>A base for other temporary directories.</description>
>   </property>
>
>   <property>
>     <name>ha.zookeeper.quorum</name>
>     <value>header-1:2181,header-2:2181,header-3:2181</value>
>   </property>
>
>   <property>
>   <name>hadoop.proxyuser.hadoop.groups</name>
>     <value>*</value>
>     </property>
>
>   <property>
>     <name>hadoop.proxyuser.hadoop.hosts</name>
>     <value>*</value>
>   </property>
>
>   <property>
>     <name>hadoop.proxyuser.hue.groups</name>
>     <value>*</value>
>     <description>User proxy groups for hue.</description>
>   </property>
>
>   <property>
>     <name>hadoop.proxyuser.hue.hosts</name>
>     <value>*</value>
>     <description>User proxy hosts for hue.</description>
>   </property>
> </configuration>
>
>
> hdfs-site.xml
>
>
> <configuration>
>   <property>
>     <name>dfs.nameservices</name>
>     <value>ha-cluster</value>
>   </property>
>
>   <property>
>     <name>dfs.ha.namenodes.cluster</name>
>     <value>nn1,nn2</value>
>   </property>
>
>   <property>
>     <name>dfs.namenode.rpc-address.cluster.nn1</name>
>     <value>header-1:8020</value>
>   </property>
>
>   <property>
>     <name>dfs.namenode.rpc-address.cluster.nn2</name>
>     <value>header-2:8020</value>
>   </property>
>
>   <property>
>     <name>dfs.namenode.http-address.cluster.nn1</name>
>     <value>header-1:50070</value>
>   </property>
>
>   <property>
>     <name>dfs.namenode.http-address.cluster.nn2</name>
>     <value>header-2:50070</value>
>   </property>
>
>   <property>
>     <name>dfs.namenode.shared.edits.dir</name>
>
> <value>qjournal://header-1:8485;header-2:8485;header-3:8485/cluster</value>
>   </property>
>
>   <property>
>     <name>dfs.client.failover.proxy.provider.cluster</name>
>
> <value>org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider</value>
>   </property>
>
>   <property>
>     <name>dfs.ha.fencing.methods</name>
>     <value>shell(/bin/true)</value>
>   </property>
>
>   <property>
>     <name>dfs.ha.automatic-failover.enabled</name>
>     <value>true</value>
>   </property>
>
>   <property>
>     <name>ha.zookeeper.quorum</name>
>     <value>header-1:2181,header-2:2181,header-3:2181</value>
>   </property>
>
>   <property>
>     <name>dfs.journalnode.edits.dir</name>
>     <value>/mnt/disk1/hdfs/journal</value>
>   </property>
>
>   <property>
>     <name>dfs.replication</name>
>     <value>2</value>
>         <description>Default block replication.
>     The actual number of replications can be specified when the file is
> created.
>     The default is used if replication is not specified in create time.
>     </description>
>   </property>
>
>   <property>
>      <name>dfs.permissions.superusergroup</name>
>      <value>hadoop</value>
>      <description>The name of the group of super-users.</description>
>   </property>
>
>   <property>
>     <name>dfs.permissions.enabled</name>
>     <value>false</value>
>   </property>
>
>   <property>
>     <name>dfs.namenode.name.dir</name>
>     <value>file:///mnt/disk1/hdfs/name</value>
>   </property>
>   <property>
>     <name>dfs.namenode.edits.dir</name>
>     <value>file:///mnt/disk1/hdfs/edits</value>
>   </property>
>   <property>
>     <name>dfs.namenode.resource.du.reserved</name>
>     <value>1073741824</value>
>   </property>
>
>   <property>
>     <name>dfs.namenode.handler.count</name>
>     <value>10</value>
>   </property>
>   <property>
>     <name>dfs.support.append</name>
>     <value>true</value>
>     </property>
>   <property>
>     <name>dfs.http.address</name>
>     <value>0.0.0.0:50070</value>
>   </property>
>   <property>
>     <name>dfs.namenode.datanode.registration.ip-hostname-check</name>
>     <value>false</value>
>   </property>
>
> <!-- DATANODE -->
>   <property>
>     <name>dfs.datanode.data.dir</name>
>     <value>
>         file:///mnt/disk1
>     </value>
>   </property>
>   <property>
>     <name>dfs.datanode.data.dir.perm</name>
>     <value>770</value>
>   </property>
>   <property>
>     <name>dfs.datanode.du.reserved</name>
>     <value>1073741824</value>
>   </property>
>   <!-- SECONDARYNAMENODE -->
>   <property>
>     <name>dfs.namenode.checkpoint.dir</name>
>     <value>
>       file:///mnt/disk1/hdfs/namesecondary
>     </value>
>   </property>
>
> </configuration>
>
>
>
> yarn-site.xml
>
> <configuration>
>
>  <property>
>    <name>yarn.resourcemanager.ha.enabled</name>
>    <value>true</value>
>  </property>
>
>   <property>
>     <name>yarn.resourcemanager.ha.automatic-failover.zk-base-path</name>
>     <value>/yarn-leader-election</value>
>   </property>
>
>   <property>
>     <name>yarn.resourcemanager.ha.automatic-failover.enabled</name>
>     <value>true</value>
>   </property>
>
>  <property>
>    <name>yarn.resourcemanager.cluster-id</name>
>    <value>ha-cluster</value>
>  </property>
>
>  <property>
>    <name>yarn.resourcemanager.ha.rm-ids</name>
>    <value>rm1,rm2</value>
>  </property>
>
>  <property>
>    <name>yarn.resourcemanager.hostname.rm1</name>
>    <value>header-1</value>
>  </property>
>
>  <property>
>    <name>yarn.resourcemanager.hostname.rm2</name>
>    <value>header-2</value>
>  </property>
>  <property>
>    <name>yarn.resourcemanager.zk-address</name>
>    <value>header-1:2181,header-2:2181,header-3:2181</value>
>  </property>
>
>   <property>
>     <name>yarn.web-proxy.address</name>
>     <value>${master_hostname}:20888</value>
>   </property>
>
>   <property>
>     <name>yarn.resourcemanager.resource-tracker.address.rm1</name>
>     <value>header-1:8025</value>
>   </property>
>
>   <property>
>     <name>yarn.resourcemanager.resource-tracker.address.rm2</name>
>     <value>header-2:8025</value>
>   </property>
>
>   <property>
>     <name>yarn.resourcemanager.address.rm1</name>
>     <value>header-1:8032</value>
>   </property>
>
>   <property>
>     <name>yarn.resourcemanager.address.rm2</name>
>     <value>header-2:8032</value>
>   </property>
>
>   <property>
>     <name>yarn.resourcemanager.scheduler.address.rm1</name>
>     <value>header-1:8030</value>
>   </property>
>
>   <property>
>     <name>yarn.resourcemanager.scheduler.address.rm2</name>
>     <value>header-2:8030</value>
>   </property>
>
>   <property>
>     <name>yarn.nodemanager.aux-services</name>
>     <value>mapreduce_shuffle,</value>
>   </property>
>
>   <property>
>     <name>yarn.log-aggregation-enable</name>
>     <value>false</value>
>   </property>
>
>   <property>
>     <description>How long to keep aggregation logs before deleting them.
>  -1 disables.
>     Be careful set this too small and you will spam the name
> node.</description>
>     <name>yarn.log-aggregation.retain-seconds</name>
>     <value>86400</value>
>   </property>
>
> <!-- hdfs dir -->
>   <property>
>     <description>Where to aggregate logs to.</description>
>     <name>yarn.nodemanager.remote-app-log-dir</name>
>     <value>/tmp/logs</value>
>   </property>
>
>   <property>
>     <name>yarn.dispatcher.exit-on-error</name>
>     <value>true</value>
>   </property>
>
>   <property>
>     <name>yarn.nodemanager.local-dirs</name>
>     <value>file:///mnt/disk1/yarn</value>
>     <final>true</final>
>   </property>
>
>   <property>
>     <description>Where to store container logs.</description>
>     <name>yarn.nodemanager.log-dirs</name>
>     <value>file:///mnt/disk1/log/hadoop-yarn/containers</value>
>   </property>
>
>   <property>
>     <description>Classpath for typical applications.</description>
>      <name>yarn.application.classpath</name>
>      <value>
>         $HADOOP_CONF_DIR,
>         $HADOOP_COMMON_HOME/share/hadoop/common/*,
>         $HADOOP_COMMON_HOME/share/hadoop/common/lib/*,
>         $HADOOP_HDFS_HOME/share/hadoop/hdfs/*,
>         $HADOOP_HDFS_HOME/share/hadoop/hdfs/lib/*,
>         $HADOOP_YARN_HOME/share/hadoop/yarn/*,
>         $HADOOP_YARN_HOME/share/hadoop/yarn/lib/*,
>         /opt/apps/extra-jars/*
>      </value>
>   </property>
>
>   ...
> </configuration>
>
>
>

Reply via email to