[ https://issues.apache.org/jira/browse/AMBARI-22248?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Jayush Luniya reopened AMBARI-22248: ------------------------------------ Reopening if incase this needs to be in 2.6.1 > HBase default.rootdir config results in deploy failure if value is not > overriden > -------------------------------------------------------------------------------- > > Key: AMBARI-22248 > URL: https://issues.apache.org/jira/browse/AMBARI-22248 > Project: Ambari > Issue Type: Bug > Components: ambari-server > Affects Versions: 2.0.0 > Reporter: Siddharth Wagle > Assignee: Siddharth Wagle > Priority: Critical > Fix For: 2.6.1 > > Attachments: AMBARI-22248.patch > > > The default value for hbase.rootdir is set as below, due to which hbase fails > to start with below exception. > hbase.rootdir=hdfs://localhost:8020/apps/hbase/data > {noformat} > 2017-10-16 17:23:06,761 FATAL > [ctr-e134-1499953498516-228160-01-000003:16000.activeMasterManager] > master.HMaster: Unhandled exception. Starting shutdown. > java.net.ConnectException: Call From > ctr-e134-1499953498516-228160-01-000003.hwx.site/172.27.63.128 to > localhost:8020 failed on connection exception: java.net.ConnectException: > Connection refused; For more details see: > http://wiki.apache.org/hadoop/ConnectionRefused > at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native > Method) > at > sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) > at > sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) > at java.lang.reflect.Constructor.newInstance(Constructor.java:423) > at org.apache.hadoop.net.NetUtils.wrapWithMessage(NetUtils.java:801) > at org.apache.hadoop.net.NetUtils.wrapException(NetUtils.java:732) > at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1558) > at org.apache.hadoop.ipc.Client.call(Client.java:1498) > at org.apache.hadoop.ipc.Client.call(Client.java:1398) > at > org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:233) > at com.sun.proxy.$Proxy16.setSafeMode(Unknown Source) > at > org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.setSafeMode(ClientNamenodeProtocolTranslatorPB.java:718) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:498) > at > org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:291) > at > org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:203) > at > org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:185) > at com.sun.proxy.$Proxy17.setSafeMode(Unknown Source) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:498) > at > org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:280) > at com.sun.proxy.$Proxy18.setSafeMode(Unknown Source) > at org.apache.hadoop.hdfs.DFSClient.setSafeMode(DFSClient.java:2669) > at > org.apache.hadoop.hdfs.DistributedFileSystem.setSafeMode(DistributedFileSystem.java:1359) > at > org.apache.hadoop.hdfs.DistributedFileSystem.setSafeMode(DistributedFileSystem.java:1343) > at org.apache.hadoop.hbase.util.FSUtils.isInSafeMode(FSUtils.java:555) > at > org.apache.hadoop.hbase.util.FSUtils.waitOnSafeMode(FSUtils.java:1001) > at > org.apache.hadoop.hbase.master.MasterFileSystem.checkRootDir(MasterFileSystem.java:455) > at > org.apache.hadoop.hbase.master.MasterFileSystem.createInitialFileSystemLayout(MasterFileSystem.java:162) > at > org.apache.hadoop.hbase.master.MasterFileSystem.<init>(MasterFileSystem.java:142) > {noformat} > Since Ambari only allows HBase to be deployed along with HDFS through the > dependency in common-services definition of HBase, we can set the root > directory as a relative path while the hadoop-common code figures out the FS > URI on its own. -- This message was sent by Atlassian JIRA (v6.4.14#64029)