Did you verify that Hadoop is really up and healthy? Look at the Hadoop monitor pages and confirm that you can use the Hadoop cli to navigate around? You may also need to update the accumulo configuration files / env to match your configuration.
You might what to look at using https://github.com/apache/fluo-uno as a quick way to stand up an instance for testing – and that might give to additional insights. From: Christine Buss <christine.buss...@gmx.de> Sent: Wednesday, July 7, 2021 11:21 AM To: user@accumulo.apache.org Subject: Hadoop ConnectException Hi, I am using: Java 11 Ubuntu 20.04.2 Hadoop 3.3.1 Zookeeper 3.7.0 Accumulo 2.0.1 I followed the instructions here: https://hadoop.apache.org/docs/stable/hadoop-project-dist/hadoop-common/SingleCluster.html and edited etc/hadoop/hadoop-env.sh, etc/hadoop/core-site.xml, etc/hadoop/hdfs-site.xml accordingly. 'ssh localhost' works without a passphrase. Then I started Zookeper, start-dfs.sh and start-yarn.sh: christine@centauri:~$ ./zookeeper-3.4.9/bin/zkServer.sh start ZooKeeper JMX enabled by default Using config: /home/christine/zookeeper-3.4.9/bin/../conf/zoo.cfg Starting zookeeper ... STARTED christine@centauri:~$ ./hadoop-3.3.1/sbin/start-dfs.sh Starting namenodes on [localhost] Starting datanodes Starting secondary namenodes [centauri] centauri: Warning: Permanently added 'centauri,2003:d4:771c:3b00:7223:40a1:4c07:7c7b' (ECDSA) to the list of known hosts. christine@centauri:~$ ./hadoop-3.3.1/sbin/start-yarn.sh Starting resourcemanager Starting nodemanagers christine@centauri:~$ jps 3921 Jps 2387 QuorumPeerMain 3171 SecondaryNameNode 3732 NodeManager 2955 DataNode 3599 ResourceManager BUT when running 'accumulo init' I get this Error: hristine@centauri:~$ ./accumulo-2.0.1/bin/accumulo init OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release. 2021-07-07 15:59:05,590 [conf.SiteConfiguration] INFO : Found Accumulo configuration on classpath at /home/christine/accumulo-2.0.1/conf/accumulo.properties 2021-07-07 15:59:08,460 [fs.VolumeManagerImpl] WARN : dfs.datanode.synconclose set to false in hdfs-site.xml: data loss is possible on hard system reset or power loss 2021-07-07 15:59:08,461 [init.Initialize] INFO : Hadoop Filesystem is hdfs://localhost:9000 2021-07-07 15:59:08,461 [init.Initialize] INFO : Accumulo data dirs are [hdfs://localhost:8020/accumulo] 2021-07-07 15:59:08,461 [init.Initialize] INFO : Zookeeper server is localhost:2181 2021-07-07 15:59:08,461 [init.Initialize] INFO : Checking if Zookeeper is available. If this hangs, then you need to make sure zookeeper is running 2021-07-07 15:59:08,938 [init.Initialize] ERROR: Fatal exception java.io.IOException: Failed to check if filesystem already initialized at org.apache.accumulo.server.init.Initialize.checkInit(Initialize.java:285) at org.apache.accumulo.server.init.Initialize.doInit(Initialize.java:323) at org.apache.accumulo.server.init.Initialize.execute(Initialize.java:991) at org.apache.accumulo.start.Main.lambda$execKeyword$0(Main.java:129) at java.base/java.lang.Thread.run(Thread.java:829) Caused by: java.net.ConnectException: Call From centauri/192.168.178.30 to localhost:8020 failed on connection exception: java.net.ConnectException: Connection refused; For more details see: http://wiki.apache.org/hadoop/ConnectionRefused at java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) at java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) at java.base/jdk.internal.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) at java.base/java.lang.reflect.Constructor.newInstance(Constructor.java:490) at org.apache.hadoop.net.NetUtils.wrapWithMessage(NetUtils.java:913) at org.apache.hadoop.net.NetUtils.wrapException(NetUtils.java:828) at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1577) at org.apache.hadoop.ipc.Client.call(Client.java:1519) at org.apache.hadoop.ipc.Client.call(Client.java:1416) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:242) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:129) at com.sun.proxy.$Proxy18.getFileInfo(Unknown Source) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.getFileInfo(ClientNamenodeProtocolTranslatorPB.java:965) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:566) at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:422) at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeMethod(RetryInvocationHandler.java:165) at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invoke(RetryInvocationHandler.java:157) at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeOnce(RetryInvocationHandler.java:95) at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:359) at com.sun.proxy.$Proxy19.getFileInfo(Unknown Source) at org.apache.hadoop.hdfs.DFSClient.getFileInfo(DFSClient.java:1731) at org.apache.hadoop.hdfs.DistributedFileSystem$29.doCall(DistributedFileSystem.java:1752) at org.apache.hadoop.hdfs.DistributedFileSystem$29.doCall(DistributedFileSystem.java:1749) at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) at org.apache.hadoop.hdfs.DistributedFileSystem.getFileStatus(DistributedFileSystem.java:1764) at org.apache.hadoop.fs.FileSystem.exists(FileSystem.java:1760) at org.apache.accumulo.server.fs.VolumeManagerImpl.exists(VolumeManagerImpl.java:254) at org.apache.accumulo.server.init.Initialize.isInitialized(Initialize.java:860) at org.apache.accumulo.server.init.Initialize.checkInit(Initialize.java:280) ... 4 more Caused by: java.net.ConnectException: Connection refused at java.base/sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) at java.base/sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:777) at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206) at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:586) at org.apache.hadoop.ipc.Client$Connection.setupConnection(Client.java:701) at org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:822) at org.apache.hadoop.ipc.Client$Connection.access$3800(Client.java:414) at org.apache.hadoop.ipc.Client.getConnection(Client.java:1647) at org.apache.hadoop.ipc.Client.call(Client.java:1463) ... 28 more 2021-07-07 15:59:08,944 [start.Main] ERROR: Thread 'init' died. java.lang.RuntimeException: java.io.IOException: Failed to check if filesystem already initialized at org.apache.accumulo.server.init.Initialize.execute(Initialize.java:997) at org.apache.accumulo.start.Main.lambda$execKeyword$0(Main.java:129) at java.base/java.lang.Thread.run(Thread.java:829) Caused by: java.io.IOException: Failed to check if filesystem already initialized at org.apache.accumulo.server.init.Initialize.checkInit(Initialize.java:285) at org.apache.accumulo.server.init.Initialize.doInit(Initialize.java:323) at org.apache.accumulo.server.init.Initialize.execute(Initialize.java:991) ... 2 more Caused by: java.net.ConnectException: Call From centauri/192.168.178.30 to localhost:8020 failed on connection exception: java.net.ConnectException: Connection refused; For more details see: http://wiki.apache.org/hadoop/ConnectionRefused at java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) at java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) at java.base/jdk.internal.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) at java.base/java.lang.reflect.Constructor.newInstance(Constructor.java:490) at org.apache.hadoop.net.NetUtils.wrapWithMessage(NetUtils.java:913) at org.apache.hadoop.net.NetUtils.wrapException(NetUtils.java:828) at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1577) at org.apache.hadoop.ipc.Client.call(Client.java:1519) at org.apache.hadoop.ipc.Client.call(Client.java:1416) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:242) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:129) at com.sun.proxy.$Proxy18.getFileInfo(Unknown Source) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.getFileInfo(ClientNamenodeProtocolTranslatorPB.java:965) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:566) at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:422) at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeMethod(RetryInvocationHandler.java:165) at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invoke(RetryInvocationHandler.java:157) at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeOnce(RetryInvocationHandler.java:95) at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:359) at com.sun.proxy.$Proxy19.getFileInfo(Unknown Source) at org.apache.hadoop.hdfs.DFSClient.getFileInfo(DFSClient.java:1731) at org.apache.hadoop.hdfs.DistributedFileSystem$29.doCall(DistributedFileSystem.java:1752) at org.apache.hadoop.hdfs.DistributedFileSystem$29.doCall(DistributedFileSystem.java:1749) at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) at org.apache.hadoop.hdfs.DistributedFileSystem.getFileStatus(DistributedFileSystem.java:1764) at org.apache.hadoop.fs.FileSystem.exists(FileSystem.java:1760) at org.apache.accumulo.server.fs.VolumeManagerImpl.exists(VolumeManagerImpl.java:254) at org.apache.accumulo.server.init.Initialize.isInitialized(Initialize.java:860) at org.apache.accumulo.server.init.Initialize.checkInit(Initialize.java:280) ... 4 more Caused by: java.net.ConnectException: Connection refused at java.base/sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) at java.base/sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:777) at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206) at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:586) at org.apache.hadoop.ipc.Client$Connection.setupConnection(Client.java:701) at org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:822) at org.apache.hadoop.ipc.Client$Connection.access$3800(Client.java:414) at org.apache.hadoop.ipc.Client.getConnection(Client.java:1647) at org.apache.hadoop.ipc.Client.call(Client.java:1463) ... 28 more I am not able to find the mistake. I found similar questions on Stackoverflow, but none of them solved my problem. Thanks in advance for any idea.