Hello,
I'm running Accumulo 1.5.2, trying to test out the GeoMesa
<http://www.geomesa.org/2014/05/28/geomesa-quickstart/> family of
spatio-temporal iterators using their quickstart demonstration tool. I
think I'm not making progress due to my Accumulo setup, though, so can
someone validate that all looks good from here?
start-all.sh output:
hduser@accumulo:~$ $ACCUMULO_HOME/bin/start-all.sh
Starting monitor on localhost
Starting tablet servers .... done
Starting tablet server on localhost
2015-01-05 21:37:18,523 [server.Accumulo] INFO : Attempting to talk to zookeeper
2015-01-05 21:37:18,772 [server.Accumulo] INFO : Zookeeper connected
and initialized, attemping to talk to HDFS
2015-01-05 21:37:19,028 [server.Accumulo] INFO : Connected to HDFS
Starting master on localhost
Starting garbage collector on localhost
Starting tracer on localhost
hduser@accumulo:~$
I do believe my HDFS is set up correctly:
hduser@accumulo:/home/ubuntu/geomesa-quickstart$ hadoop fs -ls /accumulo
Found 5 items
drwxrwxrwx - hduser supergroup 0 2014-12-10 01:04
/accumulo/instance_id
drwxrwxrwx - hduser supergroup 0 2015-01-05 21:22 /accumulo/recovery
drwxrwxrwx - hduser supergroup 0 2015-01-05 20:14 /accumulo/tables
drwxrwxrwx - hduser supergroup 0 2014-12-10 01:04 /accumulo/version
drwxrwxrwx - hduser supergroup 0 2014-12-10 01:05 /accumulo/wal
However, when I check the Accumulo monitor logs, I see these errors
post-startup:
java.io.IOException: Mkdirs failed to create directory
/accumulo/recovery/15664488-bd10-4d8d-9584-f88d8595a07c/part-r-00000
java.io.IOException: Mkdirs failed to create directory
/accumulo/recovery/15664488-bd10-4d8d-9584-f88d8595a07c/part-r-00000
at org.apache.hadoop.io.MapFile$Writer.<init>(MapFile.java:264)
at org.apache.hadoop.io.MapFile$Writer.<init>(MapFile.java:103)
at
org.apache.accumulo.server.tabletserver.log.LogSorter$LogProcessor.writeBuffer(LogSorter.java:196)
at
org.apache.accumulo.server.tabletserver.log.LogSorter$LogProcessor.sort(LogSorter.java:166)
at
org.apache.accumulo.server.tabletserver.log.LogSorter$LogProcessor.process(LogSorter.java:89)
at
org.apache.accumulo.server.zookeeper.DistributedWorkQueue$1.run(DistributedWorkQueue.java:101)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at
org.apache.accumulo.trace.instrument.TraceRunnable.run(TraceRunnable.java:47)
at
org.apache.accumulo.core.util.LoggingRunnable.run(LoggingRunnable.java:34)
at java.lang.Thread.run(Thread.java:745)
I don't really understand - I started accumulo as the hduser, which is the
same user that has access to the HDFS directory /accumulo/recovery, and it
looks like the directory was created actually, except for the last
directory (part-r-0000):
hduser@accumulo:~$ hadoop fs -ls /accumulo0/recovery/
Found 1 items
drwxr-xr-x - hduser supergroup 0 2015-01-05 22:11
/accumulo/recovery/87fb7aac-0274-4aea-8014-9d53dbbdfbbc
I'm not out of physical disk space:
hduser@accumulo:~$ df -h
Filesystem Size Used Avail Use% Mounted on
/dev/xvda1 1008G 8.5G 959G 1% /
What could be going on here? Any ideas on something simple I could have
missed?
Thanks,
Mike