on all 5 nodes there is a conf/accumulo-site.xml with the same values for all 5.
someone is going to kill me when they find out that i left off a bracket or semi-colon somewhere... On Tue, Mar 18, 2014 at 12:02 PM, Josh Elser <[email protected]> wrote: > No, running `accumulo init` on a single host is sufficient. > > Is accumulo-site.xml consistent across all machines? > > > On 3/18/14, 11:57 AM, Benjamin Parrish wrote: > >> So here is the error from the tablet server... >> >> 2014-03-18 10:38:43,456 [client.ZooKeeperInstance] ERROR: unable obtain >> instance id at /accumulo/instance_id >> 2014-03-18 10:38:43,456 [tabletserver.TabletServer] ERROR: Uncaught >> exception in TabletServer.main, exiting >> java.lang.RuntimeException: Accumulo not initialized, there is no >> instance id at /accumulo/instance_id >> at >> org.apache.accumulo.core.client.ZooKeeperInstance.getInstanceIDFromHdfs( >> ZooKeeperInstance.java:295) >> at >> org.apache.accumulo.server.client.HdfsZooInstance._ >> getInstanceID(HdfsZooInstance.java:126) >> at >> org.apache.accumulo.server.client.HdfsZooInstance. >> getInstanceID(HdfsZooInstance.java:119) >> at >> org.apache.accumulo.server.conf.ZooConfiguration. >> getInstance(ZooConfiguration.java:55) >> at >> org.apache.accumulo.server.conf.ServerConfiguration.getZooConfiguration( >> ServerConfiguration.java:50) >> at >> org.apache.accumulo.server.conf.ServerConfiguration.getConfiguration( >> ServerConfiguration.java:104) >> at org.apache.accumulo.server.Accumulo.init(Accumulo.java:98) >> at >> org.apache.accumulo.server.tabletserver.TabletServer. >> main(TabletServer.java:3249) >> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >> at >> sun.reflect.NativeMethodAccessorImpl.invoke( >> NativeMethodAccessorImpl.java:57) >> at >> sun.reflect.DelegatingMethodAccessorImpl.invoke( >> DelegatingMethodAccessorImpl.java:43) >> at java.lang.reflect.Method.invoke(Method.java:606) >> at org.apache.accumulo.start.Main$1.run(Main.java:103) >> at java.lang.Thread.run(Thread.java:744) >> >> Do I need to run bin/accumulo init on every box in the cluster? >> >> >> On Tue, Mar 18, 2014 at 11:19 AM, Eric Newton <[email protected] >> <mailto:[email protected]>> wrote: >> >> Port numbers (for 1.5+) >> >> 4560 Accumulo monitor (for centralized log display) >> 9997 Tablet Server >> 9999 Master Server >> 12234 Accumulo Tracer >> 50091 Accumulo GC >> 50095 Accumulo HTTP monitor >> >> >> On Tue, Mar 18, 2014 at 11:04 AM, Benjamin Parrish >> <[email protected] <mailto:[email protected]>> >> >> wrote: >> >> First off, are there specific ports that need to be opened up >> for accumulo? I have hadoop operating without any issues as a 5 >> node cluster. Zookeeper seems to be operating with 2181, 3888, >> 2888 ports open. >> >> Here is some data from trying to get everything started and >> getting into the shell. I discluded the bash portion as Eric >> suggested because the mailing list rejected it for length and >> thinking it was spam. >> >> bin/start-all.sh >> >> [root@hadoop-node-1 zookeeper]# bash -x >> /usr/local/accumulo/bin/start-all.sh >> Starting monitor on hadoop-node-1 >> WARN : Max files open on hadoop-node-1 is 1024, recommend 65536 >> Starting tablet servers ....... done >> Starting tablet server on hadoop-node-3 >> Starting tablet server on hadoop-node-5 >> Starting tablet server on hadoop-node-2 >> Starting tablet server on hadoop-node-4 >> WARN : Max files open on hadoop-node-3 is 1024, recommend 65536 >> WARN : Max files open on hadoop-node-2 is 1024, recommend 65536 >> WARN : Max files open on hadoop-node-5 is 1024, recommend 65536 >> WARN : Max files open on hadoop-node-4 is 1024, recommend 65536 >> Java HotSpot(TM) 64-Bit Server VM warning: You have loaded >> library /usr/local/hadoop/lib/native/libhadoop.so.1.0.0 which >> might have disabled stack guard. The VM will try to fix the >> stack guard no >> w. >> It's highly recommended that you fix the library with 'execstack >> -c <libfile>', or link it with '-z noexecstack'. >> 2014-03-18 10:38:43,143 [util.NativeCodeLoader] WARN : Unable to >> load native-hadoop library for your platform... using >> builtin-java classes where applicable >> 2014-03-18 10:38:44,194 [server.Accumulo] INFO : Attempting to >> talk to zookeeper >> 2014-03-18 10:38:44,389 [server.Accumulo] INFO : Zookeeper >> connected and initialized, attemping to talk to HDFS >> 2014-03-18 10:38:44,558 [server.Accumulo] INFO : Connected to HDFS >> Starting master on hadoop-node-1 >> WARN : Max files open on hadoop-node-1 is 1024, recommend 65536 >> Starting garbage collector on hadoop-node-1 >> WARN : Max files open on hadoop-node-1 is 1024, recommend 65536 >> Starting tracer on hadoop-node-1 >> WARN : Max files open on hadoop-node-1 is 1024, recommend 65536 >> >> starting shell as root... >> >> [root@hadoop-node-1 zookeeper]# bash -x >> /usr/local/accumulo/bin/accumulo shell -u root >> Java HotSpot(TM) 64-Bit Server VM warning: You have loaded >> library /usr/local/hadoop/lib/native/libhadoop.so.1.0.0 which >> might have disabled stack guard. The VM will try to fix the >> stack guard no >> w. >> It's highly recommended that you fix the library with 'execstack >> -c <libfile>', or link it with '-z noexecstack'. >> 2014-03-18 10:38:56,002 [util.NativeCodeLoader] WARN : Unable to >> load native-hadoop library for your platform... using >> builtin-java classes where applicable >> Password: **** >> 2014-03-18 10:38:58,762 [impl.ServerClient] WARN : There are no >> tablet servers: check that zookeeper and accumulo are running. >> >> ... this is the point where it sits and acts like it doesn't do >> anything >> >> -- LOGS -- (most of this looks to be that I cannot connect to >> anything) >> >> here is the tail -f >> $ACCUMULO_HOME/logs/monitor_hadoop-node-1.local.debug.log >> >> 2014-03-18 10:42:54,617 [impl.ThriftScanner] DEBUG: Failed to >> locate tablet for table : !0 row : ~err_ >> 2014-03-18 10:42:57,625 [monitor.Monitor] INFO : Failed to >> obtain problem reports >> java.lang.RuntimeException: >> org.apache.accumulo.core.client.impl.ThriftScanner$ >> ScanTimedOutException >> at >> org.apache.accumulo.core.client.impl.ScannerIterator. >> hasNext(ScannerIterator.java:174) >> at >> org.apache.accumulo.server.problems.ProblemReports$3. >> hasNext(ProblemReports.java:241) >> at >> org.apache.accumulo.server.problems.ProblemReports. >> summarize(ProblemReports.java:299) >> at >> org.apache.accumulo.server.monitor.Monitor.fetchData( >> Monitor.java:399) >> at >> org.apache.accumulo.server.monitor.Monitor$1.run(Monitor. >> java:530) >> at >> org.apache.accumulo.core.util.LoggingRunnable.run( >> LoggingRunnable.java:34) >> at java.lang.Thread.run(Thread.java:744) >> Caused by: >> org.apache.accumulo.core.client.impl.ThriftScanner$ >> ScanTimedOutException >> at >> org.apache.accumulo.core.client.impl.ThriftScanner. >> scan(ThriftScanner.java:212) >> at >> org.apache.accumulo.core.client.impl.ScannerIterator$ >> Reader.run(ScannerIterator.java:82) >> at >> org.apache.accumulo.core.client.impl.ScannerIterator. >> hasNext(ScannerIterator.java:164) >> ... 6 more >> >> here is the tail -f >> $ACCUMULO+HOME/logs/tracer_hadoop-node-1.local.debug.log >> >> 2014-03-18 10:47:44,759 [impl.ServerClient] DEBUG: ClientService >> request failed null, retrying ... >> org.apache.thrift.transport.TTransportException: Failed to >> connect to a server >> at >> org.apache.accumulo.core.client.impl.ThriftTransportPool. >> getAnyTransport(ThriftTransportPool.java:455) >> at >> org.apache.accumulo.core.client.impl.ServerClient. >> getConnection(ServerClient.java:154) >> at >> org.apache.accumulo.core.client.impl.ServerClient. >> getConnection(ServerClient.java:128) >> at >> org.apache.accumulo.core.client.impl.ServerClient. >> getConnection(ServerClient.java:123) >> at >> org.apache.accumulo.core.client.impl.ServerClient. >> executeRaw(ServerClient.java:105) >> at >> org.apache.accumulo.core.client.impl.ServerClient. >> execute(ServerClient.java:71) >> at >> org.apache.accumulo.core.client.impl.ConnectorImpl.< >> init>(ConnectorImpl.java:64) >> at >> org.apache.accumulo.server.client.HdfsZooInstance. >> getConnector(HdfsZooInstance.java:154) >> at >> org.apache.accumulo.server.client.HdfsZooInstance. >> getConnector(HdfsZooInstance.java:149) >> at >> org.apache.accumulo.server.trace.TraceServer.<init>( >> TraceServer.java:200) >> at >> org.apache.accumulo.server.trace.TraceServer.main( >> TraceServer.java:295) >> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native >> Method) >> at >> sun.reflect.NativeMethodAccessorImpl.invoke( >> NativeMethodAccessorImpl.java:57) >> at >> sun.reflect.DelegatingMethodAccessorImpl.invoke( >> DelegatingMethodAccessorImpl.java:43) >> at java.lang.reflect.Method.invoke(Method.java:606) >> at org.apache.accumulo.start.Main$1.run(Main.java:103) >> at java.lang.Thread.run(Thread.java:744) >> >> >> On Tue, Mar 18, 2014 at 9:37 AM, Eric Newton >> <[email protected] <mailto:[email protected]>> wrote: >> >> Can you post the exact error message you are seeing? >> >> Verify that your HADOOP_PREFIX and HADOOP_CONF_DIR are being >> set properly in accumulo-site.xml. >> >> The output of: >> >> bash -x $ACCUMULO_HOME/bin/accumulo shell -u root >> >> >> would also help. >> >> It's going to be something simple. >> >> >> On Tue, Mar 18, 2014 at 9:14 AM, Benjamin Parrish >> <[email protected] >> <mailto:[email protected]>> wrote: >> >> Looking to see if there was an answer to this issue or >> if you could point me in a direction or example that >> could lead to a solution. >> >> >> On Sun, Mar 16, 2014 at 9:52 PM, Benjamin Parrish >> <[email protected] >> <mailto:[email protected]>> wrote: >> >> I am running Accumulo 1.5.1 >> >> <?xml version="1.0" encoding="UTF-8"?> >> <!-- >> Licensed to the Apache Software Foundation (ASF) >> under one or more >> contributor license agreements. See the NOTICE >> file distributed with >> this work for additional information regarding >> copyright ownership. >> The ASF licenses this file to You under the >> Apache License, Version 2.0 >> (the "License"); you may not use this file except >> in compliance with >> the License. You may obtain a copy of the License >> at >> >> http://www.apache.org/licenses/LICENSE-2.0 >> >> Unless required by applicable law or agreed to in >> writing, software >> distributed under the License is distributed on >> an "AS IS" BASIS, >> WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, >> either express or implied. >> See the License for the specific language >> governing permissions and >> limitations under the License. >> --> >> <?xml-stylesheet type="text/xsl" >> href="configuration.xsl"?> >> >> <configuration> >> <!-- Put your site-specific accumulo >> configurations here. The available configuration >> values along with their defaults are documented in >> docs/config.html Unless >> you are simply testing at your workstation, you >> will most definitely need to change the three >> entries below. --> >> >> <property> >> <name>instance.zookeeper.host</name> >> >> <value>hadoop-node-1:2181,hadoop-node-2:2181,hadoop- >> node-3:2181,hadoop-node-4:2181,hadoop-node-5:2181</value> >> <description>comma separated list of zookeeper >> servers</description> >> </property> >> >> <property> >> <name>logger.dir.walog</name> >> <value>walogs</value> >> <description>The property only needs to be set >> if upgrading from 1.4 which used to store >> write-ahead logs on the local >> filesystem. In 1.5 write-ahead logs are >> stored in DFS. When 1.5 is started for the first >> time it will copy any 1.4 >> write ahead logs into DFS. It is possible to >> specify a comma-separated list of directories. >> </description> >> </property> >> >> <property> >> <name>instance.secret</name> >> <value></value> >> <description>A secret unique to a given >> instance that all servers must know in order to >> communicate with one another. >> Change it before initialization. To >> change it later use ./bin/accumulo >> org.apache.accumulo.server.util.ChangeSecret --old >> [oldpasswd] --new [newpasswd], >> and then update this file. >> </description> >> </property> >> >> <property> >> <name>tserver.memory.maps.max</name> >> <value>1G</value> >> </property> >> >> <property> >> <name>tserver.cache.data.size</name> >> <value>128M</value> >> </property> >> >> <property> >> <name>tserver.cache.index.size</name> >> <value>128M</value> >> </property> >> >> <property> >> <name>trace.token.property.password</name> >> <!-- change this to the root user's password, >> and/or change the user below --> >> <value></value> >> </property> >> >> <property> >> <name>trace.user</name> >> <value>root</value> >> </property> >> >> <property> >> <name>general.classpaths</name> >> <value> >> $HADOOP_PREFIX/share/hadoop/common/.*.jar, >> $HADOOP_PREFIX/share/hadoop/common/lib/.*.jar, >> $HADOOP_PREFIX/share/hadoop/hdfs/.*.jar, >> $HADOOP_PREFIX/share/hadoop/mapreduce/.*.jar, >> $HADOOP_PREFIX/share/hadoop/yarn/.*.jar, >> /usr/lib/hadoop/.*.jar, >> /usr/lib/hadoop/lib/.*.jar, >> /usr/lib/hadoop-hdfs/.*.jar, >> /usr/lib/hadoop-mapreduce/.*.jar, >> /usr/lib/hadoop-yarn/.*.jar, >> $ACCUMULO_HOME/server/target/classes/, >> $ACCUMULO_HOME/lib/accumulo-server.jar, >> $ACCUMULO_HOME/core/target/classes/, >> $ACCUMULO_HOME/lib/accumulo-core.jar, >> $ACCUMULO_HOME/start/target/classes/, >> $ACCUMULO_HOME/lib/accumulo-start.jar, >> $ACCUMULO_HOME/fate/target/classes/, >> $ACCUMULO_HOME/lib/accumulo-fate.jar, >> $ACCUMULO_HOME/proxy/target/classes/, >> $ACCUMULO_HOME/lib/accumulo-proxy.jar, >> $ACCUMULO_HOME/lib/[^.].*.jar, >> $ZOOKEEPER_HOME/zookeeper[^.].*.jar, >> $HADOOP_CONF_DIR, >> $HADOOP_PREFIX/[^.].*.jar, >> $HADOOP_PREFIX/lib/[^.].*.jar, >> </value> >> <description>Classpaths that accumulo checks >> for updates and class files. >> When using the Security Manager, please >> remove the ".../target/classes/" values. >> </description> >> </property> >> </configuration> >> >> >> On Sun, Mar 16, 2014 at 9:06 PM, Josh Elser >> <[email protected] <mailto:[email protected]>> >> >> wrote: >> >> Posting your accumulo-site.xml (filtering out >> instance.secret and trace.password before you >> post) would also help us figure out what exactly >> is going on. >> >> >> On 3/16/14, 8:41 PM, Mike Drob wrote: >> >> Which version of Accumulo are you using? >> >> You might be missing the hadoop libraries >> from your classpath. For this, >> you would check your accumulo-site.xml and >> find the comment about Hadoop >> 2 in the file. >> >> >> On Sun, Mar 16, 2014 at 8:28 PM, Benjamin >> Parrish >> <[email protected] >> <mailto:[email protected]> >> <mailto:benjamin.d.parrish@__gmail.com >> >> <mailto:[email protected]>>> >> wrote: >> >> I have a couple of issues when trying >> to use Accumulo on Hadoop 2.2.0 >> >> 1) I start with accumulo init and >> everything runs through just fine, >> but I can find '/accumulo' using >> 'hadoop fs -ls /' >> >> 2) I try to run 'accumulo shell -u >> root' and it says that that >> Hadoop and ZooKeeper are not started, >> but if I run 'jps' on the each >> cluster node it shows all the necessary >> processes for both in the >> JVM. Is there something I am missing? >> >> -- >> Benjamin D. Parrish >> H: 540-597-7860 <tel:540-597-7860> >> <tel:540-597-7860 <tel:540-597-7860>> >> >> >> >> >> >> >> -- >> Benjamin D. Parrish >> H: 540-597-7860 <tel:540-597-7860> >> >> >> >> >> -- >> Benjamin D. Parrish >> H: 540-597-7860 <tel:540-597-7860> >> >> >> >> >> >> -- >> Benjamin D. Parrish >> H: 540-597-7860 <tel:540-597-7860> >> >> >> >> >> >> -- >> Benjamin D. Parrish >> H: 540-597-7860 >> > -- Benjamin D. Parrish H: 540-597-7860
