> > Java HotSpot(TM) 64-Bit Server VM warning: You have loaded library > /usr/local/hadoop/lib/native/libhadoop.so.1.0.0 which might have disabled > stack guard. The VM will try to fix the stack guard no > w. > It's highly recommended that you fix the library with 'execstack -c > <libfile>', or link it with '-z noexecstack'.
This is not causing your problem, but the binaries are 32-bit and need to be recompiled for a 64-bit environment. This message is just bogus. -Eric On Tue, Mar 18, 2014 at 12:07 PM, Benjamin Parrish < [email protected]> wrote: > on all 5 nodes there is a conf/accumulo-site.xml with the same values for > all 5. > > someone is going to kill me when they find out that i left off a bracket > or semi-colon somewhere... > > > On Tue, Mar 18, 2014 at 12:02 PM, Josh Elser <[email protected]> wrote: > >> No, running `accumulo init` on a single host is sufficient. >> >> Is accumulo-site.xml consistent across all machines? >> >> >> On 3/18/14, 11:57 AM, Benjamin Parrish wrote: >> >>> So here is the error from the tablet server... >>> >>> 2014-03-18 10:38:43,456 [client.ZooKeeperInstance] ERROR: unable obtain >>> instance id at /accumulo/instance_id >>> 2014-03-18 10:38:43,456 [tabletserver.TabletServer] ERROR: Uncaught >>> exception in TabletServer.main, exiting >>> java.lang.RuntimeException: Accumulo not initialized, there is no >>> instance id at /accumulo/instance_id >>> at >>> org.apache.accumulo.core.client.ZooKeeperInstance.getInstanceIDFromHdfs( >>> ZooKeeperInstance.java:295) >>> at >>> org.apache.accumulo.server.client.HdfsZooInstance._ >>> getInstanceID(HdfsZooInstance.java:126) >>> at >>> org.apache.accumulo.server.client.HdfsZooInstance. >>> getInstanceID(HdfsZooInstance.java:119) >>> at >>> org.apache.accumulo.server.conf.ZooConfiguration. >>> getInstance(ZooConfiguration.java:55) >>> at >>> org.apache.accumulo.server.conf.ServerConfiguration.getZooConfiguration( >>> ServerConfiguration.java:50) >>> at >>> org.apache.accumulo.server.conf.ServerConfiguration.getConfiguration( >>> ServerConfiguration.java:104) >>> at org.apache.accumulo.server.Accumulo.init(Accumulo.java:98) >>> at >>> org.apache.accumulo.server.tabletserver.TabletServer. >>> main(TabletServer.java:3249) >>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >>> at >>> sun.reflect.NativeMethodAccessorImpl.invoke( >>> NativeMethodAccessorImpl.java:57) >>> at >>> sun.reflect.DelegatingMethodAccessorImpl.invoke( >>> DelegatingMethodAccessorImpl.java:43) >>> at java.lang.reflect.Method.invoke(Method.java:606) >>> at org.apache.accumulo.start.Main$1.run(Main.java:103) >>> at java.lang.Thread.run(Thread.java:744) >>> >>> Do I need to run bin/accumulo init on every box in the cluster? >>> >>> >>> On Tue, Mar 18, 2014 at 11:19 AM, Eric Newton <[email protected] >>> <mailto:[email protected]>> wrote: >>> >>> Port numbers (for 1.5+) >>> >>> 4560 Accumulo monitor (for centralized log display) >>> 9997 Tablet Server >>> 9999 Master Server >>> 12234 Accumulo Tracer >>> 50091 Accumulo GC >>> 50095 Accumulo HTTP monitor >>> >>> >>> On Tue, Mar 18, 2014 at 11:04 AM, Benjamin Parrish >>> <[email protected] <mailto:[email protected]>> >>> >>> wrote: >>> >>> First off, are there specific ports that need to be opened up >>> for accumulo? I have hadoop operating without any issues as a 5 >>> node cluster. Zookeeper seems to be operating with 2181, 3888, >>> 2888 ports open. >>> >>> Here is some data from trying to get everything started and >>> getting into the shell. I discluded the bash portion as Eric >>> suggested because the mailing list rejected it for length and >>> thinking it was spam. >>> >>> bin/start-all.sh >>> >>> [root@hadoop-node-1 zookeeper]# bash -x >>> /usr/local/accumulo/bin/start-all.sh >>> Starting monitor on hadoop-node-1 >>> WARN : Max files open on hadoop-node-1 is 1024, recommend 65536 >>> Starting tablet servers ....... done >>> Starting tablet server on hadoop-node-3 >>> Starting tablet server on hadoop-node-5 >>> Starting tablet server on hadoop-node-2 >>> Starting tablet server on hadoop-node-4 >>> WARN : Max files open on hadoop-node-3 is 1024, recommend 65536 >>> WARN : Max files open on hadoop-node-2 is 1024, recommend 65536 >>> WARN : Max files open on hadoop-node-5 is 1024, recommend 65536 >>> WARN : Max files open on hadoop-node-4 is 1024, recommend 65536 >>> Java HotSpot(TM) 64-Bit Server VM warning: You have loaded >>> library /usr/local/hadoop/lib/native/libhadoop.so.1.0.0 which >>> might have disabled stack guard. The VM will try to fix the >>> stack guard no >>> w. >>> It's highly recommended that you fix the library with 'execstack >>> -c <libfile>', or link it with '-z noexecstack'. >>> 2014-03-18 10:38:43,143 [util.NativeCodeLoader] WARN : Unable to >>> load native-hadoop library for your platform... using >>> builtin-java classes where applicable >>> 2014-03-18 10:38:44,194 [server.Accumulo] INFO : Attempting to >>> talk to zookeeper >>> 2014-03-18 10:38:44,389 [server.Accumulo] INFO : Zookeeper >>> connected and initialized, attemping to talk to HDFS >>> 2014-03-18 10:38:44,558 [server.Accumulo] INFO : Connected to >>> HDFS >>> Starting master on hadoop-node-1 >>> WARN : Max files open on hadoop-node-1 is 1024, recommend 65536 >>> Starting garbage collector on hadoop-node-1 >>> WARN : Max files open on hadoop-node-1 is 1024, recommend 65536 >>> Starting tracer on hadoop-node-1 >>> WARN : Max files open on hadoop-node-1 is 1024, recommend 65536 >>> >>> starting shell as root... >>> >>> [root@hadoop-node-1 zookeeper]# bash -x >>> /usr/local/accumulo/bin/accumulo shell -u root >>> Java HotSpot(TM) 64-Bit Server VM warning: You have loaded >>> library /usr/local/hadoop/lib/native/libhadoop.so.1.0.0 which >>> might have disabled stack guard. The VM will try to fix the >>> stack guard no >>> w. >>> It's highly recommended that you fix the library with 'execstack >>> -c <libfile>', or link it with '-z noexecstack'. >>> 2014-03-18 10:38:56,002 [util.NativeCodeLoader] WARN : Unable to >>> load native-hadoop library for your platform... using >>> builtin-java classes where applicable >>> Password: **** >>> 2014-03-18 10:38:58,762 [impl.ServerClient] WARN : There are no >>> tablet servers: check that zookeeper and accumulo are running. >>> >>> ... this is the point where it sits and acts like it doesn't do >>> anything >>> >>> -- LOGS -- (most of this looks to be that I cannot connect to >>> anything) >>> >>> here is the tail -f >>> $ACCUMULO_HOME/logs/monitor_hadoop-node-1.local.debug.log >>> >>> 2014-03-18 10:42:54,617 [impl.ThriftScanner] DEBUG: Failed to >>> locate tablet for table : !0 row : ~err_ >>> 2014-03-18 10:42:57,625 [monitor.Monitor] INFO : Failed to >>> obtain problem reports >>> java.lang.RuntimeException: >>> org.apache.accumulo.core.client.impl.ThriftScanner$ >>> ScanTimedOutException >>> at >>> org.apache.accumulo.core.client.impl.ScannerIterator. >>> hasNext(ScannerIterator.java:174) >>> at >>> org.apache.accumulo.server.problems.ProblemReports$3. >>> hasNext(ProblemReports.java:241) >>> at >>> org.apache.accumulo.server.problems.ProblemReports. >>> summarize(ProblemReports.java:299) >>> at >>> org.apache.accumulo.server.monitor.Monitor.fetchData( >>> Monitor.java:399) >>> at >>> org.apache.accumulo.server.monitor.Monitor$1.run(Monitor. >>> java:530) >>> at >>> org.apache.accumulo.core.util.LoggingRunnable.run( >>> LoggingRunnable.java:34) >>> at java.lang.Thread.run(Thread.java:744) >>> Caused by: >>> org.apache.accumulo.core.client.impl.ThriftScanner$ >>> ScanTimedOutException >>> at >>> org.apache.accumulo.core.client.impl.ThriftScanner. >>> scan(ThriftScanner.java:212) >>> at >>> org.apache.accumulo.core.client.impl.ScannerIterator$ >>> Reader.run(ScannerIterator.java:82) >>> at >>> org.apache.accumulo.core.client.impl.ScannerIterator. >>> hasNext(ScannerIterator.java:164) >>> ... 6 more >>> >>> here is the tail -f >>> $ACCUMULO+HOME/logs/tracer_hadoop-node-1.local.debug.log >>> >>> 2014-03-18 10:47:44,759 [impl.ServerClient] DEBUG: ClientService >>> request failed null, retrying ... >>> org.apache.thrift.transport.TTransportException: Failed to >>> connect to a server >>> at >>> org.apache.accumulo.core.client.impl.ThriftTransportPool. >>> getAnyTransport(ThriftTransportPool.java:455) >>> at >>> org.apache.accumulo.core.client.impl.ServerClient. >>> getConnection(ServerClient.java:154) >>> at >>> org.apache.accumulo.core.client.impl.ServerClient. >>> getConnection(ServerClient.java:128) >>> at >>> org.apache.accumulo.core.client.impl.ServerClient. >>> getConnection(ServerClient.java:123) >>> at >>> org.apache.accumulo.core.client.impl.ServerClient. >>> executeRaw(ServerClient.java:105) >>> at >>> org.apache.accumulo.core.client.impl.ServerClient. >>> execute(ServerClient.java:71) >>> at >>> org.apache.accumulo.core.client.impl.ConnectorImpl.< >>> init>(ConnectorImpl.java:64) >>> at >>> org.apache.accumulo.server.client.HdfsZooInstance. >>> getConnector(HdfsZooInstance.java:154) >>> at >>> org.apache.accumulo.server.client.HdfsZooInstance. >>> getConnector(HdfsZooInstance.java:149) >>> at >>> org.apache.accumulo.server.trace.TraceServer.<init>( >>> TraceServer.java:200) >>> at >>> org.apache.accumulo.server.trace.TraceServer.main( >>> TraceServer.java:295) >>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native >>> Method) >>> at >>> sun.reflect.NativeMethodAccessorImpl.invoke( >>> NativeMethodAccessorImpl.java:57) >>> at >>> sun.reflect.DelegatingMethodAccessorImpl.invoke( >>> DelegatingMethodAccessorImpl.java:43) >>> at java.lang.reflect.Method.invoke(Method.java:606) >>> at org.apache.accumulo.start.Main$1.run(Main.java:103) >>> at java.lang.Thread.run(Thread.java:744) >>> >>> >>> On Tue, Mar 18, 2014 at 9:37 AM, Eric Newton >>> <[email protected] <mailto:[email protected]>> wrote: >>> >>> Can you post the exact error message you are seeing? >>> >>> Verify that your HADOOP_PREFIX and HADOOP_CONF_DIR are being >>> set properly in accumulo-site.xml. >>> >>> The output of: >>> >>> bash -x $ACCUMULO_HOME/bin/accumulo shell -u root >>> >>> >>> would also help. >>> >>> It's going to be something simple. >>> >>> >>> On Tue, Mar 18, 2014 at 9:14 AM, Benjamin Parrish >>> <[email protected] >>> <mailto:[email protected]>> wrote: >>> >>> Looking to see if there was an answer to this issue or >>> if you could point me in a direction or example that >>> could lead to a solution. >>> >>> >>> On Sun, Mar 16, 2014 at 9:52 PM, Benjamin Parrish >>> <[email protected] >>> <mailto:[email protected]>> wrote: >>> >>> I am running Accumulo 1.5.1 >>> >>> <?xml version="1.0" encoding="UTF-8"?> >>> <!-- >>> Licensed to the Apache Software Foundation (ASF) >>> under one or more >>> contributor license agreements. See the NOTICE >>> file distributed with >>> this work for additional information regarding >>> copyright ownership. >>> The ASF licenses this file to You under the >>> Apache License, Version 2.0 >>> (the "License"); you may not use this file except >>> in compliance with >>> the License. You may obtain a copy of the >>> License at >>> >>> http://www.apache.org/licenses/LICENSE-2.0 >>> >>> Unless required by applicable law or agreed to in >>> writing, software >>> distributed under the License is distributed on >>> an "AS IS" BASIS, >>> WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, >>> either express or implied. >>> See the License for the specific language >>> governing permissions and >>> limitations under the License. >>> --> >>> <?xml-stylesheet type="text/xsl" >>> href="configuration.xsl"?> >>> >>> <configuration> >>> <!-- Put your site-specific accumulo >>> configurations here. The available configuration >>> values along with their defaults are documented in >>> docs/config.html Unless >>> you are simply testing at your workstation, you >>> will most definitely need to change the three >>> entries below. --> >>> >>> <property> >>> <name>instance.zookeeper.host</name> >>> >>> <value>hadoop-node-1:2181,hadoop-node-2:2181,hadoop- >>> node-3:2181,hadoop-node-4:2181,hadoop-node-5:2181</value> >>> <description>comma separated list of zookeeper >>> servers</description> >>> </property> >>> >>> <property> >>> <name>logger.dir.walog</name> >>> <value>walogs</value> >>> <description>The property only needs to be set >>> if upgrading from 1.4 which used to store >>> write-ahead logs on the local >>> filesystem. In 1.5 write-ahead logs are >>> stored in DFS. When 1.5 is started for the first >>> time it will copy any 1.4 >>> write ahead logs into DFS. It is possible to >>> specify a comma-separated list of directories. >>> </description> >>> </property> >>> >>> <property> >>> <name>instance.secret</name> >>> <value></value> >>> <description>A secret unique to a given >>> instance that all servers must know in order to >>> communicate with one another. >>> Change it before initialization. To >>> change it later use ./bin/accumulo >>> org.apache.accumulo.server.util.ChangeSecret --old >>> [oldpasswd] --new [newpasswd], >>> and then update this file. >>> </description> >>> </property> >>> >>> <property> >>> <name>tserver.memory.maps.max</name> >>> <value>1G</value> >>> </property> >>> >>> <property> >>> <name>tserver.cache.data.size</name> >>> <value>128M</value> >>> </property> >>> >>> <property> >>> <name>tserver.cache.index.size</name> >>> <value>128M</value> >>> </property> >>> >>> <property> >>> <name>trace.token.property.password</name> >>> <!-- change this to the root user's password, >>> and/or change the user below --> >>> <value></value> >>> </property> >>> >>> <property> >>> <name>trace.user</name> >>> <value>root</value> >>> </property> >>> >>> <property> >>> <name>general.classpaths</name> >>> <value> >>> $HADOOP_PREFIX/share/hadoop/common/.*.jar, >>> $HADOOP_PREFIX/share/hadoop/ >>> common/lib/.*.jar, >>> $HADOOP_PREFIX/share/hadoop/hdfs/.*.jar, >>> $HADOOP_PREFIX/share/hadoop/mapreduce/.*.jar, >>> $HADOOP_PREFIX/share/hadoop/yarn/.*.jar, >>> /usr/lib/hadoop/.*.jar, >>> /usr/lib/hadoop/lib/.*.jar, >>> /usr/lib/hadoop-hdfs/.*.jar, >>> /usr/lib/hadoop-mapreduce/.*.jar, >>> /usr/lib/hadoop-yarn/.*.jar, >>> $ACCUMULO_HOME/server/target/classes/, >>> $ACCUMULO_HOME/lib/accumulo-server.jar, >>> $ACCUMULO_HOME/core/target/classes/, >>> $ACCUMULO_HOME/lib/accumulo-core.jar, >>> $ACCUMULO_HOME/start/target/classes/, >>> $ACCUMULO_HOME/lib/accumulo-start.jar, >>> $ACCUMULO_HOME/fate/target/classes/, >>> $ACCUMULO_HOME/lib/accumulo-fate.jar, >>> $ACCUMULO_HOME/proxy/target/classes/, >>> $ACCUMULO_HOME/lib/accumulo-proxy.jar, >>> $ACCUMULO_HOME/lib/[^.].*.jar, >>> $ZOOKEEPER_HOME/zookeeper[^.].*.jar, >>> $HADOOP_CONF_DIR, >>> $HADOOP_PREFIX/[^.].*.jar, >>> $HADOOP_PREFIX/lib/[^.].*.jar, >>> </value> >>> <description>Classpaths that accumulo checks >>> for updates and class files. >>> When using the Security Manager, please >>> remove the ".../target/classes/" values. >>> </description> >>> </property> >>> </configuration> >>> >>> >>> On Sun, Mar 16, 2014 at 9:06 PM, Josh Elser >>> <[email protected] <mailto:[email protected]>> >>> >>> wrote: >>> >>> Posting your accumulo-site.xml (filtering out >>> instance.secret and trace.password before you >>> post) would also help us figure out what exactly >>> is going on. >>> >>> >>> On 3/16/14, 8:41 PM, Mike Drob wrote: >>> >>> Which version of Accumulo are you using? >>> >>> You might be missing the hadoop libraries >>> from your classpath. For this, >>> you would check your accumulo-site.xml and >>> find the comment about Hadoop >>> 2 in the file. >>> >>> >>> On Sun, Mar 16, 2014 at 8:28 PM, Benjamin >>> Parrish >>> <[email protected] >>> <mailto:[email protected]> >>> <mailto:benjamin.d.parrish@__gmail.com >>> >>> <mailto:[email protected]>>> >>> wrote: >>> >>> I have a couple of issues when trying >>> to use Accumulo on Hadoop 2.2.0 >>> >>> 1) I start with accumulo init and >>> everything runs through just fine, >>> but I can find '/accumulo' using >>> 'hadoop fs -ls /' >>> >>> 2) I try to run 'accumulo shell -u >>> root' and it says that that >>> Hadoop and ZooKeeper are not started, >>> but if I run 'jps' on the each >>> cluster node it shows all the necessary >>> processes for both in the >>> JVM. Is there something I am missing? >>> >>> -- >>> Benjamin D. Parrish >>> H: 540-597-7860 <tel:540-597-7860> >>> <tel:540-597-7860 <tel:540-597-7860>> >>> >>> >>> >>> >>> >>> >>> -- >>> Benjamin D. Parrish >>> H: 540-597-7860 <tel:540-597-7860> >>> >>> >>> >>> >>> -- >>> Benjamin D. Parrish >>> H: 540-597-7860 <tel:540-597-7860> >>> >>> >>> >>> >>> >>> -- >>> Benjamin D. Parrish >>> H: 540-597-7860 <tel:540-597-7860> >>> >>> >>> >>> >>> >>> -- >>> Benjamin D. Parrish >>> H: 540-597-7860 >>> >> > > > -- > Benjamin D. Parrish > H: 540-597-7860 >
