Benjamin,

It may be better to step back for a second and make sure you have the Hadoop 
environment set up correctly.  You are very close but it seems like there is 
just an issue with the Accumulo classpath or your environment variables.

In regard to ensuring zookeeper is working, you can use the Command line 
interface script "zkCli.sh" to connect.  However, you need to specify which 
server zookeeper is running on.

# zkCli.sh -server <yourZooServer>

My setup is a bit different, but If you connect you should be able to do ls / 
and see what znodes are created in zookeeper.  If the accumulo shell -init 
worked correctly, you should see something like this when you connect and run 
ls /accumulo:


[root@1620-Megatron bin]# ./zookeeper-client -server 1620-Megatron
Connecting to 1620-Megatron
2014-03-19 09:50:57,286 [myid:] - INFO  [main:Environment@100] - Client 
environment:zookeeper.version=3.4.5-cdh4.5.0--1, built on 11/20/2013 22:29 GMT
2014-03-19 09:50:57,289 [myid:] - INFO  [main:Environment@100] - Client 
environment:host.name=1620-Megatron
2014-03-19 09:50:57,290 [myid:] - INFO  [main:Environment@100] - Client 
environment:java.version=1.7.0_45
2014-03-19 09:50:57,290 [myid:] - INFO  [main:Environment@100] - Client 
environment:java.vendor=Oracle Corporation
2014-03-19 09:50:57,291 [myid:] - INFO  [main:Environment@100] - Client 
environment:java.home=/usr/java/jdk1.7.0_45/jre
2014-03-19 09:50:57,291 [myid:] - INFO  [main:Environment@100] - Client 
environment:java.class.path=/opt/cloudera/parcels/CDH-4.5.0-1.cdh4.5.0.p0.30/lib/zookeeper/bin/../build/classes:/opt/cloudera/parcels/CDH-4.5.0-1.cdh4.5.0.p0.30/lib/zookeeper/bin/../build/lib/*.jar:/opt/cloudera/parcels/CDH-4.5.0-1.cdh4.5.0.p0.30/lib/zookeeper/bin/../lib/slf4j-log4j12-1.6.1.jar:/opt/cloudera/parcels/CDH-4.5.0-1.cdh4.5.0.p0.30/lib/zookeeper/bin/../lib/slf4j-api-1.6.1.jar:/opt/cloudera/parcels/CDH-4.5.0-1.cdh4.5.0.p0.30/lib/zookeeper/bin/../lib/netty-3.2.2.Final.jar:/opt/cloudera/parcels/CDH-4.5.0-1.cdh4.5.0.p0.30/lib/zookeeper/bin/../lib/log4j-1.2.15.jar:/opt/cloudera/parcels/CDH-4.5.0-1.cdh4.5.0.p0.30/lib/zookeeper/bin/../lib/jline-0.9.94.jar:/opt/cloudera/parcels/CDH-4.5.0-1.cdh4.5.0.p0.30/lib/zookeeper/bin/../zookeeper-3.4.5-cdh4.5.0.jar:/opt/cloudera/parcels/CDH-4.5.0-1.cdh4.5.0.p0.30/lib/zookeeper/bin/../src/java/lib/*.jar:/etc/zookeeper/conf::/etc/zookeeper/conf:/opt/cloudera/parcels/CDH-4.5.0-1.cdh4.5.0.p0.30/bin/../lib/zookeeper/zookeeper-3.4.5-cdh4.5.0.jar:/opt/cloudera/parcels/CDH-4.5.0-1.cdh4.5.0.p0.30/bin/../lib/zookeeper/zookeeper.jar:/opt/cloudera/parcels/CDH-4.5.0-1.cdh4.5.0.p0.30/bin/../lib/zookeeper/lib/log4j-1.2.15.jar:/opt/cloudera/parcels/CDH-4.5.0-1.cdh4.5.0.p0.30/bin/../lib/zookeeper/lib/jline-0.9.94.jar:/opt/cloudera/parcels/CDH-4.5.0-1.cdh4.5.0.p0.30/bin/../lib/zookeeper/lib/slf4j-api-1.6.1.jar:/opt/cloudera/parcels/CDH-4.5.0-1.cdh4.5.0.p0.30/bin/../lib/zookeeper/lib/netty-3.2.2.Final.jar:/opt/cloudera/parcels/CDH-4.5.0-1.cdh4.5.0.p0.30/bin/../lib/zookeeper/lib/slf4j-log4j12-1.6.1.jar
2014-03-19 09:50:57,292 [myid:] - INFO  [main:Environment@100] - Client 
environment:java.library.path=/usr/java/packages/lib/amd64:/usr/lib64:/lib64:/lib:/usr/lib
2014-03-19 09:50:57,292 [myid:] - INFO  [main:Environment@100] - Client 
environment:java.io.tmpdir=/tmp
2014-03-19 09:50:57,293 [myid:] - INFO  [main:Environment@100] - Client 
environment:java.compiler=<NA>
2014-03-19 09:50:57,293 [myid:] - INFO  [main:Environment@100] - Client 
environment:os.name=Linux
2014-03-19 09:50:57,294 [myid:] - INFO  [main:Environment@100] - Client 
environment:os.arch=amd64
2014-03-19 09:50:57,294 [myid:] - INFO  [main:Environment@100] - Client 
environment:os.version=2.6.32-431.el6.x86_64
2014-03-19 09:50:57,295 [myid:] - INFO  [main:Environment@100] - Client 
environment:user.name=root
2014-03-19 09:50:57,295 [myid:] - INFO  [main:Environment@100] - Client 
environment:user.home=/root
2014-03-19 09:50:57,295 [myid:] - INFO  [main:Environment@100] - Client 
environment:user.dir=/opt/cloudera/parcels/CDH-4.5.0-1.cdh4.5.0.p0.30/bin
2014-03-19 09:50:57,297 [myid:] - INFO  [main:ZooKeeper@438] - Initiating 
client connection, connectString=1620-Megatron sessionTimeout=30000 
watcher=org.apache.zookeeper.ZooKeeperMain$MyWatcher@40e51e67
Welcome to ZooKeeper!
2014-03-19 09:50:57,319 [myid:] - INFO  
[main-SendThread(1620-Megatron:2181):ClientCnxn$SendThread@966] - Opening 
socket connection to server 1620-Megatron/10.35.56.87:2181. Will not attempt to 
authenticate using SASL (unknown error)
2014-03-19 09:50:57,324 [myid:] - INFO  
[main-SendThread(1620-Megatron:2181):ClientCnxn$SendThread@849] - Socket 
connection established to 1620-Megatron /10.35.56.87:2181, initiating session
JLine support is enabled
[zk: 1620-Megatron(CONNECTING) 0] 2014-03-19 09:50:57,360 [myid:] - INFO  
[main-SendThread(1620-Megatron:2181):ClientCnxn$SendThread@1207] - Session 
establishment complete on server 1620-Megatron /10.35.56.87:2181, sessionid = 
0x244d57a1c511db9, negotiated timeout = 30000

WATCHER::

WatchedEvent state:SyncConnected type:None path:null

[zk: 1620-Megatron(CONNECTED) 0] ls /
[accumulo, zookeeper]
[zk: 1620-Megatron(CONNECTED) 1] ls /accumulo
[a85286bf-031c-4e24-9b47-f6aca34401b8, a531e027-6154-47ec-8b0e-eebf20a9a902, 
instances, ec960eed-e05f-448c-b417-74620a242764, 
1b491f7e-d2a2-4353-a691-fdcab06592bd, 675bbabc-0e0d-4ae1-9de0-0dba3cd4d1f0]

If you are able to see that there is an /accumulo node, then you're accumulo 
init probably worked fine.  Then you need to make sure the accumulo-env.sh is 
defining the correct paths to the Hadoop settings.   Remember that the slaves 
running tablet servers must be setup the same, and ensure passwordless ssh is 
working for the account you are running start-all.sh.

Hope this helps.

From: [email protected] 
[mailto:[email protected]] On 
Behalf Of Benjamin Parrish
Sent: Wednesday, March 19, 2014 9:28 AM
To: [email protected]
Subject: Re: Installing with Hadoop 2.2.0

So, I am back to no clue now...

On Wed, Mar 19, 2014 at 9:13 AM, Josh Elser 
<[email protected]<mailto:[email protected]>> wrote:

I think by default zkCli.sh will just try to connect to localhost. You can 
change this by providing the quorum string to the script with the -server 
option.
On Mar 19, 2014 8:29 AM, "Benjamin Parrish" 
<[email protected]<mailto:[email protected]>> wrote:
I adjusted accumulo-env.sh to have hard coded values as seen below.

Are there any logs that could shed some light on this issue?

If it also helps I am using CentOS 6.5, Hadoop 2.2.0, ZooKeeper 3.4.6.

I also ran across this, that didn't look right...

Welcome to ZooKeeper!
2014-03-19 08:25:53,479 [myid:] - INFO  
[main-SendThread(localhost:2181):ClientCnxn$SendThread@975] - Opening socket 
connection to server localhost/127.0.0.1:2181<http://127.0.0.1:2181>. Will not 
attempt to authenticat
e using SASL (unknown error)
2014-03-19 08:25:53,483 [myid:] - INFO  
[main-SendThread(localhost:2181):ClientCnxn$SendThread@852] - Socket connection 
established to localhost/127.0.0.1:2181<http://127.0.0.1:2181>, initiating 
session
JLine support is enabled
[zk: localhost:2181(CONNECTING) 0] 2014-03-19 08:25:53,523 [myid:] - INFO  
[main-SendThread(localhost:2181):ClientCnxn$SendThread@1235] - Session 
establishment complete on server localhost/127.0.
0.1:2181, sessionid = 0x144da4e00d90000, negotiated timeout = 30000

should ZooKeeper try to hit localhost/127.0.0.1<http://127.0.0.1>?

my zoo.cfg looks like this....
tickTime=2000
initLimit=10
syncLimit=5
dataDir=/usr/local/zookeeper/data
clientPort=2181
server.1=hadoop-node-1:2888:3888
server.2=hadoop-node-2:2888:3888
server.3=hadoop-node-3:2888:3888
server.4=hadoop-node-4:2888:3888
server.5=hadoop-node-5:2888:3888

#! /usr/bin/env bash

# Licensed to the Apache Software Foundation (ASF) under one or more
# contributor license agreements.  See the NOTICE file distributed with
# this work for additional information regarding copyright ownership.
# The ASF licenses this file to You under the Apache License, Version 2.0
# (the "License"); you may not use this file except in compliance with
# the License.  You may obtain a copy of the License at
#
#     http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.

###
### Configure these environment variables to point to your local installations.
###
### The functional tests require conditional values, so keep this style:
###
### test -z "$JAVA_HOME" && export JAVA_HOME=/usr/local/lib/jdk-1.6.0
###
###
### Note that the -Xmx -Xms settings below require substantial free memory:
### you may want to use smaller values, especially when running everything
### on a single machine.
###
if [ -z "$HADOOP_HOME" ]
then
   test -z "$HADOOP_PREFIX"      && export HADOOP_PREFIX=/usr/local/hadoop
else
   HADOOP_PREFIX="$HADOOP_HOME"
   unset HADOOP_HOME
fi
# test -z "$HADOOP_CONF_DIR"       && export 
HADOOP_CONF_DIR="/usr/local/hadoop/conf"
# hadoop-2.0:
test -z "$HADOOP_CONF_DIR"     && export 
HADOOP_CONF_DIR="/usr/local/hadoop/etc/hadoop"

#! /usr/bin/env bash

# Licensed to the Apache Software Foundation (ASF) under one or more
# contributor license agreements.  See the NOTICE file distributed with
# this work for additional information regarding copyright ownership.
# The ASF licenses this file to You under the Apache License, Version 2.0
# (the "License"); you may not use this file except in compliance with
# the License.  You may obtain a copy of the License at
#
#     http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.

###
### Configure these environment variables to point to your local installations.
###
### The functional tests require conditional values, so keep this style:
###
### test -z "$JAVA_HOME" && export JAVA_HOME=/usr/local/lib/jdk-1.6.0
###
###
### Note that the -Xmx -Xms settings below require substantial free memory:
### you may want to use smaller values, especially when running everything
### on a single machine.
###
if [ -z "$HADOOP_HOME" ]
then
   test -z "$HADOOP_PREFIX"      && export HADOOP_PREFIX=/usr/local/hadoop
else
   HADOOP_PREFIX="$HADOOP_HOME"
   unset HADOOP_HOME
fi
# test -z "$HADOOP_CONF_DIR"       && export 
HADOOP_CONF_DIR="/usr/local/hadoop/conf"
# hadoop-2.0:
test -z "$HADOOP_CONF_DIR"     && export 
HADOOP_CONF_DIR="/usr/local/hadoop/etc/hadoop"

test -z "$JAVA_HOME"             && export JAVA_HOME=/usr/lib/jvm/jdk1.7.0
test -z "$ZOOKEEPER_HOME"        && export ZOOKEEPER_HOME=/usr/local/zookeeper
test -z "$ACCUMULO_LOG_DIR"      && export 
ACCUMULO_LOG_DIR=/usr/local/accumulo/logs
if [ -f /usr/local/accumulo/conf/accumulo.policy ]
then
   POLICY="-Djava.security.manager 
-Djava.security.policy=/usr/local/accumulo/conf/accumulo.policy"
fi
test -z "$ACCUMULO_TSERVER_OPTS" && export ACCUMULO_TSERVER_OPTS="${POLICY} 
-Xmx1g -Xms1g -XX:NewSize=500m -XX:MaxNewSize=500m "
test -z "$ACCUMULO_MASTER_OPTS"  && export ACCUMULO_MASTER_OPTS="${POLICY} 
-Xmx1g -Xms1g"
test -z "$ACCUMULO_MONITOR_OPTS" && export ACCUMULO_MONITOR_OPTS="${POLICY} 
-Xmx1g -Xms256m"
test -z "$ACCUMULO_GC_OPTS"      && export ACCUMULO_GC_OPTS="-Xmx256m -Xms256m"
test -z "$ACCUMULO_GENERAL_OPTS" && export 
ACCUMULO_GENERAL_OPTS="-XX:+UseConcMarkSweepGC 
-XX:CMSInitiatingOccupancyFraction=75 -Djava.net.preferIPv4Stack=true"
test -z "$ACCUMULO_OTHER_OPTS"   && export ACCUMULO_OTHER_OPTS="-Xmx1g -Xms256m"
# what do when the JVM runs out of heap memory
export ACCUMULO_KILL_CMD='kill -9 %p'

# Should the monitor bind to all network interfaces -- default: false
# export ACCUMULO_MONITOR_BIND_ALL="true"

On Tue, Mar 18, 2014 at 8:58 PM, Sean Busbey 
<[email protected]<mailto:[email protected]>> wrote:

On Mar 18, 2014 7:51 PM, "Benjamin Parrish" 
<[email protected]<mailto:[email protected]>> wrote:
>
> HADOOP_CONF_DIR=/usr/local/hadoop/etc/hadoop is set in all ~/.bash_profile 
> files as needed.
>
>

Can you add to the gist the output of running

$> find $HADOOP_CONF_DIR

As the user who runs the tablet server on the same host you ran the classpath 
command on?

-Sean



--
Benjamin D. Parrish
H: 540-597-7860<tel:540-597-7860>



--
Benjamin D. Parrish
H: 540-597-7860

Reply via email to