Try giving your hbase master more memory.  See item 3 in the FAQ:
http://wiki.apache.org/lucene-hadoop/Hbase/FAQ#3.
St.Ack

Thiago Jackiw wrote:
I've just checked out hadoop from svn trunk, 'ant' compiled it by
following the instructions on the wiki and I'm getting a strange
behavior when going to HBase's web interface. Note that I don't even
have any data stored on hadoop's dfs nor have a crazy-custom cluster
configuration.

Tailing hbase's master log file this is what happens, first:
====
2007-12-04 18:25:18,625 WARN org.apache.hadoop.ipc.Server: Out of
Memory in server select
java.lang.OutOfMemoryError: Java heap space
        at java.nio.HeapByteBuffer.<init>(HeapByteBuffer.java:39)
        at java.nio.ByteBuffer.allocate(ByteBuffer.java:312)
        at 
org.apache.hadoop.ipc.Server$Connection.readAndProcess(Server.java:503)
        at org.apache.hadoop.ipc.Server$Listener.doRead(Server.java:379)
        at org.apache.hadoop.ipc.Server$Listener.run(Server.java:296)


Then it turns into this:
====
2007-12-04 18:37:55,054 INFO org.apache.hadoop.ipc.Client: Retrying
connect to server: /192.168.9.8:60020. Already tried 1 time(s).
2007-12-04 18:37:56,055 INFO org.apache.hadoop.ipc.Client: Retrying
connect to server: /192.168.9.8:60020. Already tried 2 time(s).
2007-12-04 18:37:57,056 INFO org.apache.hadoop.ipc.Client: Retrying
connect to server: /192.168.9.8:60020. Already tried 3 time(s).
2007-12-04 18:37:58,056 INFO org.apache.hadoop.ipc.Client: Retrying
connect to server: /192.168.9.8:60020. Already tried 4 time(s).
2007-12-04 18:37:59,057 INFO org.apache.hadoop.ipc.Client: Retrying
connect to server: /192.168.9.8:60020. Already tried 5 time(s).
2007-12-04 18:38:00,058 INFO org.apache.hadoop.ipc.Client: Retrying
connect to server: /192.168.9.8:60020. Already tried 6 time(s).
2007-12-04 18:38:01,058 INFO org.apache.hadoop.ipc.Client: Retrying
connect to server: /192.168.9.8:60020. Already tried 7 time(s).
2007-12-04 18:38:02,059 INFO org.apache.hadoop.ipc.Client: Retrying
connect to server: /192.168.9.8:60020. Already tried 8 time(s).
2007-12-04 18:38:03,060 INFO org.apache.hadoop.ipc.Client: Retrying
connect to server: /192.168.9.8:60020. Already tried 9 time(s).
2007-12-04 18:38:04,060 INFO org.apache.hadoop.ipc.Client: Retrying
connect to server: /192.168.9.8:60020. Already tried 10 time(s)
java.net.ConnectException: Connection refused
        at java.net.PlainSocketImpl.socketConnect(Native Method)


I've setup my cluster as follows:

hadoop-site.xml
====
<configuration>
<property>
    <name>fs.default.name</name>
    <value>localhost:9000</value>
  </property>
  <property>
    <name>mapred.job.tracker</name>
    <value>localhost:9001</value>
  </property>
</configuration>


hadoop-env.sh:
====
export JAVA_HOME=/usr/lib/java


hbase-site.xml
====
<configuration>
<property>
     <name>hbase.master</name>
     <value>localhost:60000</value>
     <description>The host and port that the HBase master runs at</description>
   </property>
   <property>
     <name>hbase.rootdir</name>
     <value>/hbase</value>
     <description>location of HBase instance in dfs</description>
   </property>
</configuration>


/etc/hosts
====
127.0.0.1     localhost
192.168.9.8 localhost

====

Any ideas?

Also, 'bin/stop-hbase.sh' isn't actually stopping the server when the
problem "org.apache.hadoop.ipc.Client: Retrying connect to server:
/192.168.9.8:60020. Already tried 8 time(s)" exists.

Thanks


Reply via email to