I have 3 PC cluster.(pc1 , pc2 , pc3)
Hadoop master (pc1), 2 slaves (pc2,pc3)

HBase and ZK running on pc1, two region servers (pc2,pc3)

pc1 : Intel core2 , 2.4GHz , RAM 1G

pc2 : Intel core2 , 2.4GHz , RAM 1G

pc3 : Intel core2 , 1.86GHZ, RAM 2G

-----------------------------------------------------------

hbase-env.sh
 export HBASE_MANAGES_ZK=true

-----------------------------------------------------------
<configuration>

 <property>
    <name>hbase.cluster.distributed</name>
    <value>true</value>
    <description>true:fully-distributed with unmanaged Zookeeper Quorum
    </description>
  </property>

  <property>
    <name>hbase.rootdir</name>
    <value>hdfs://convera:9000/hbase</value>
    <description>The directory shared by region servers.
    Should be fully-qualified to include the filesystem to use.
    E.g: hdfs://NAMENODE_SERVER:PORT/HBASE_ROOTDIR
    </description>
  </property>

  <property>
    <name>hbase.master</name>
    <value>10.42.253.182:60000</value>
    <description>The host and port that the HBase master runs at.
    A value of 'local' runs the master and a regionserver in
    a single process.
    </description>
  </property>
 <property>
    <name>hbase.zookeeper.quorum</name>
    <value>convera</value>
    <description>Comma separated list of servers in the ZooKeeper Quorum.
    For example,
"host1.mydomain.com,host2.mydomain.com,host3.mydomain.com".
    By default this is set to localhost for local and pseudo-distributed
modes
    of operation. For a fully-distributed setup, this should be set to a
full
    list of ZooKeeper quorum servers. If HBASE_MANAGES_ZK is set in
hbase-env.sh
    this is the list of servers which we will start/stop ZooKeeper on.
    </description>
  </property>

<property>
    <name>hbase.zookeeper.property.maxClientCnxns</name>
    <value>30</value>
    <description>Property from ZooKeeper's config zoo.cfg.
    Limit on number of concurrent connections (at the socket level) that a
    single client, identified by IP address, may make to a single member of
    the ZooKeeper ensemble. Set high to avoid zk connection issues running
    standalone and pseudo-distributed.
    </description>
  </property>

</configuration>






                                                                                
                                                                      
                      Amandeep Khurana                                          
                                                                      
                      <[email protected]        To:      
[email protected]                                                    
              
                      >                        cc:      (bcc: Y_823910/TSMC)    
                                                                      
                                               Subject: Re: HBase-0.20.0 multi 
read                                                                   
                      2009/08/21 11:54                                          
                                                                      
                      AM                                                        
                                                                      
                      Please respond to                                         
                                                                      
                      hbase-user                                                
                                                                      
                                                                                
                                                                      
                                                                                
                                                                      




You ideally want to have 3-5 servers outside the hbase servers... 1
server is not enough. That could to be causing you the trouble.

Post logs from the master and the region server where the read failed.

Also, what's your configuration? How many nodes, ram, cpus etc?

On 8/20/09, [email protected] <[email protected]> wrote:
>
> Hi there,
>
> It worked well while I fired 5 threads to fetch data from HBASE,but
> it failed after I incresed to 6 threads.
> Although it showed some WARN, the thread job can't be done!
> My hbase is the latest version hbase0.20.
> I want to test HBase multi read performance.
> Any suggestion?
> Thank you
>
> Fleming
>
>
> hbase-env.sh
>    export HBASE_MANAGES_ZK=true
>
> 09/08/21 09:54:07 WARN zookeeper.ZooKeeperWrapper: Failed to create
/hbase
> -- check quorum servers, currently=10.42.253.182:2181
> org.apache.zookeeper.KeeperException$ConnectionLossException:
> KeeperErrorCode = ConnectionLoss for /hbase
>       at
> org.apache.zookeeper.KeeperException.create(KeeperException.java:90)
>       at
> org.apache.zookeeper.KeeperException.create(KeeperException.java:42)
>       at org.apache.zookeeper.ZooKeeper.create(ZooKeeper.java:522)
>       at
>
org.apache.hadoop.hbase.zookeeper.ZooKeeperWrapper.ensureExists(ZooKeeperWrapper.java:342)

>       at
>
org.apache.hadoop.hbase.zookeeper.ZooKeeperWrapper.ensureParentExists(ZooKeeperWrapper.java:365)

>       at
>
org.apache.hadoop.hbase.zookeeper.ZooKeeperWrapper.checkOutOfSafeMode(ZooKeeperWrapper.java:478)

>       at
>
org.apache.hadoop.hbase.client.HConnectionManager$TableServers.locateRootRegion(HConnectionManager.java:846)

>       at
>
org.apache.hadoop.hbase.client.HConnectionManager$TableServers.locateRegion(HConnectionManager.java:515)

>       at
>
org.apache.hadoop.hbase.client.HConnectionManager$TableServers.locateRegion(HConnectionManager.java:491)

>       at
>
org.apache.hadoop.hbase.client.HConnectionManager$TableServers.locateRegionInMeta(HConnectionManager.java:565)

>       at
>
org.apache.hadoop.hbase.client.HConnectionManager$TableServers.locateRegion(HConnectionManager.java:524)

>       at
>
org.apache.hadoop.hbase.client.HConnectionManager$TableServers.locateRegion(HConnectionManager.java:491)

>       at
>
org.apache.hadoop.hbase.client.HConnectionManager$TableServers.locateRegionInMeta(HConnectionManager.java:565)

>       at
>
org.apache.hadoop.hbase.client.HConnectionManager$TableServers.locateRegion(HConnectionManager.java:528)

>       at
>
org.apache.hadoop.hbase.client.HConnectionManager$TableServers.locateRegion(HConnectionManager.java:491)

>       at org.apache.hadoop.hbase.client.HTable.<init>(HTable.java:123)
>       at org.apache.hadoop.hbase.client.HTable.<init>(HTable.java:101)
>       at
> org.gridgain.examples.executor.FlowJob.getHBaseData(FlowJob.java:144)
>       at org.gridgain.examples.executor.FlowJob.call(FlowJob.java:78)
>       at org.gridgain.examples.executor.FlowJob.call(FlowJob.java:1)
>       at
>
org.gridgain.grid.kernal.executor.GridExecutorCallableTask$1.execute(GridExecutorCallableTask.java:57)

>       at
>
org.gridgain.grid.kernal.processors.job.GridJobWorker.body(GridJobWorker.java:406)

>       at
> org.gridgain.grid.util.runnable.GridRunnable$1.run(GridRunnable.java:142)
>       at
> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:441)
>       at
java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303)
>       at java.util.concurrent.FutureTask.run(FutureTask.java:138)
>       at
> org.gridgain.grid.util.runnable.GridRunnable.run(GridRunnable.java:194)
>       at
>
java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)

>       at
>
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)

>       at java.lang.Thread.run(Thread.java:619)
>
---------------------------------------------------------------------------
>                                                          TSMC PROPERTY
>  This email communication (and any attachments) is proprietary
information
>  for the sole use of its
>  intended recipient. Any unauthorized review, use or distribution by
anyone
>  other than the intended
>  recipient is strictly prohibited.  If you are not the intended
recipient,
>  please notify the sender by
>  replying to this email, and then delete this email and any copies of it
>  immediately. Thank you.
>
---------------------------------------------------------------------------
>
>
>
>


--


Amandeep Khurana
Computer Science Graduate Student
University of California, Santa Cruz




 --------------------------------------------------------------------------- 
                                                         TSMC PROPERTY       
 This email communication (and any attachments) is proprietary information   
 for the sole use of its                                                     
 intended recipient. Any unauthorized review, use or distribution by anyone  
 other than the intended                                                     
 recipient is strictly prohibited.  If you are not the intended recipient,   
 please notify the sender by                                                 
 replying to this email, and then delete this email and any copies of it     
 immediately. Thank you.                                                     
 --------------------------------------------------------------------------- 



Reply via email to