No there's something else here, your machines are weak indeed but it
should be alright.

So you say you are fetching data from HBase... how? How big is each
row? What kind of load do you see on your machines? Lots of IO wait?

The error you pasted in the first mail seems to imply that your client
wasn't able to connect to Zookeeper. Is there a connectivity issue? Do
you see load on the machine hosting the ZK server? Did you monitor
that? Did you try to put a ZK process on each server (just add the
other machines in hbase.zookeeper.quorum)?

Without that kind of info it is hard to tell what's exactly wrong.

J-D

On Fri, Aug 21, 2009 at 4:12 AM, <[email protected]> wrote:
> You mean my PCs are not good enough to run HBase well ?
> I've put 5 oracle tables to HBase successfully , the biggest table record
> count is only 50,000.
> Is there a client request limit for region server?
> Two region server just serve 5 clients, it's a little strange!
> Any suggestion hardware spec for HBase?
> For that spec, how many clients can fetch data from HBase  concurrently?
>
> Fleming
>
>
>
>
>
>                      Amandeep Khurana
>                      <[email protected]        To:      
> [email protected]
>                      >                        cc:      (bcc: Y_823910/TSMC)
>                                               Subject: Re: HBase-0.20.0 multi 
> read
>                      2009/08/21 03:49
>                      PM
>                      Please respond to
>                      hbase-user
>
>
>
>
>
>
> On Fri, Aug 21, 2009 at 12:45 AM, <[email protected]> wrote:
>
>>
>> I have 3 PC cluster.(pc1 , pc2 , pc3)
>> Hadoop master (pc1), 2 slaves (pc2,pc3)
>>
>> HBase and ZK running on pc1, two region servers (pc2,pc3)
>>
>> pc1 : Intel core2 , 2.4GHz , RAM 1G
>>
>> pc2 : Intel core2 , 2.4GHz , RAM 1G
>>
>> pc3 : Intel core2 , 1.86GHZ, RAM 2G
>>
>
> This is a very low config for HBase. I doubt if you'll be able to get a
> remotely stable hbase instance going in this. More so, if you are trying to
> test how much load it can take...
>
>
>>
>> -----------------------------------------------------------
>>
>> hbase-env.sh
>>  export HBASE_MANAGES_ZK=true
>>
>> -----------------------------------------------------------
>> <configuration>
>>
>>  <property>
>>    <name>hbase.cluster.distributed</name>
>>    <value>true</value>
>>    <description>true:fully-distributed with unmanaged Zookeeper Quorum
>>    </description>
>>  </property>
>>
>>   <property>
>>    <name>hbase.rootdir</name>
>>    <value>hdfs://convera:9000/hbase</value>
>>    <description>The directory shared by region servers.
>>    Should be fully-qualified to include the filesystem to use.
>>    E.g: hdfs://NAMENODE_SERVER:PORT/HBASE_ROOTDIR
>>    </description>
>>  </property>
>>
>>  <property>
>>    <name>hbase.master</name>
>>    <value>10.42.253.182:60000</value>
>>    <description>The host and port that the HBase master runs at.
>>    A value of 'local' runs the master and a regionserver in
>>    a single process.
>>    </description>
>>  </property>
>>  <property>
>>    <name>hbase.zookeeper.quorum</name>
>>    <value>convera</value>
>>     <description>Comma separated list of servers in the ZooKeeper Quorum.
>>    For example,
>> "host1.mydomain.com,host2.mydomain.com,host3.mydomain.com".
>>    By default this is set to localhost for local and pseudo-distributed
>> modes
>>    of operation. For a fully-distributed setup, this should be set to a
>> full
>>    list of ZooKeeper quorum servers. If HBASE_MANAGES_ZK is set in
>> hbase-env.sh
>>    this is the list of servers which we will start/stop ZooKeeper on.
>>    </description>
>>  </property>
>>
>> <property>
>>     <name>hbase.zookeeper.property.maxClientCnxns</name>
>>    <value>30</value>
>>    <description>Property from ZooKeeper's config zoo.cfg.
>>    Limit on number of concurrent connections (at the socket level) that a
>>    single client, identified by IP address, may make to a single member
> of
>>    the ZooKeeper ensemble. Set high to avoid zk connection issues running
>>    standalone and pseudo-distributed.
>>    </description>
>>  </property>
>>
>> </configuration>
>>
>>
>>
>>
>>
>>
>>
>>                      Amandeep Khurana
>>                      <[email protected]        To:
>> [email protected]
>>                      >                        cc:      (bcc:
> Y_823910/TSMC)
>>                                               Subject: Re: HBase-0.20.0
>> multi read
>>                       2009/08/21 11:54
>>                       AM
>>                      Please respond to
>>                      hbase-user
>>
>>
>>
>>
>>
>>
>> You ideally want to have 3-5 servers outside the hbase servers... 1
>> server is not enough. That could to be causing you the trouble.
>>
>> Post logs from the master and the region server where the read failed.
>>
>> Also, what's your configuration? How many nodes, ram, cpus etc?
>>
>> On 8/20/09, [email protected] <[email protected]> wrote:
>> >
>> > Hi there,
>> >
>> > It worked well while I fired 5 threads to fetch data from HBASE,but
>> > it failed after I incresed to 6 threads.
>> > Although it showed some WARN, the thread job can't be done!
>> > My hbase is the latest version hbase0.20.
>> > I want to test HBase multi read performance.
>> > Any suggestion?
>> > Thank you
>> >
>> > Fleming
>> >
>> >
>> > hbase-env.sh
>> >    export HBASE_MANAGES_ZK=true
>> >
>> > 09/08/21 09:54:07 WARN zookeeper.ZooKeeperWrapper: Failed to create
>> /hbase
>> > -- check quorum servers, currently=10.42.253.182:2181
>> > org.apache.zookeeper.KeeperException$ConnectionLossException:
>> > KeeperErrorCode = ConnectionLoss for /hbase
>> >       at
>> > org.apache.zookeeper.KeeperException.create(KeeperException.java:90)
>> >       at
>> > org.apache.zookeeper.KeeperException.create(KeeperException.java:42)
>> >       at org.apache.zookeeper.ZooKeeper.create(ZooKeeper.java:522)
>> >       at
>> >
>>
>>
> org.apache.hadoop.hbase.zookeeper.ZooKeeperWrapper.ensureExists(ZooKeeperWrapper.java:342)
>
>>
>> >       at
>> >
>>
>>
> org.apache.hadoop.hbase.zookeeper.ZooKeeperWrapper.ensureParentExists(ZooKeeperWrapper.java:365)
>
>>
>> >       at
>> >
>>
>>
> org.apache.hadoop.hbase.zookeeper.ZooKeeperWrapper.checkOutOfSafeMode(ZooKeeperWrapper.java:478)
>
>>
>> >       at
>> >
>>
>>
> org.apache.hadoop.hbase.client.HConnectionManager$TableServers.locateRootRegion(HConnectionManager.java:846)
>
>>
>> >       at
>> >
>>
>>
> org.apache.hadoop.hbase.client.HConnectionManager$TableServers.locateRegion(HConnectionManager.java:515)
>
>>
>> >       at
>> >
>>
>>
> org.apache.hadoop.hbase.client.HConnectionManager$TableServers.locateRegion(HConnectionManager.java:491)
>
>>
>> >       at
>> >
>>
>>
> org.apache.hadoop.hbase.client.HConnectionManager$TableServers.locateRegionInMeta(HConnectionManager.java:565)
>
>>
>> >       at
>> >
>>
>>
> org.apache.hadoop.hbase.client.HConnectionManager$TableServers.locateRegion(HConnectionManager.java:524)
>
>>
>> >       at
>> >
>>
>>
> org.apache.hadoop.hbase.client.HConnectionManager$TableServers.locateRegion(HConnectionManager.java:491)
>
>>
>> >       at
>> >
>>
>>
> org.apache.hadoop.hbase.client.HConnectionManager$TableServers.locateRegionInMeta(HConnectionManager.java:565)
>
>>
>> >       at
>> >
>>
>>
> org.apache.hadoop.hbase.client.HConnectionManager$TableServers.locateRegion(HConnectionManager.java:528)
>
>>
>> >       at
>> >
>>
>>
> org.apache.hadoop.hbase.client.HConnectionManager$TableServers.locateRegion(HConnectionManager.java:491)
>
>>
>> >       at org.apache.hadoop.hbase.client.HTable.<init>(HTable.java:123)
>> >       at org.apache.hadoop.hbase.client.HTable.<init>(HTable.java:101)
>> >       at
>> > org.gridgain.examples.executor.FlowJob.getHBaseData(FlowJob.java:144)
>> >       at org.gridgain.examples.executor.FlowJob.call(FlowJob.java:78)
>> >       at org.gridgain.examples.executor.FlowJob.call(FlowJob.java:1)
>> >       at
>> >
>>
>>
> org.gridgain.grid.kernal.executor.GridExecutorCallableTask$1.execute(GridExecutorCallableTask.java:57)
>
>>
>> >       at
>> >
>>
>>
> org.gridgain.grid.kernal.processors.job.GridJobWorker.body(GridJobWorker.java:406)
>
>>
>> >       at
>> >
> org.gridgain.grid.util.runnable.GridRunnable$1.run(GridRunnable.java:142)
>> >       at
>> > java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:441)
>> >       at
>> java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303)
>> >       at java.util.concurrent.FutureTask.run(FutureTask.java:138)
>> >       at
>> > org.gridgain.grid.util.runnable.GridRunnable.run(GridRunnable.java:194)
>> >       at
>> >
>>
>>
> java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
>
>>
>> >       at
>> >
>>
>>
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
>
>>
>> >       at java.lang.Thread.run(Thread.java:619)
>> >
>>
> ---------------------------------------------------------------------------
>> >                                                          TSMC PROPERTY
>> >  This email communication (and any attachments) is proprietary
>> information
>> >  for the sole use of its
>> >  intended recipient. Any unauthorized review, use or distribution by
>> anyone
>> >  other than the intended
>> >  recipient is strictly prohibited.  If you are not the intended
>> recipient,
>> >  please notify the sender by
>> >  replying to this email, and then delete this email and any copies of
> it
>> >  immediately. Thank you.
>> >
>>
> ---------------------------------------------------------------------------
>> >
>> >
>> >
>> >
>>
>>
>> --
>>
>>
>> Amandeep Khurana
>> Computer Science Graduate Student
>> University of California, Santa Cruz
>>
>>
>>
>>
>>
>>
> ---------------------------------------------------------------------------
>>                                                          TSMC PROPERTY
>>  This email communication (and any attachments) is proprietary
> information
>>  for the sole use of its
>>  intended recipient. Any unauthorized review, use or distribution by
> anyone
>>  other than the intended
>>  recipient is strictly prohibited.  If you are not the intended
> recipient,
>>  please notify the sender by
>>  replying to this email, and then delete this email and any copies of it
>>  immediately. Thank you.
>>
>>
> ---------------------------------------------------------------------------
>>
>>
>>
>>
>
>
>
>
>  ---------------------------------------------------------------------------
>                                                         TSMC PROPERTY
>  This email communication (and any attachments) is proprietary information
>  for the sole use of its
>  intended recipient. Any unauthorized review, use or distribution by anyone
>  other than the intended
>  recipient is strictly prohibited.  If you are not the intended recipient,
>  please notify the sender by
>  replying to this email, and then delete this email and any copies of it
>  immediately. Thank you.
>  ---------------------------------------------------------------------------
>
>
>
>

Reply via email to