Hi Anil, thanks again for the useful info. My replies inline as well:

On 30/09/15 23:45, anil gupta wrote:
Please find my reply inline.

On Wed, Sep 30, 2015 at 3:29 PM, Konstantinos Kougios <[email protected] <mailto:[email protected]>> wrote:

    Thanks for the reply and the useful information Anil.

    I am aware of the difficulties of distributed joins and
    aggregations and that phoenix is a layer on top of hbase. It would
    be great if it could be configured to run the queries, even if it
    takes a lot of time for the queries to complete.

Anil: I think, it is doable. But, this might require little bit of hit & trial with HBase and Phoenix conf. I would start with increasing HBase and Phoenix timeouts.
I believe I did that and now my queries run for a long time without timeout exceptions (before failing):

  zookeeper.session.timeout = 9000000 # not sure if this should be so high
  hbase.client.pause = 10000
  hbase.client.retries.number = 2
  hbase.client.scanner.timeout.period = 6000000
  phoenix.query.timeoutMs = 60000000
  phoenix.query.keepAliveMs = 60000
  hbase.client.operation.timeout = 60000000
  hbase.client.backpressure.enabled = true
  hbase.client.retries.number = 1
  hbase.rpc.timeout = 6000000

failure might all be due to

2015-10-01 09:13:22,002 WARN [B.defaultRpcServer.handler=14,queue=2,port=16020] ipc.RpcServer: (responseTooSlow): {"call":"Scan(org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ScanRequest)","starttimems":1443685974126,"responsesize":10,"method":"Scan","processingtimems":1227740,"client":"192.168.0.11:34656","queuetimems":0,"class":"HRegionServer"} 2015-10-01 09:13:22,296 ERROR [IndexRpcServer.handler=15,queue=0,port=16020] ipc.RpcServer: Unexpected throwable object
java.lang.ArrayIndexOutOfBoundsException
        at org.apache.hadoop.hbase.util.Bytes.putBytes(Bytes.java:299)
at org.apache.hadoop.hbase.KeyValue.createByteArray(KeyValue.java:1102)
        at org.apache.hadoop.hbase.KeyValue.<init>(KeyValue.java:650)
        at org.apache.hadoop.hbase.KeyValue.<init>(KeyValue.java:578)
at org.apache.phoenix.util.KeyValueUtil.newKeyValue(KeyValueUtil.java:63) at org.apache.phoenix.cache.aggcache.SpillManager.getAggregators(SpillManager.java:204) at org.apache.phoenix.cache.aggcache.SpillManager.toCacheEntry(SpillManager.java:243) at org.apache.phoenix.cache.aggcache.SpillableGroupByCache$EntryIterator.next(SpillableGroupByCache.java:285) at org.apache.phoenix.cache.aggcache.SpillableGroupByCache$EntryIterator.next(SpillableGroupByCache.java:261) at org.apache.phoenix.cache.aggcache.SpillableGroupByCache$2.next(SpillableGroupByCache.java:364) at org.apache.phoenix.coprocessor.BaseRegionScanner.next(BaseRegionScanner.java:40) at org.apache.phoenix.coprocessor.BaseRegionScanner.nextRaw(BaseRegionScanner.java:60) at org.apache.phoenix.coprocessor.DelegateRegionScanner.nextRaw(DelegateRegionScanner.java:77) at org.apache.hadoop.hbase.regionserver.RSRpcServices.scan(RSRpcServices.java:2395) at org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:32205)
        at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114)
        at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130) at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
        at java.lang.Thread.run(Thread.java:745)

The exception appears frequently in the logs when I run the query.
Also an other issue I've noticed is that sqlline.py starts using the cpu quite a lot after running my query for a while. Is't that just the client and should just wait for the server to respond? I believe it is due of using all it's memory (4GB). Also I think it is writing a lot of files under /tmp ... is it doing part of the query???

Thanks




    I got mainly 2 tables of 170GB and 550GB. Aggregation queries on
    both fail and even make region servers crash (there is no info in
    the logs and still don't know why. My server proved to be rock
    stable so far on other things but you never know).

Anil: RS should not crash. Are you doing heavy writes along with full table scans at same time? In one of your email, i saw stack trace regarding Region split and compactions?
I used to, but now I am just trying to run a query without any other load on the server. Or create an index.


    I am doing full table scans only because so far I was unable to
    create the indexes. I tried async indexes too with the map reduce
    job to create them but it runs extremely slowly.

Anil: This doesnt not sounds good. I haven't use those yet. So, i wont be able to help debug the problem. Hopefully, someone else will be able to chime in.


    In theory full table scans are possible with hbase, so even if it
    was slow it shouldn't fail.

Anil: IMO, if you are doing full table scans, then maybe you should turn off blockCache for those queries. Basically, there is a lot of cache churn due to full table scans. Cache churn will lead to JVM GC's.


    My setup is a 64GB AMD opteron server with 16 cores. 3 lxc virtual
    machines as region servers with Xmx8G, each running on a 3TB
    7200rpm disk. So somehow I simulate 3x low spec servers with
    enough ram.

    Next thing I will try is give region servers 16GB of RAM. WIth 8GB
    they seem to have some memory pressure and I see some slow GC's in
    the logs.

Anil: 16GB ram should help in some cases. Try to disable blockcache for full table scans.


    Cheers





    On 30/09/15 21:18, anil gupta wrote:
    Hi Konstantinos,
    Please find my reply inline.

    On Wed, Sep 30, 2015 at 12:10 PM, Konstantinos Kougios
    <[email protected]
    <mailto:[email protected]>> wrote:

        Hi all,

        I had various issues with big tables while experimenting the
        couple last weeks.

        The thing that goes to my mind is that hbase (+phoenix) works
        only when there is a fairly powerful cluster and say 1/2 the
        data can fit into the combined servers memory and disks are
        fast (SSD?) as well. It doesn't seem to be able to work when
        tables are 2x as large as the memory allocated to region
        servers (frankly I think it is less)

    Anil: Phoenix is just a SQL layer over HBase. From the query in
    your previous emails, it seems like you are doing full table
    scans with group by clauses. IMO, HBase is not a DB to be used
    for full table scans. If 90% of your use cases are small range
    scan or gets then HBase should work nicely with Terabytes of
    data. I have a 40 TB table in prod on 60 node cluster where every
    RS only has 16GB of heap. What kind of workload you are trying to
    run with HBase?


        Things that constantly fail:

        - non-trivial queries on large tables (with group by, counts,
        joins) with region server out of memory errors or crashes
        without any reason for Xmx of 4G or 8G

    Anil: Can you convert these queries into short range based scans?
    If you are always going to do full table scan, then maybe you
    need to use MR or Spark for those computation and then tune
    cluster for full table scans. Cluster tuning varies with full
    table scan workload.

        - index creation on the same big tables. Those always fail I
        think around the point when hbase has to flush it's memory
        regions to the disk and couldn't find a solution

        - spark jobs fail unless they are throttled to feed hbase
        with the data it can take . No backpressure?


        There were no replies to my emails regarding the issues,
        which makes me think there aren't solutions (or solutions are
        pretty hard to find and not many ppl know them).

        So after 21 tweaks to the default config, I am still not able
        to operate it as a normal database.

    Anil: HBase is actually not a normal RDBMS DB. Its a **keyvalue
    store**. Phoenix is providing a SQL layer using HBase API. So,
    user will need to deal with pros/cons of a key/value store.


        Should I start believing my config is all wrong or that
        hbase+phoenix is only working if there is a sufficiently
        powerful cluster to handle the data?

    Anil: **As per my experience**, HBase+Phoenix will work nicely if
    you are doing keyvalue lookups and short range scans.
    I would suggest you to evaluate data model of HBase tables and
    try to convert queries to small range scan or lookups.


        I believe it is a great project and the functionality is
        really useful. What's lacking is 3 sample configs for 3
        different strength clusters.

    Anil: I agree that guidance on configuration of HBase and Phoenix
    can be improved so that people can get going quickly.


        Thanks




-- Thanks & Regards,
    Anil Gupta




--
Thanks & Regards,
Anil Gupta

Reply via email to