[ 
https://issues.apache.org/jira/browse/PHOENIX-1215?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vikas Vishwakarma resolved PHOENIX-1215.
----------------------------------------
    Resolution: Invalid

This turned out to be a issue in our client implementation. Closing.

> some phoenix transactions are hanging indefinitely on ps.execute() call
> -----------------------------------------------------------------------
>
>                 Key: PHOENIX-1215
>                 URL: https://issues.apache.org/jira/browse/PHOENIX-1215
>             Project: Phoenix
>          Issue Type: Bug
>    Affects Versions: 4.0.0
>         Environment: Num Cores: 24, Model : Intel(R) Xeon(R) CPU E5-2420 0 @ 
> 1.90GHz
> MemTotal:       65801264 kB, 1 Gbps Network
> Red Hat Enterprise Linux Server release 6.2
>            Reporter: Vikas Vishwakarma
>
> Test Setup:
> ==========
> 4 load client with 16 threads each running against a HBase cluster with 6 
> RegionServers. 
> Each load client spawns a thread and runs a upsert query in batch mode
> Pseudo code:
> ==========
> SQL_UPSERT_HBASE_DATA = UPSERT INTO TABLE_NAME  (ROW_KEY_ID, ROW_VAL) VALUES 
> (?, ?)
>     for (int i=0; i<6400; i++) {
>       upsertSql.append(SQL_UPSERT_HBASE_DATA);
>       ps = conn.prepareStatement(preparedSqlTableStmt.toString());
>       ps.setString(1, rowKey);
>       ps.setString(2, rowVal);
>       ps.execute();
>     }
>     conn.commit();
>     tryClose(conn);
> Observation:
> ==============
> Observing some issues where a small number of transactions (approx 40 out of 
> 80,000) are hanging indefinitely on preparedStatement.execute() call. 
> I tried to put a timeout on the transaction but looks like phoenix still does 
> not support preparedStatement.setQueryTimeout().
> Log Analysis:
> =============
> On the client side there are no errors or exceptions except some threads in 
> hung state 
> On the service side, the observation is as follows:
> Also after some log analysis,  it looks like all the batches that got stuck, 
> are during the start of the test. From the RegionServer logs, the phoenix 
> co-processor loading and index updates is in progress during this time (which 
> involves some failures and retries). Once this phase is over, remaining write 
> batches that started later have succeeded.
> RegionServer log snapshots during the time period this issue was observed on 
> the client side:
> ---------
> 2014-08-27 03:28:00,114 INFO 
> org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost: Loaded 
> coprocessor org.apache.phoenix.coprocessor.ScanRegionObserver from HTD of 
> HT_42 successfully.
> 2014-08-27 03:28:00,114 INFO 
> org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost: Loaded 
> coprocessor org.apache.phoenix.coprocessor.ScanRegionObserver from HTD of 
> HT_42 successfully.
> 2014-08-27 03:28:00,252 INFO org.apache.phoenix.hbase.index.Indexer: Found 
> some outstanding index updates that didn't succeed during WAL replay - 
> attempting to replay now.
> 2014-08-27 03:28:00,252 INFO org.apache.phoenix.hbase.index.Indexer: Found 
> some outstanding index updates that didn't succeed during WAL replay - 
> attempting to replay now.
> 2014-08-27 03:28:00,252 INFO org.apache.phoenix.hbase.index.Indexer: Found 
> some outstanding index updates that didn't succeed during WAL replay - 
> attempting to replay now.
> 2014-08-27 03:28:00,931 INFO 
> org.apache.phoenix.coprocessor.UngroupedAggregateRegionObserver: Starting 
> ungrouped coprocessor scan 
> {"timeRange":[0,1409110080691],"batch":-1,"startRow":"\\x0E","stopRow":"\\x0F","loadColumnFamiliesOnDemand":true,"totalColumns":1,"cacheBlocks":true,"families":{"0":["ALL"]},"maxResultSize":-1,"maxVersions":1,"caching":-1}
> 2014-08-27 03:28:00,931 INFO 
> org.apache.phoenix.coprocessor.UngroupedAggregateRegionObserver: Starting 
> ungrouped coprocessor scan 
> {"timeRange":[0,1409110080691],"batch":-1,"startRow":"\\x07","stopRow":"\\x08","loadColumnFamiliesOnDemand":true,"totalColumns":1,"cacheBlocks":true,"families":{"0":["ALL"]},"maxResultSize":-1,"maxVersions":1,"caching":-1}
> ....
> 2014-08-27 03:28:06,507 DEBUG 
> org.apache.hadoop.hbase.coprocessor.CoprocessorHost: Loading coprocessor 
> class org.apache.phoenix.coprocessor.ServerCachingEndpointImpl with path null 
> and priority 1
> 2014-08-27 03:28:06,507 INFO 
> org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost: Loaded 
> coprocessor org.apache.phoenix.coprocessor.ServerCachingEndpointImpl from HTD 
> of HT_44 successfully.
> ....
> 2014-08-27 03:28:06,508 DEBUG 
> org.apache.hadoop.hbase.coprocessor.CoprocessorHost: Loading coprocessor 
> class org.apache.phoenix.coprocessor.GroupedAggregateRegionObserver with path 
> null and priority 1
> 2014-08-27 03:28:06,508 INFO 
> org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost: Loaded 
> coprocessor org.apache.phoenix.coprocessor.GroupedAggregateRegionObserver 
> from HTD of HT_44 successfully.
> ...
> 2014-08-27 03:28:19,525 INFO org.apache.phoenix.hbase.index.Indexer: Found 
> some outstanding index updates that didn't succeed during WAL replay - 
> attempting to replay now.
> 2014-08-27 03:28:19,547 INFO org.apache.phoenix.hbase.index.Indexer: Found 
> some outstanding index updates that didn't succeed during WAL replay - 
> attempting to replay now.
> ..
> 2014-08-27 03:28:20,660 INFO 
> org.apache.phoenix.coprocessor.UngroupedAggregateRegionObserver: Starting 
> ungrouped coprocessor scan 
> {"timeRange":[0,1409110100554],"batch":-1,"startRow":"\\x0C","stopRow":"\\x0D","loadColumnFamiliesOnDemand":true,"totalColumns":1,"cacheBlocks":true,"families":{"0":["ALL"]},"maxResultSize":-1,"maxVersions":1,"caching":-1}
> 2014-08-27 03:28:20,660 INFO 
> org.apache.phoenix.coprocessor.UngroupedAggregateRegionObserver: Finished 
> scanning 0 rows for ungrouped coprocessor scan 
> {"timeRange":[0,1409110100554],"batch":-1,"startRow":"\\x0C","stopRow":"\\x0D","loadColumnFamiliesOnDemand":true,"totalColumns":1,"cacheBlocks":true,"families":{"0":["ALL"]},"maxResultSize":-1,"maxVersions":1,"caching":-1}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to