Hi Maryann, Its the same query:
*select c.c_first_name, ca.ca_city, cd.cd_education_status from CUSTOMER_30000 c join CUSTOMER_DEMOGRAPHICS_1 cd on c.c_current_cdemo_sk = cd.cd_demo_sk join CUSTOMER_ADDRESS_1 ca on c.c_current_addr_sk = ca.ca_address_sk group by ca.ca_city, cd.cd_education_status, c.c_first_name;* *The size of CUSTOMER_30000 is 4.1 GB with 30million records. The CUSTOMER_DEMOGRAPHICS contains 2M records and CUSTOMER_ADDRESS contains 50000 records.* *Regards,* *Vijay Raajaa G S* On Tue, Oct 7, 2014 at 9:39 PM, Maryann Xue <maryann....@gmail.com> wrote: > Hi Ashish, > > The warning you got was exactly showing the reason why you finally got > that error: one of the join table query had taken too long so that the > cache for other join tables expired and got invalidated. Again, could you > please share your query and the size of the tables used in your query? > Instead of changing the parameters to get around the problems, it might be > much more efficient just to adjust the query itself. And if doable, mostly > likely the query is gonna run faster as well. > > Besides, you might find this document helpful : > http://phoenix.apache.org/joins.html > > > Thanks, > Maryann > > > On Tue, Oct 7, 2014 at 11:01 AM, ashish tapdiya <ashishtapd...@gmail.com> > wrote: > >> Maryann, >> >> hbase-site.xml was not on CLASSPATH and that was the issue. Thanks for >> the help. I appreciate it. >> >> ~Ashish >> >> >> >> On Sat, Oct 4, 2014 at 3:40 PM, Maryann Xue <maryann....@gmail.com> >> wrote: >> >>> Hi Ashish, >>> >>> The "phoenix.query.maxServerCacheBytes" is a client parameter while the >>> other two are server parameters. But it looks like the configuration change >>> did not take effect at your client side. Could you please make sure that >>> this is the only configuration that goes to the CLASSPATH of your phoenix >>> client execution environment? >>> >>> Another thing is the exception you got was a different problem from >>> Vijay's. It happened in an even earlier stage. Could you please also share >>> you query? We could probably re-write it so that it can better fit the >>> hash-join scheme. (Since table stats are not used in joins yet, we >>> currently have to do it manually.) >>> >>> >>> Thanks, >>> Maryann >>> >>> On Tue, Sep 30, 2014 at 1:22 PM, ashish tapdiya <ashishtapd...@gmail.com >>> > wrote: >>> >>>> Here it is, >>>> >>>> java.sql.SQLException: Encountered exception in hash plan [1] execution. >>>> at >>>> org.apache.phoenix.execute.HashJoinPlan.iterator(HashJoinPlan.java:146) >>>> at >>>> org.apache.phoenix.jdbc.PhoenixStatement$1.call(PhoenixStatement.java:211) >>>> at >>>> org.apache.phoenix.jdbc.PhoenixStatement$1.call(PhoenixStatement.java:204) >>>> at >>>> org.apache.phoenix.util.PhoenixContextExecutor.call(PhoenixContextExecutor.java:54) >>>> at >>>> org.apache.phoenix.jdbc.PhoenixStatement.executeQuery(PhoenixStatement.java:204) >>>> at >>>> org.apache.phoenix.jdbc.PhoenixPreparedStatement.executeQuery(PhoenixPreparedStatement.java:158) >>>> at Query.sel_Cust_Order_OrderLine_Tables(Query.java:135) >>>> at Query.main(Query.java:25) >>>> Caused by: org.apache.phoenix.join.MaxServerCacheSizeExceededException: >>>> Size of hash cache (104857684 bytes) exceeds the maximum allowed size >>>> (104857600 bytes) >>>> at >>>> org.apache.phoenix.join.HashCacheClient.serialize(HashCacheClient.java:106) >>>> at >>>> org.apache.phoenix.join.HashCacheClient.addHashCache(HashCacheClient.java:77) >>>> at >>>> org.apache.phoenix.execute.HashJoinPlan$1.call(HashJoinPlan.java:119) >>>> at >>>> org.apache.phoenix.execute.HashJoinPlan$1.call(HashJoinPlan.java:114) >>>> at java.util.concurrent.FutureTask.run(FutureTask.java:262) >>>> at >>>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) >>>> at >>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) >>>> at java.lang.Thread.run(Thread.java:744) >>>> >>>> I am setting hbase heap to 4 GB and phoenix properties are set as below >>>> >>>> <property> >>>> <name>phoenix.query.maxServerCacheBytes</name> >>>> <value>2004857600</value> >>>> </property> >>>> <property> >>>> <name>phoenix.query.maxGlobalMemoryPercentage</name> >>>> <value>40</value> >>>> </property> >>>> <property> >>>> <name>phoenix.query.maxGlobalMemorySize</name> >>>> <value>1504857600</value> >>>> </property> >>>> >>>> Thanks, >>>> ~Ashish >>>> >>>> On Tue, Sep 30, 2014 at 12:13 PM, Maryann Xue <maryann....@gmail.com> >>>> wrote: >>>> >>>>> Hi Ashish, >>>>> >>>>> Could you please let us see your error message? >>>>> >>>>> >>>>> Thanks, >>>>> Maryann >>>>> >>>>> On Tue, Sep 30, 2014 at 12:58 PM, ashish tapdiya < >>>>> ashishtapd...@gmail.com> wrote: >>>>> >>>>>> Hey Maryann, >>>>>> >>>>>> Thanks for your input. I tried both the properties but no luck. >>>>>> >>>>>> ~Ashish >>>>>> >>>>>> On Sun, Sep 28, 2014 at 8:31 PM, Maryann Xue <maryann....@gmail.com> >>>>>> wrote: >>>>>> >>>>>>> Hi Ashish, >>>>>>> >>>>>>> The global cache size is set to either " >>>>>>> *phoenix.query.maxGlobalMemorySize*" or >>>>>>> "phoenix.query.maxGlobalMemoryPercentage >>>>>>> * heapSize" (Sorry about the mistake I made earlier). The "" >>>>>>> phoenix.query.maxServerCacheBytes" is a client parameter and is >>>>>>> most likely NOT the thing you should worry about. So you can try >>>>>>> adjusting " >>>>>>> phoenix.query.maxGlobalMemoryPercentage" and the heap size in >>>>>>> region server configurations and see how it works. >>>>>>> >>>>>>> >>>>>>> Thanks, >>>>>>> Maryann >>>>>>> >>>>>>> On Fri, Sep 26, 2014 at 10:48 PM, ashish tapdiya < >>>>>>> ashishtapd...@gmail.com> wrote: >>>>>>> >>>>>>>> I have tried that as well...but "phoenix.query.maxServerCacheBytes" >>>>>>>> remains the default value of 100 MB. I get to see it when join fails. >>>>>>>> >>>>>>>> Thanks, >>>>>>>> ~Ashish >>>>>>>> >>>>>>>> On Fri, Sep 26, 2014 at 8:02 PM, Maryann Xue <maryann....@gmail.com >>>>>>>> > wrote: >>>>>>>> >>>>>>>>> Hi Ashish, >>>>>>>>> >>>>>>>>> The global cache size is set to either >>>>>>>>> "phoenix.query.maxServerCacheBytes" >>>>>>>>> or "phoenix.query.maxGlobalMemoryPercentage * heapSize", >>>>>>>>> whichever is *smaller*. You can try setting "phoenix.query. >>>>>>>>> maxGlobalMemoryPercentage" instead, which is recommended, and see >>>>>>>>> how it goes. >>>>>>>>> >>>>>>>>> >>>>>>>>> Thanks, >>>>>>>>> Maryann >>>>>>>>> >>>>>>>>> On Fri, Sep 26, 2014 at 5:37 PM, ashish tapdiya < >>>>>>>>> ashishtapd...@gmail.com> wrote: >>>>>>>>> >>>>>>>>>> Hi Maryann, >>>>>>>>>> >>>>>>>>>> I am having the same issue where star join is failing with >>>>>>>>>> MaxServerCacheSizeExceededException. >>>>>>>>>> I set phoenix.query.maxServerCacheBytes to 1 GB both in client >>>>>>>>>> and server hbase-site.xml's. However, it does not take effect. >>>>>>>>>> >>>>>>>>>> Phoenix 3.1 >>>>>>>>>> HBase .94 >>>>>>>>>> >>>>>>>>>> Thanks, >>>>>>>>>> ~Ashish >>>>>>>>>> >>>>>>>>>> On Fri, Sep 26, 2014 at 2:56 PM, Maryann Xue < >>>>>>>>>> maryann....@gmail.com> wrote: >>>>>>>>>> >>>>>>>>>>> Yes, you should make your modification on each region server, >>>>>>>>>>> since this is a server-side configuration. >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> On Thu, Sep 25, 2014 at 4:15 AM, G.S.Vijay Raajaa < >>>>>>>>>>> gsvijayraa...@gmail.com> wrote: >>>>>>>>>>> >>>>>>>>>>>> Hi Xue, >>>>>>>>>>>> >>>>>>>>>>>> Thanks for replying. I did modify the hbase-site.xml >>>>>>>>>>>> by increasing the default value of >>>>>>>>>>>> phoenix.query.maxGlobalMemoryPercentage >>>>>>>>>>>> . Also increased the Region server heap space memory . The >>>>>>>>>>>> change didn't get reflected and I still get the error with an >>>>>>>>>>>> indication >>>>>>>>>>>> that "global pool of 319507660 bytes" is present. Should I >>>>>>>>>>>> modify the hbase-site.xml in every region server or just the file >>>>>>>>>>>> present >>>>>>>>>>>> in the class path of Phoenix client? >>>>>>>>>>>> >>>>>>>>>>>> Regards, >>>>>>>>>>>> Vijay Raajaa G S >>>>>>>>>>>> >>>>>>>>>>>> On Thu, Sep 25, 2014 at 1:47 AM, Maryann Xue < >>>>>>>>>>>> maryann....@gmail.com> wrote: >>>>>>>>>>>> >>>>>>>>>>>>> Hi Vijay, >>>>>>>>>>>>> >>>>>>>>>>>>> I think here the query plan is scanning table *CUSTOMER_30000 >>>>>>>>>>>>> *while joining the other two tables at the same time, which >>>>>>>>>>>>> means the region server memory for Phoenix should be large enough >>>>>>>>>>>>> to hold 2 >>>>>>>>>>>>> tables together and you also need to expect some memory expansion >>>>>>>>>>>>> for java >>>>>>>>>>>>> objects. >>>>>>>>>>>>> >>>>>>>>>>>>> Do you mean that after you had modified the parameters you >>>>>>>>>>>>> mentioned, you were still getting the same error message with >>>>>>>>>>>>> exactly the >>>>>>>>>>>>> same numbers as "global pool of 319507660 bytes"? Did you >>>>>>>>>>>>> make sure that the parameters actually took effect after >>>>>>>>>>>>> modification? >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> Thanks, >>>>>>>>>>>>> Maryann >>>>>>>>>>>>> >>>>>>>>>>>>> On Tue, Sep 23, 2014 at 1:43 AM, G.S.Vijay Raajaa < >>>>>>>>>>>>> gsvijayraa...@gmail.com> wrote: >>>>>>>>>>>>> >>>>>>>>>>>>>> Hi, >>>>>>>>>>>>>> >>>>>>>>>>>>>> I am trying to do a join of three tables usng the >>>>>>>>>>>>>> following query: >>>>>>>>>>>>>> >>>>>>>>>>>>>> *select c.c_first_name, ca.ca_city, cd.cd_education_status >>>>>>>>>>>>>> from CUSTOMER_30000 c join CUSTOMER_DEMOGRAPHICS_1 cd on >>>>>>>>>>>>>> c.c_current_cdemo_sk = cd.cd_demo_sk join CUSTOMER_ADDRESS_1 ca >>>>>>>>>>>>>> on >>>>>>>>>>>>>> c.c_current_addr_sk = ca.ca_address_sk group by ca.ca_city, >>>>>>>>>>>>>> cd.cd_education_status, c.c_first_name;* >>>>>>>>>>>>>> >>>>>>>>>>>>>> *The size of CUSTOMER_30000 is 4.1 GB with 30million records.* >>>>>>>>>>>>>> >>>>>>>>>>>>>> *I get the following error:* >>>>>>>>>>>>>> >>>>>>>>>>>>>> ./psql.py 10.10.5.55 test.sql >>>>>>>>>>>>>> java.sql.SQLException: Encountered exception in hash plan [0] >>>>>>>>>>>>>> execution. >>>>>>>>>>>>>> at >>>>>>>>>>>>>> org.apache.phoenix.execute.HashJoinPlan.iterator(HashJoinPlan.java:146) >>>>>>>>>>>>>> at >>>>>>>>>>>>>> org.apache.phoenix.jdbc.PhoenixStatement$1.call(PhoenixStatement.java:211) >>>>>>>>>>>>>> at >>>>>>>>>>>>>> org.apache.phoenix.jdbc.PhoenixStatement$1.call(PhoenixStatement.java:204) >>>>>>>>>>>>>> at >>>>>>>>>>>>>> org.apache.phoenix.util.PhoenixContextExecutor.call(PhoenixContextExecutor.java:54) >>>>>>>>>>>>>> at >>>>>>>>>>>>>> org.apache.phoenix.jdbc.PhoenixStatement.executeQuery(PhoenixStatement.java:204) >>>>>>>>>>>>>> at >>>>>>>>>>>>>> org.apache.phoenix.jdbc.PhoenixStatement.execute(PhoenixStatement.java:193) >>>>>>>>>>>>>> at >>>>>>>>>>>>>> org.apache.phoenix.jdbc.PhoenixPreparedStatement.execute(PhoenixPreparedStatement.java:147) >>>>>>>>>>>>>> at >>>>>>>>>>>>>> org.apache.phoenix.jdbc.PhoenixPreparedStatement.execute(PhoenixPreparedStatement.java:152) >>>>>>>>>>>>>> at >>>>>>>>>>>>>> org.apache.phoenix.jdbc.PhoenixConnection.executeStatements(PhoenixConnection.java:220) >>>>>>>>>>>>>> at >>>>>>>>>>>>>> org.apache.phoenix.util.PhoenixRuntime.executeStatements(PhoenixRuntime.java:193) >>>>>>>>>>>>>> at >>>>>>>>>>>>>> org.apache.phoenix.util.PhoenixRuntime.main(PhoenixRuntime.java:140) >>>>>>>>>>>>>> Caused by: java.sql.SQLException: >>>>>>>>>>>>>> java.util.concurrent.ExecutionException: >>>>>>>>>>>>>> java.lang.reflect.UndeclaredThrowableException >>>>>>>>>>>>>> at >>>>>>>>>>>>>> org.apache.phoenix.cache.ServerCacheClient.addServerCache(ServerCacheClient.java:199) >>>>>>>>>>>>>> at >>>>>>>>>>>>>> org.apache.phoenix.join.HashCacheClient.addHashCache(HashCacheClient.java:78) >>>>>>>>>>>>>> at >>>>>>>>>>>>>> org.apache.phoenix.execute.HashJoinPlan$1.call(HashJoinPlan.java:119) >>>>>>>>>>>>>> at >>>>>>>>>>>>>> org.apache.phoenix.execute.HashJoinPlan$1.call(HashJoinPlan.java:114) >>>>>>>>>>>>>> at >>>>>>>>>>>>>> java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) >>>>>>>>>>>>>> at java.util.concurrent.FutureTask.run(FutureTask.java:138) >>>>>>>>>>>>>> at >>>>>>>>>>>>>> java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) >>>>>>>>>>>>>> at >>>>>>>>>>>>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) >>>>>>>>>>>>>> at java.lang.Thread.run(Thread.java:662) >>>>>>>>>>>>>> Caused by: java.util.concurrent.ExecutionException: >>>>>>>>>>>>>> java.lang.reflect.UndeclaredThrowableException >>>>>>>>>>>>>> at >>>>>>>>>>>>>> java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) >>>>>>>>>>>>>> at java.util.concurrent.FutureTask.get(FutureTask.java:91) >>>>>>>>>>>>>> at >>>>>>>>>>>>>> org.apache.phoenix.cache.ServerCacheClient.addServerCache(ServerCacheClient.java:191) >>>>>>>>>>>>>> ... 8 more >>>>>>>>>>>>>> Caused by: java.lang.reflect.UndeclaredThrowableException >>>>>>>>>>>>>> at $Proxy10.addServerCache(Unknown Source) >>>>>>>>>>>>>> at >>>>>>>>>>>>>> org.apache.phoenix.cache.ServerCacheClient$1.call(ServerCacheClient.java:169) >>>>>>>>>>>>>> at >>>>>>>>>>>>>> org.apache.phoenix.cache.ServerCacheClient$1.call(ServerCacheClient.java:164) >>>>>>>>>>>>>> ... 5 more >>>>>>>>>>>>>> Caused by: >>>>>>>>>>>>>> org.apache.hadoop.hbase.client.RetriesExhaustedException: Failed >>>>>>>>>>>>>> after >>>>>>>>>>>>>> attempts=14, exceptions: >>>>>>>>>>>>>> Tue Sep 23 00:25:53 CDT 2014, >>>>>>>>>>>>>> org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@100e398, >>>>>>>>>>>>>> java.io.IOException: java.io.IOException: >>>>>>>>>>>>>> org.apache.phoenix.memory.InsufficientMemoryException: Requested >>>>>>>>>>>>>> memory of >>>>>>>>>>>>>> 446623727 bytes is larger than global pool of 319507660 bytes. >>>>>>>>>>>>>> Tue Sep 23 00:26:02 CDT 2014, >>>>>>>>>>>>>> org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@100e398, >>>>>>>>>>>>>> java.io.IOException: java.io.IOException: >>>>>>>>>>>>>> org.apache.phoenix.memory.InsufficientMemoryException: Requested >>>>>>>>>>>>>> memory of >>>>>>>>>>>>>> 446623727 bytes is larger than global pool of 319507660 bytes. >>>>>>>>>>>>>> Tue Sep 23 00:26:18 CDT 2014, >>>>>>>>>>>>>> org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@100e398, >>>>>>>>>>>>>> java.io.IOException: java.io.IOException: >>>>>>>>>>>>>> org.apache.phoenix.memory.InsufficientMemoryException: Requested >>>>>>>>>>>>>> memory of >>>>>>>>>>>>>> 446623727 bytes is larger than global pool of 319507660 bytes. >>>>>>>>>>>>>> Tue Sep 23 00:26:43 CDT 2014, >>>>>>>>>>>>>> org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@100e398, >>>>>>>>>>>>>> java.io.IOException: java.io.IOException: >>>>>>>>>>>>>> org.apache.phoenix.memory.InsufficientMemoryException: Requested >>>>>>>>>>>>>> memory of >>>>>>>>>>>>>> 446623727 bytes is larger than global pool of 319507660 bytes. >>>>>>>>>>>>>> Tue Sep 23 00:27:01 CDT 2014, >>>>>>>>>>>>>> org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@100e398, >>>>>>>>>>>>>> java.io.IOException: java.io.IOException: >>>>>>>>>>>>>> org.apache.phoenix.memory.InsufficientMemoryException: Requested >>>>>>>>>>>>>> memory of >>>>>>>>>>>>>> 446623727 bytes is larger than global pool of 319507660 bytes. >>>>>>>>>>>>>> Tue Sep 23 00:27:10 CDT 2014, >>>>>>>>>>>>>> org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@100e398, >>>>>>>>>>>>>> java.io.IOException: java.io.IOException: >>>>>>>>>>>>>> org.apache.phoenix.memory.InsufficientMemoryException: Requested >>>>>>>>>>>>>> memory of >>>>>>>>>>>>>> 446623727 bytes is larger than global pool of 319507660 bytes. >>>>>>>>>>>>>> Tue Sep 23 00:27:24 CDT 2014, >>>>>>>>>>>>>> org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@100e398, >>>>>>>>>>>>>> java.io.IOException: java.io.IOException: >>>>>>>>>>>>>> org.apache.phoenix.memory.InsufficientMemoryException: Requested >>>>>>>>>>>>>> memory of >>>>>>>>>>>>>> 446623727 bytes is larger than global pool of 319507660 bytes. >>>>>>>>>>>>>> Tue Sep 23 00:28:16 CDT 2014, >>>>>>>>>>>>>> org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@100e398, >>>>>>>>>>>>>> java.io.IOException: java.io.IOException: >>>>>>>>>>>>>> org.apache.phoenix.memory.InsufficientMemoryException: Requested >>>>>>>>>>>>>> memory of >>>>>>>>>>>>>> 446623727 bytes is larger than global pool of 319507660 bytes. >>>>>>>>>>>>>> Tue Sep 23 00:28:35 CDT 2014, >>>>>>>>>>>>>> org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@100e398, >>>>>>>>>>>>>> java.io.IOException: java.io.IOException: >>>>>>>>>>>>>> org.apache.phoenix.memory.InsufficientMemoryException: Requested >>>>>>>>>>>>>> memory of >>>>>>>>>>>>>> 446623727 bytes is larger than global pool of 319507660 bytes. >>>>>>>>>>>>>> Tue Sep 23 00:29:09 CDT 2014, >>>>>>>>>>>>>> org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@100e398, >>>>>>>>>>>>>> java.io.IOException: java.io.IOException: >>>>>>>>>>>>>> org.apache.phoenix.memory.InsufficientMemoryException: Requested >>>>>>>>>>>>>> memory of >>>>>>>>>>>>>> 446623727 bytes is larger than global pool of 319507660 bytes. >>>>>>>>>>>>>> Tue Sep 23 00:30:16 CDT 2014, >>>>>>>>>>>>>> org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@100e398, >>>>>>>>>>>>>> java.io.IOException: java.io.IOException: >>>>>>>>>>>>>> org.apache.phoenix.memory.InsufficientMemoryException: Requested >>>>>>>>>>>>>> memory of >>>>>>>>>>>>>> 446623727 bytes is larger than global pool of 319507660 bytes. >>>>>>>>>>>>>> Tue Sep 23 00:31:22 CDT 2014, >>>>>>>>>>>>>> org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@100e398, >>>>>>>>>>>>>> java.io.IOException: java.io.IOException: >>>>>>>>>>>>>> org.apache.phoenix.memory.InsufficientMemoryException: Requested >>>>>>>>>>>>>> memory of >>>>>>>>>>>>>> 446623727 bytes is larger than global pool of 319507660 bytes. >>>>>>>>>>>>>> Tue Sep 23 00:32:29 CDT 2014, >>>>>>>>>>>>>> org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@100e398, >>>>>>>>>>>>>> java.io.IOException: java.io.IOException: >>>>>>>>>>>>>> org.apache.phoenix.memory.InsufficientMemoryException: Requested >>>>>>>>>>>>>> memory of >>>>>>>>>>>>>> 446623727 bytes is larger than global pool of 319507660 bytes. >>>>>>>>>>>>>> Tue Sep 23 00:33:35 CDT 2014, >>>>>>>>>>>>>> org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@100e398, >>>>>>>>>>>>>> java.io.IOException: java.io.IOException: >>>>>>>>>>>>>> org.apache.phoenix.memory.InsufficientMemoryException: Requested >>>>>>>>>>>>>> memory of >>>>>>>>>>>>>> 446623727 bytes is larger than global pool of 319507660 bytes. >>>>>>>>>>>>>> >>>>>>>>>>>>>> at >>>>>>>>>>>>>> org.apache.hadoop.hbase.client.ServerCallable.withRetries(ServerCallable.java:187) >>>>>>>>>>>>>> at >>>>>>>>>>>>>> org.apache.hadoop.hbase.ipc.ExecRPCInvoker.invoke(ExecRPCInvoker.java:79) >>>>>>>>>>>>>> ... 8 more >>>>>>>>>>>>>> >>>>>>>>>>>>>> Trials: >>>>>>>>>>>>>> >>>>>>>>>>>>>> I tried to increase the Region Server Heap space , >>>>>>>>>>>>>> modified phoenix.query.maxGlobalMemoryPercentage as well. >>>>>>>>>>>>>> >>>>>>>>>>>>>> I am not able to increase the global memory . >>>>>>>>>>>>>> >>>>>>>>>>>>>> Regards, >>>>>>>>>>>>>> Vijay Raajaa >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> -- >>>>>>>>>>>>> Thanks, >>>>>>>>>>>>> Maryann >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> -- >>>>>>>>>>> Thanks, >>>>>>>>>>> Maryann >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> -- >>>>>>>>> Thanks, >>>>>>>>> Maryann >>>>>>>>> >>>>>>>> >>>>>>>> >>>>>>> >>>>>>> >>>>>>> -- >>>>>>> Thanks, >>>>>>> Maryann >>>>>>> >>>>>> >>>>>> >>>>> >>>>> >>>>> -- >>>>> Thanks, >>>>> Maryann >>>>> >>>> >>>> >>> >>> >>> -- >>> Thanks, >>> Maryann >>> >> >> > > > -- > Thanks, > Maryann >