Hi Maryann,

                 After executing the same query after increasing the heap
space in region server, I get a strange error:

./psql.py 10.10.5.55 test.sql
14/10/07 02:59:52 WARN execute.HashJoinPlan: Hash plan [0] execution seems
too slow. Earlier hash cache(s) might have expired on servers.
org.apache.phoenix.exception.PhoenixIOException:
org.apache.phoenix.exception.PhoenixIOException:
org.apache.hadoop.hbase.DoNotRetryIOException: Could not find hash cache
for joinId: {/�PY�L�. The cache might have expired and have been removed.
at
org.apache.phoenix.util.ServerUtil.parseServerException(ServerUtil.java:97)
at
org.apache.phoenix.iterate.ParallelIterators.getIterators(ParallelIterators.java:279)
at
org.apache.phoenix.iterate.MergeSortResultIterator.getIterators(MergeSortResultIterator.java:48)
at
org.apache.phoenix.iterate.MergeSortResultIterator.minIterator(MergeSortResultIterator.java:63)
at
org.apache.phoenix.iterate.MergeSortResultIterator.next(MergeSortResultIterator.java:90)
at
org.apache.phoenix.iterate.GroupedAggregatingResultIterator.next(GroupedAggregatingResultIterator.java:68)
at
org.apache.phoenix.iterate.DelegateResultIterator.next(DelegateResultIterator.java:40)
at org.apache.phoenix.jdbc.PhoenixResultSet.next(PhoenixResultSet.java:732)
at
org.apache.phoenix.jdbc.PhoenixConnection.executeStatements(PhoenixConnection.java:223)
at
org.apache.phoenix.util.PhoenixRuntime.executeStatements(PhoenixRuntime.java:193)
at org.apache.phoenix.util.PhoenixRuntime.main(PhoenixRuntime.java:140)
Caused by: java.util.concurrent.ExecutionException:
org.apache.phoenix.exception.PhoenixIOException:
org.apache.hadoop.hbase.DoNotRetryIOException: Could not find hash cache
for joinId: {/�PY�L�. The cache might have expired and have been removed.
at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232)
at java.util.concurrent.FutureTask.get(FutureTask.java:91)
at
org.apache.phoenix.iterate.ParallelIterators.getIterators(ParallelIterators.java:275)
... 9 more
Caused by: org.apache.phoenix.exception.PhoenixIOException:
org.apache.hadoop.hbase.DoNotRetryIOException: Could not find hash cache
for joinId: {/�PY�L�. The cache might have expired and have been removed.
at
org.apache.phoenix.util.ServerUtil.parseServerException(ServerUtil.java:97)
at
org.apache.phoenix.iterate.TableResultIterator.<init>(TableResultIterator.java:57)
at
org.apache.phoenix.iterate.ParallelIterators$3.call(ParallelIterators.java:351)
at
org.apache.phoenix.iterate.ParallelIterators$3.call(ParallelIterators.java:346)
at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303)
at java.util.concurrent.FutureTask.run(FutureTask.java:138)
at
java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
at java.lang.Thread.run(Thread.java:662)
Caused by: org.apache.hadoop.hbase.DoNotRetryIOException:
org.apache.hadoop.hbase.DoNotRetryIOException: Could not find hash cache
for joinId: {/�PY�L�. The cache might have expired and have been removed.
at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
at
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:39)
at
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27)
at java.lang.reflect.Constructor.newInstance(Constructor.java:513)
at
org.apache.hadoop.ipc.RemoteException.instantiateException(RemoteException.java:95)
at
org.apache.hadoop.ipc.RemoteException.unwrapRemoteException(RemoteException.java:79)
at
org.apache.hadoop.hbase.client.ServerCallable.translateException(ServerCallable.java:256)
at
org.apache.hadoop.hbase.client.ServerCallable.withRetries(ServerCallable.java:166)
at
org.apache.hadoop.hbase.client.ClientScanner.nextScanner(ClientScanner.java:211)
at
org.apache.hadoop.hbase.client.ClientScanner.initializeScannerInConstruction(ClientScanner.java:126)
at
org.apache.hadoop.hbase.client.ClientScanner.<init>(ClientScanner.java:121)
at org.apache.hadoop.hbase.client.HTable.getScanner(HTable.java:702)
at
org.apache.phoenix.iterate.TableResultIterator.<init>(TableResultIterator.java:54)
... 7 more
Caused by: org.apache.hadoop.ipc.RemoteException:
org.apache.hadoop.hbase.DoNotRetryIOException: Could not find hash cache
for joinId: {/�PY�L�. The cache might have expired and have been removed.
at
org.apache.phoenix.coprocessor.HashJoinRegionScanner.<init>(HashJoinRegionScanner.java:90)
at
org.apache.phoenix.coprocessor.GroupedAggregateRegionObserver.doPostScannerOpen(GroupedAggregateRegionObserver.java:121)
at
org.apache.phoenix.coprocessor.BaseScannerRegionObserver.postScannerOpen(BaseScannerRegionObserver.java:89)
at
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.postScannerOpen(RegionCoprocessorHost.java:1313)
at
org.apache.hadoop.hbase.regionserver.HRegionServer.openScanner(HRegionServer.java:2429)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at
org.apache.hadoop.hbase.ipc.WritableRpcEngine$Server.call(WritableRpcEngine.java:320)
at
org.apache.hadoop.hbase.ipc.HBaseServer$Handler.run(HBaseServer.java:1428)

at org.apache.hadoop.hbase.ipc.HBaseClient.call(HBaseClient.java:1012)
at
org.apache.hadoop.hbase.ipc.WritableRpcEngine$Invoker.invoke(WritableRpcEngine.java:87)
at $Proxy6.openScanner(Unknown Source)
at
org.apache.hadoop.hbase.client.ScannerCallable.openScanner(ScannerCallable.java:224)
at
org.apache.hadoop.hbase.client.ScannerCallable.call(ScannerCallable.java:126)
at
org.apache.hadoop.hbase.client.ScannerCallable.call(ScannerCallable.java:42)
at
org.apache.hadoop.hbase.client.ServerCallable.withRetries(ServerCallable.java:164)

Kindly direct me.

Regards,
Vijay Raajaa G S

On Sun, Oct 5, 2014 at 2:10 AM, Maryann Xue <maryann....@gmail.com> wrote:

> Hi Ashish,
>
> The "phoenix.query.maxServerCacheBytes" is a client parameter while the
> other two are server parameters. But it looks like the configuration change
> did not take effect at your client side. Could you please make sure that
> this is the only configuration that goes to the CLASSPATH of your phoenix
> client execution environment?
>
> Another thing is the exception you got was a different problem from
> Vijay's. It happened in an even earlier stage. Could you please also share
> you query? We could probably re-write it so that it can better fit the
> hash-join scheme. (Since table stats are not used in joins yet, we
> currently have to do it manually.)
>
>
> Thanks,
> Maryann
>
> On Tue, Sep 30, 2014 at 1:22 PM, ashish tapdiya <ashishtapd...@gmail.com>
> wrote:
>
>> Here it is,
>>
>> java.sql.SQLException: Encountered exception in hash plan [1] execution.
>>         at
>> org.apache.phoenix.execute.HashJoinPlan.iterator(HashJoinPlan.java:146)
>>         at
>> org.apache.phoenix.jdbc.PhoenixStatement$1.call(PhoenixStatement.java:211)
>>         at
>> org.apache.phoenix.jdbc.PhoenixStatement$1.call(PhoenixStatement.java:204)
>>         at
>> org.apache.phoenix.util.PhoenixContextExecutor.call(PhoenixContextExecutor.java:54)
>>         at
>> org.apache.phoenix.jdbc.PhoenixStatement.executeQuery(PhoenixStatement.java:204)
>>         at
>> org.apache.phoenix.jdbc.PhoenixPreparedStatement.executeQuery(PhoenixPreparedStatement.java:158)
>>         at Query.sel_Cust_Order_OrderLine_Tables(Query.java:135)
>>         at Query.main(Query.java:25)
>> Caused by: org.apache.phoenix.join.MaxServerCacheSizeExceededException:
>> Size of hash cache (104857684 bytes) exceeds the maximum allowed size
>> (104857600 bytes)
>>         at
>> org.apache.phoenix.join.HashCacheClient.serialize(HashCacheClient.java:106)
>>         at
>> org.apache.phoenix.join.HashCacheClient.addHashCache(HashCacheClient.java:77)
>>         at
>> org.apache.phoenix.execute.HashJoinPlan$1.call(HashJoinPlan.java:119)
>>         at
>> org.apache.phoenix.execute.HashJoinPlan$1.call(HashJoinPlan.java:114)
>>         at java.util.concurrent.FutureTask.run(FutureTask.java:262)
>>         at
>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>>         at
>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>>         at java.lang.Thread.run(Thread.java:744)
>>
>> I am setting hbase heap to 4 GB and phoenix properties are set as below
>>
>> <property>
>>       <name>phoenix.query.maxServerCacheBytes</name>
>>       <value>2004857600</value>
>> </property>
>> <property>
>>       <name>phoenix.query.maxGlobalMemoryPercentage</name>
>>       <value>40</value>
>> </property>
>> <property>
>>       <name>phoenix.query.maxGlobalMemorySize</name>
>>       <value>1504857600</value>
>> </property>
>>
>> Thanks,
>> ~Ashish
>>
>> On Tue, Sep 30, 2014 at 12:13 PM, Maryann Xue <maryann....@gmail.com>
>> wrote:
>>
>>> Hi Ashish,
>>>
>>> Could you please let us see your error message?
>>>
>>>
>>> Thanks,
>>> Maryann
>>>
>>> On Tue, Sep 30, 2014 at 12:58 PM, ashish tapdiya <
>>> ashishtapd...@gmail.com> wrote:
>>>
>>>> Hey Maryann,
>>>>
>>>> Thanks for your input. I tried both the properties but no luck.
>>>>
>>>> ~Ashish
>>>>
>>>> On Sun, Sep 28, 2014 at 8:31 PM, Maryann Xue <maryann....@gmail.com>
>>>> wrote:
>>>>
>>>>> Hi Ashish,
>>>>>
>>>>> The global cache size is set to either "
>>>>> *phoenix.query.maxGlobalMemorySize*" or 
>>>>> "phoenix.query.maxGlobalMemoryPercentage
>>>>> * heapSize" (Sorry about the mistake I made earlier). The ""
>>>>> phoenix.query.maxServerCacheBytes" is a client parameter and is most
>>>>> likely NOT the thing you should worry about. So you can try adjusting "
>>>>> phoenix.query.maxGlobalMemoryPercentage" and the heap size in region
>>>>> server configurations and see how it works.
>>>>>
>>>>>
>>>>> Thanks,
>>>>> Maryann
>>>>>
>>>>> On Fri, Sep 26, 2014 at 10:48 PM, ashish tapdiya <
>>>>> ashishtapd...@gmail.com> wrote:
>>>>>
>>>>>> I have tried that as well...but "phoenix.query.maxServerCacheBytes"
>>>>>> remains the default value of 100 MB. I get to see it when join fails.
>>>>>>
>>>>>> Thanks,
>>>>>> ~Ashish
>>>>>>
>>>>>> On Fri, Sep 26, 2014 at 8:02 PM, Maryann Xue <maryann....@gmail.com>
>>>>>> wrote:
>>>>>>
>>>>>>> Hi Ashish,
>>>>>>>
>>>>>>> The global cache size is set to either 
>>>>>>> "phoenix.query.maxServerCacheBytes"
>>>>>>> or "phoenix.query.maxGlobalMemoryPercentage * heapSize", whichever
>>>>>>> is *smaller*. You can try setting "phoenix.query.
>>>>>>> maxGlobalMemoryPercentage" instead, which is recommended, and see
>>>>>>> how it goes.
>>>>>>>
>>>>>>>
>>>>>>> Thanks,
>>>>>>> Maryann
>>>>>>>
>>>>>>> On Fri, Sep 26, 2014 at 5:37 PM, ashish tapdiya <
>>>>>>> ashishtapd...@gmail.com> wrote:
>>>>>>>
>>>>>>>> Hi Maryann,
>>>>>>>>
>>>>>>>> I am having the same issue where star join is failing with 
>>>>>>>> MaxServerCacheSizeExceededException.
>>>>>>>> I set phoenix.query.maxServerCacheBytes to 1 GB both in client and
>>>>>>>> server hbase-site.xml's. However, it does not take effect.
>>>>>>>>
>>>>>>>> Phoenix 3.1
>>>>>>>> HBase .94
>>>>>>>>
>>>>>>>> Thanks,
>>>>>>>> ~Ashish
>>>>>>>>
>>>>>>>> On Fri, Sep 26, 2014 at 2:56 PM, Maryann Xue <maryann....@gmail.com
>>>>>>>> > wrote:
>>>>>>>>
>>>>>>>>> Yes, you should make your modification on each region server,
>>>>>>>>> since this is a server-side configuration.
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> On Thu, Sep 25, 2014 at 4:15 AM, G.S.Vijay Raajaa <
>>>>>>>>> gsvijayraa...@gmail.com> wrote:
>>>>>>>>>
>>>>>>>>>> Hi Xue,
>>>>>>>>>>
>>>>>>>>>>           Thanks for replying. I did modify the hbase-site.xml by
>>>>>>>>>> increasing the default value of 
>>>>>>>>>> phoenix.query.maxGlobalMemoryPercentage
>>>>>>>>>> . Also increased the Region server heap space memory . The
>>>>>>>>>> change didn't get reflected and I still get the error with an 
>>>>>>>>>> indication
>>>>>>>>>> that "global pool of 319507660 bytes" is present. Should I
>>>>>>>>>> modify the hbase-site.xml in every region server or just the file 
>>>>>>>>>> present
>>>>>>>>>> in the class path of Phoenix client?
>>>>>>>>>>
>>>>>>>>>> Regards,
>>>>>>>>>> Vijay Raajaa G S
>>>>>>>>>>
>>>>>>>>>> On Thu, Sep 25, 2014 at 1:47 AM, Maryann Xue <
>>>>>>>>>> maryann....@gmail.com> wrote:
>>>>>>>>>>
>>>>>>>>>>> Hi Vijay,
>>>>>>>>>>>
>>>>>>>>>>> I think here the query plan is scanning table *CUSTOMER_30000 *while
>>>>>>>>>>> joining the other two tables at the same time, which means the 
>>>>>>>>>>> region
>>>>>>>>>>> server memory for Phoenix should be large enough to hold 2 tables 
>>>>>>>>>>> together
>>>>>>>>>>> and you also need to expect some memory expansion for java objects.
>>>>>>>>>>>
>>>>>>>>>>> Do you mean that after you had modified the parameters you
>>>>>>>>>>> mentioned, you were still getting the same error message with 
>>>>>>>>>>> exactly the
>>>>>>>>>>> same numbers as "global pool of 319507660 bytes"? Did you make
>>>>>>>>>>> sure that the parameters actually took effect after modification?
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> Thanks,
>>>>>>>>>>> Maryann
>>>>>>>>>>>
>>>>>>>>>>> On Tue, Sep 23, 2014 at 1:43 AM, G.S.Vijay Raajaa <
>>>>>>>>>>> gsvijayraa...@gmail.com> wrote:
>>>>>>>>>>>
>>>>>>>>>>>> Hi,
>>>>>>>>>>>>
>>>>>>>>>>>>     I am trying to do a join of three tables usng the following
>>>>>>>>>>>> query:
>>>>>>>>>>>>
>>>>>>>>>>>> *select c.c_first_name, ca.ca_city, cd.cd_education_status from
>>>>>>>>>>>> CUSTOMER_30000 c join CUSTOMER_DEMOGRAPHICS_1 cd on 
>>>>>>>>>>>> c.c_current_cdemo_sk =
>>>>>>>>>>>> cd.cd_demo_sk join CUSTOMER_ADDRESS_1 ca on c.c_current_addr_sk =
>>>>>>>>>>>> ca.ca_address_sk group by ca.ca_city, cd.cd_education_status,
>>>>>>>>>>>> c.c_first_name;*
>>>>>>>>>>>>
>>>>>>>>>>>> *The size of CUSTOMER_30000 is 4.1 GB with 30million records.*
>>>>>>>>>>>>
>>>>>>>>>>>> *I get the following error:*
>>>>>>>>>>>>
>>>>>>>>>>>> ./psql.py 10.10.5.55 test.sql
>>>>>>>>>>>> java.sql.SQLException: Encountered exception in hash plan [0]
>>>>>>>>>>>> execution.
>>>>>>>>>>>> at
>>>>>>>>>>>> org.apache.phoenix.execute.HashJoinPlan.iterator(HashJoinPlan.java:146)
>>>>>>>>>>>> at
>>>>>>>>>>>> org.apache.phoenix.jdbc.PhoenixStatement$1.call(PhoenixStatement.java:211)
>>>>>>>>>>>> at
>>>>>>>>>>>> org.apache.phoenix.jdbc.PhoenixStatement$1.call(PhoenixStatement.java:204)
>>>>>>>>>>>> at
>>>>>>>>>>>> org.apache.phoenix.util.PhoenixContextExecutor.call(PhoenixContextExecutor.java:54)
>>>>>>>>>>>> at
>>>>>>>>>>>> org.apache.phoenix.jdbc.PhoenixStatement.executeQuery(PhoenixStatement.java:204)
>>>>>>>>>>>> at
>>>>>>>>>>>> org.apache.phoenix.jdbc.PhoenixStatement.execute(PhoenixStatement.java:193)
>>>>>>>>>>>> at
>>>>>>>>>>>> org.apache.phoenix.jdbc.PhoenixPreparedStatement.execute(PhoenixPreparedStatement.java:147)
>>>>>>>>>>>> at
>>>>>>>>>>>> org.apache.phoenix.jdbc.PhoenixPreparedStatement.execute(PhoenixPreparedStatement.java:152)
>>>>>>>>>>>> at
>>>>>>>>>>>> org.apache.phoenix.jdbc.PhoenixConnection.executeStatements(PhoenixConnection.java:220)
>>>>>>>>>>>> at
>>>>>>>>>>>> org.apache.phoenix.util.PhoenixRuntime.executeStatements(PhoenixRuntime.java:193)
>>>>>>>>>>>> at
>>>>>>>>>>>> org.apache.phoenix.util.PhoenixRuntime.main(PhoenixRuntime.java:140)
>>>>>>>>>>>> Caused by: java.sql.SQLException:
>>>>>>>>>>>> java.util.concurrent.ExecutionException:
>>>>>>>>>>>> java.lang.reflect.UndeclaredThrowableException
>>>>>>>>>>>> at
>>>>>>>>>>>> org.apache.phoenix.cache.ServerCacheClient.addServerCache(ServerCacheClient.java:199)
>>>>>>>>>>>> at
>>>>>>>>>>>> org.apache.phoenix.join.HashCacheClient.addHashCache(HashCacheClient.java:78)
>>>>>>>>>>>> at
>>>>>>>>>>>> org.apache.phoenix.execute.HashJoinPlan$1.call(HashJoinPlan.java:119)
>>>>>>>>>>>> at
>>>>>>>>>>>> org.apache.phoenix.execute.HashJoinPlan$1.call(HashJoinPlan.java:114)
>>>>>>>>>>>> at
>>>>>>>>>>>> java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303)
>>>>>>>>>>>> at java.util.concurrent.FutureTask.run(FutureTask.java:138)
>>>>>>>>>>>> at
>>>>>>>>>>>> java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
>>>>>>>>>>>> at
>>>>>>>>>>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
>>>>>>>>>>>> at java.lang.Thread.run(Thread.java:662)
>>>>>>>>>>>> Caused by: java.util.concurrent.ExecutionException:
>>>>>>>>>>>> java.lang.reflect.UndeclaredThrowableException
>>>>>>>>>>>> at
>>>>>>>>>>>> java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232)
>>>>>>>>>>>> at java.util.concurrent.FutureTask.get(FutureTask.java:91)
>>>>>>>>>>>> at
>>>>>>>>>>>> org.apache.phoenix.cache.ServerCacheClient.addServerCache(ServerCacheClient.java:191)
>>>>>>>>>>>> ... 8 more
>>>>>>>>>>>> Caused by: java.lang.reflect.UndeclaredThrowableException
>>>>>>>>>>>> at $Proxy10.addServerCache(Unknown Source)
>>>>>>>>>>>> at
>>>>>>>>>>>> org.apache.phoenix.cache.ServerCacheClient$1.call(ServerCacheClient.java:169)
>>>>>>>>>>>> at
>>>>>>>>>>>> org.apache.phoenix.cache.ServerCacheClient$1.call(ServerCacheClient.java:164)
>>>>>>>>>>>> ... 5 more
>>>>>>>>>>>> Caused by:
>>>>>>>>>>>> org.apache.hadoop.hbase.client.RetriesExhaustedException: Failed 
>>>>>>>>>>>> after
>>>>>>>>>>>> attempts=14, exceptions:
>>>>>>>>>>>> Tue Sep 23 00:25:53 CDT 2014,
>>>>>>>>>>>> org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@100e398,
>>>>>>>>>>>> java.io.IOException: java.io.IOException:
>>>>>>>>>>>> org.apache.phoenix.memory.InsufficientMemoryException: Requested 
>>>>>>>>>>>> memory of
>>>>>>>>>>>> 446623727 bytes is larger than global pool of 319507660 bytes.
>>>>>>>>>>>> Tue Sep 23 00:26:02 CDT 2014,
>>>>>>>>>>>> org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@100e398,
>>>>>>>>>>>> java.io.IOException: java.io.IOException:
>>>>>>>>>>>> org.apache.phoenix.memory.InsufficientMemoryException: Requested 
>>>>>>>>>>>> memory of
>>>>>>>>>>>> 446623727 bytes is larger than global pool of 319507660 bytes.
>>>>>>>>>>>> Tue Sep 23 00:26:18 CDT 2014,
>>>>>>>>>>>> org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@100e398,
>>>>>>>>>>>> java.io.IOException: java.io.IOException:
>>>>>>>>>>>> org.apache.phoenix.memory.InsufficientMemoryException: Requested 
>>>>>>>>>>>> memory of
>>>>>>>>>>>> 446623727 bytes is larger than global pool of 319507660 bytes.
>>>>>>>>>>>> Tue Sep 23 00:26:43 CDT 2014,
>>>>>>>>>>>> org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@100e398,
>>>>>>>>>>>> java.io.IOException: java.io.IOException:
>>>>>>>>>>>> org.apache.phoenix.memory.InsufficientMemoryException: Requested 
>>>>>>>>>>>> memory of
>>>>>>>>>>>> 446623727 bytes is larger than global pool of 319507660 bytes.
>>>>>>>>>>>> Tue Sep 23 00:27:01 CDT 2014,
>>>>>>>>>>>> org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@100e398,
>>>>>>>>>>>> java.io.IOException: java.io.IOException:
>>>>>>>>>>>> org.apache.phoenix.memory.InsufficientMemoryException: Requested 
>>>>>>>>>>>> memory of
>>>>>>>>>>>> 446623727 bytes is larger than global pool of 319507660 bytes.
>>>>>>>>>>>> Tue Sep 23 00:27:10 CDT 2014,
>>>>>>>>>>>> org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@100e398,
>>>>>>>>>>>> java.io.IOException: java.io.IOException:
>>>>>>>>>>>> org.apache.phoenix.memory.InsufficientMemoryException: Requested 
>>>>>>>>>>>> memory of
>>>>>>>>>>>> 446623727 bytes is larger than global pool of 319507660 bytes.
>>>>>>>>>>>> Tue Sep 23 00:27:24 CDT 2014,
>>>>>>>>>>>> org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@100e398,
>>>>>>>>>>>> java.io.IOException: java.io.IOException:
>>>>>>>>>>>> org.apache.phoenix.memory.InsufficientMemoryException: Requested 
>>>>>>>>>>>> memory of
>>>>>>>>>>>> 446623727 bytes is larger than global pool of 319507660 bytes.
>>>>>>>>>>>> Tue Sep 23 00:28:16 CDT 2014,
>>>>>>>>>>>> org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@100e398,
>>>>>>>>>>>> java.io.IOException: java.io.IOException:
>>>>>>>>>>>> org.apache.phoenix.memory.InsufficientMemoryException: Requested 
>>>>>>>>>>>> memory of
>>>>>>>>>>>> 446623727 bytes is larger than global pool of 319507660 bytes.
>>>>>>>>>>>> Tue Sep 23 00:28:35 CDT 2014,
>>>>>>>>>>>> org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@100e398,
>>>>>>>>>>>> java.io.IOException: java.io.IOException:
>>>>>>>>>>>> org.apache.phoenix.memory.InsufficientMemoryException: Requested 
>>>>>>>>>>>> memory of
>>>>>>>>>>>> 446623727 bytes is larger than global pool of 319507660 bytes.
>>>>>>>>>>>> Tue Sep 23 00:29:09 CDT 2014,
>>>>>>>>>>>> org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@100e398,
>>>>>>>>>>>> java.io.IOException: java.io.IOException:
>>>>>>>>>>>> org.apache.phoenix.memory.InsufficientMemoryException: Requested 
>>>>>>>>>>>> memory of
>>>>>>>>>>>> 446623727 bytes is larger than global pool of 319507660 bytes.
>>>>>>>>>>>> Tue Sep 23 00:30:16 CDT 2014,
>>>>>>>>>>>> org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@100e398,
>>>>>>>>>>>> java.io.IOException: java.io.IOException:
>>>>>>>>>>>> org.apache.phoenix.memory.InsufficientMemoryException: Requested 
>>>>>>>>>>>> memory of
>>>>>>>>>>>> 446623727 bytes is larger than global pool of 319507660 bytes.
>>>>>>>>>>>> Tue Sep 23 00:31:22 CDT 2014,
>>>>>>>>>>>> org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@100e398,
>>>>>>>>>>>> java.io.IOException: java.io.IOException:
>>>>>>>>>>>> org.apache.phoenix.memory.InsufficientMemoryException: Requested 
>>>>>>>>>>>> memory of
>>>>>>>>>>>> 446623727 bytes is larger than global pool of 319507660 bytes.
>>>>>>>>>>>> Tue Sep 23 00:32:29 CDT 2014,
>>>>>>>>>>>> org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@100e398,
>>>>>>>>>>>> java.io.IOException: java.io.IOException:
>>>>>>>>>>>> org.apache.phoenix.memory.InsufficientMemoryException: Requested 
>>>>>>>>>>>> memory of
>>>>>>>>>>>> 446623727 bytes is larger than global pool of 319507660 bytes.
>>>>>>>>>>>> Tue Sep 23 00:33:35 CDT 2014,
>>>>>>>>>>>> org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@100e398,
>>>>>>>>>>>> java.io.IOException: java.io.IOException:
>>>>>>>>>>>> org.apache.phoenix.memory.InsufficientMemoryException: Requested 
>>>>>>>>>>>> memory of
>>>>>>>>>>>> 446623727 bytes is larger than global pool of 319507660 bytes.
>>>>>>>>>>>>
>>>>>>>>>>>> at
>>>>>>>>>>>> org.apache.hadoop.hbase.client.ServerCallable.withRetries(ServerCallable.java:187)
>>>>>>>>>>>> at
>>>>>>>>>>>> org.apache.hadoop.hbase.ipc.ExecRPCInvoker.invoke(ExecRPCInvoker.java:79)
>>>>>>>>>>>> ... 8 more
>>>>>>>>>>>>
>>>>>>>>>>>> Trials:
>>>>>>>>>>>>
>>>>>>>>>>>> I tried to increase the Region Server Heap space ,
>>>>>>>>>>>> modified phoenix.query.maxGlobalMemoryPercentage as well.
>>>>>>>>>>>>
>>>>>>>>>>>> I am not able to increase the global memory .
>>>>>>>>>>>>
>>>>>>>>>>>> Regards,
>>>>>>>>>>>> Vijay Raajaa
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> --
>>>>>>>>>>> Thanks,
>>>>>>>>>>> Maryann
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> --
>>>>>>>>> Thanks,
>>>>>>>>> Maryann
>>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> --
>>>>>>> Thanks,
>>>>>>> Maryann
>>>>>>>
>>>>>>
>>>>>>
>>>>>
>>>>>
>>>>> --
>>>>> Thanks,
>>>>> Maryann
>>>>>
>>>>
>>>>
>>>
>>>
>>> --
>>> Thanks,
>>> Maryann
>>>
>>
>>
>
>
> --
> Thanks,
> Maryann
>

Reply via email to