You are welcome, Vijay!
On Thu, Oct 9, 2014 at 12:48 AM, G.S.Vijay Raajaa <[email protected]> wrote: > Modifying phoenix.coprocessor.maxServerCacheTimeToLiveMs parameter which > defaults to *30,000 *solved the problem. > > *Thanks !!* > > On Wed, Oct 8, 2014 at 10:25 AM, G.S.Vijay Raajaa <[email protected] > > wrote: > >> Hi Maryann, >> >> Its the same query: >> >> *select c.c_first_name, ca.ca_city, cd.cd_education_status from >> CUSTOMER_30000 c join CUSTOMER_DEMOGRAPHICS_1 cd on c.c_current_cdemo_sk = >> cd.cd_demo_sk join CUSTOMER_ADDRESS_1 ca on c.c_current_addr_sk = >> ca.ca_address_sk group by ca.ca_city, cd.cd_education_status, >> c.c_first_name;* >> >> *The size of CUSTOMER_30000 is 4.1 GB with 30million records. The >> CUSTOMER_DEMOGRAPHICS contains 2M records and CUSTOMER_ADDRESS contains >> 50000 records.* >> >> *Regards,* >> *Vijay Raajaa G S* >> >> On Tue, Oct 7, 2014 at 9:39 PM, Maryann Xue <[email protected]> >> wrote: >> >>> Hi Ashish, >>> >>> The warning you got was exactly showing the reason why you finally got >>> that error: one of the join table query had taken too long so that the >>> cache for other join tables expired and got invalidated. Again, could you >>> please share your query and the size of the tables used in your query? >>> Instead of changing the parameters to get around the problems, it might be >>> much more efficient just to adjust the query itself. And if doable, mostly >>> likely the query is gonna run faster as well. >>> >>> Besides, you might find this document helpful : >>> http://phoenix.apache.org/joins.html >>> >>> >>> Thanks, >>> Maryann >>> >>> >>> On Tue, Oct 7, 2014 at 11:01 AM, ashish tapdiya <[email protected] >>> > wrote: >>> >>>> Maryann, >>>> >>>> hbase-site.xml was not on CLASSPATH and that was the issue. Thanks for >>>> the help. I appreciate it. >>>> >>>> ~Ashish >>>> >>>> >>>> >>>> On Sat, Oct 4, 2014 at 3:40 PM, Maryann Xue <[email protected]> >>>> wrote: >>>> >>>>> Hi Ashish, >>>>> >>>>> The "phoenix.query.maxServerCacheBytes" is a client parameter while >>>>> the other two are server parameters. But it looks like the configuration >>>>> change did not take effect at your client side. Could you please make sure >>>>> that this is the only configuration that goes to the CLASSPATH of your >>>>> phoenix client execution environment? >>>>> >>>>> Another thing is the exception you got was a different problem from >>>>> Vijay's. It happened in an even earlier stage. Could you please also share >>>>> you query? We could probably re-write it so that it can better fit the >>>>> hash-join scheme. (Since table stats are not used in joins yet, we >>>>> currently have to do it manually.) >>>>> >>>>> >>>>> Thanks, >>>>> Maryann >>>>> >>>>> On Tue, Sep 30, 2014 at 1:22 PM, ashish tapdiya < >>>>> [email protected]> wrote: >>>>> >>>>>> Here it is, >>>>>> >>>>>> java.sql.SQLException: Encountered exception in hash plan [1] >>>>>> execution. >>>>>> at >>>>>> org.apache.phoenix.execute.HashJoinPlan.iterator(HashJoinPlan.java:146) >>>>>> at >>>>>> org.apache.phoenix.jdbc.PhoenixStatement$1.call(PhoenixStatement.java:211) >>>>>> at >>>>>> org.apache.phoenix.jdbc.PhoenixStatement$1.call(PhoenixStatement.java:204) >>>>>> at >>>>>> org.apache.phoenix.util.PhoenixContextExecutor.call(PhoenixContextExecutor.java:54) >>>>>> at >>>>>> org.apache.phoenix.jdbc.PhoenixStatement.executeQuery(PhoenixStatement.java:204) >>>>>> at >>>>>> org.apache.phoenix.jdbc.PhoenixPreparedStatement.executeQuery(PhoenixPreparedStatement.java:158) >>>>>> at Query.sel_Cust_Order_OrderLine_Tables(Query.java:135) >>>>>> at Query.main(Query.java:25) >>>>>> Caused by: >>>>>> org.apache.phoenix.join.MaxServerCacheSizeExceededException: Size of hash >>>>>> cache (104857684 bytes) exceeds the maximum allowed size (104857600 >>>>>> bytes) >>>>>> at >>>>>> org.apache.phoenix.join.HashCacheClient.serialize(HashCacheClient.java:106) >>>>>> at >>>>>> org.apache.phoenix.join.HashCacheClient.addHashCache(HashCacheClient.java:77) >>>>>> at >>>>>> org.apache.phoenix.execute.HashJoinPlan$1.call(HashJoinPlan.java:119) >>>>>> at >>>>>> org.apache.phoenix.execute.HashJoinPlan$1.call(HashJoinPlan.java:114) >>>>>> at java.util.concurrent.FutureTask.run(FutureTask.java:262) >>>>>> at >>>>>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) >>>>>> at >>>>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) >>>>>> at java.lang.Thread.run(Thread.java:744) >>>>>> >>>>>> I am setting hbase heap to 4 GB and phoenix properties are set as >>>>>> below >>>>>> >>>>>> <property> >>>>>> <name>phoenix.query.maxServerCacheBytes</name> >>>>>> <value>2004857600</value> >>>>>> </property> >>>>>> <property> >>>>>> <name>phoenix.query.maxGlobalMemoryPercentage</name> >>>>>> <value>40</value> >>>>>> </property> >>>>>> <property> >>>>>> <name>phoenix.query.maxGlobalMemorySize</name> >>>>>> <value>1504857600</value> >>>>>> </property> >>>>>> >>>>>> Thanks, >>>>>> ~Ashish >>>>>> >>>>>> On Tue, Sep 30, 2014 at 12:13 PM, Maryann Xue <[email protected]> >>>>>> wrote: >>>>>> >>>>>>> Hi Ashish, >>>>>>> >>>>>>> Could you please let us see your error message? >>>>>>> >>>>>>> >>>>>>> Thanks, >>>>>>> Maryann >>>>>>> >>>>>>> On Tue, Sep 30, 2014 at 12:58 PM, ashish tapdiya < >>>>>>> [email protected]> wrote: >>>>>>> >>>>>>>> Hey Maryann, >>>>>>>> >>>>>>>> Thanks for your input. I tried both the properties but no luck. >>>>>>>> >>>>>>>> ~Ashish >>>>>>>> >>>>>>>> On Sun, Sep 28, 2014 at 8:31 PM, Maryann Xue <[email protected] >>>>>>>> > wrote: >>>>>>>> >>>>>>>>> Hi Ashish, >>>>>>>>> >>>>>>>>> The global cache size is set to either " >>>>>>>>> *phoenix.query.maxGlobalMemorySize*" or >>>>>>>>> "phoenix.query.maxGlobalMemoryPercentage >>>>>>>>> * heapSize" (Sorry about the mistake I made earlier). The "" >>>>>>>>> phoenix.query.maxServerCacheBytes" is a client parameter and is >>>>>>>>> most likely NOT the thing you should worry about. So you can try >>>>>>>>> adjusting " >>>>>>>>> phoenix.query.maxGlobalMemoryPercentage" and the heap size in >>>>>>>>> region server configurations and see how it works. >>>>>>>>> >>>>>>>>> >>>>>>>>> Thanks, >>>>>>>>> Maryann >>>>>>>>> >>>>>>>>> On Fri, Sep 26, 2014 at 10:48 PM, ashish tapdiya < >>>>>>>>> [email protected]> wrote: >>>>>>>>> >>>>>>>>>> I have tried that as well...but "phoenix.query.maxServerCacheBytes" >>>>>>>>>> remains the default value of 100 MB. I get to see it when join fails. >>>>>>>>>> >>>>>>>>>> Thanks, >>>>>>>>>> ~Ashish >>>>>>>>>> >>>>>>>>>> On Fri, Sep 26, 2014 at 8:02 PM, Maryann Xue < >>>>>>>>>> [email protected]> wrote: >>>>>>>>>> >>>>>>>>>>> Hi Ashish, >>>>>>>>>>> >>>>>>>>>>> The global cache size is set to either >>>>>>>>>>> "phoenix.query.maxServerCacheBytes" >>>>>>>>>>> or "phoenix.query.maxGlobalMemoryPercentage * heapSize", >>>>>>>>>>> whichever is *smaller*. You can try setting "phoenix.query. >>>>>>>>>>> maxGlobalMemoryPercentage" instead, which is recommended, and >>>>>>>>>>> see how it goes. >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> Thanks, >>>>>>>>>>> Maryann >>>>>>>>>>> >>>>>>>>>>> On Fri, Sep 26, 2014 at 5:37 PM, ashish tapdiya < >>>>>>>>>>> [email protected]> wrote: >>>>>>>>>>> >>>>>>>>>>>> Hi Maryann, >>>>>>>>>>>> >>>>>>>>>>>> I am having the same issue where star join is failing with >>>>>>>>>>>> MaxServerCacheSizeExceededException. >>>>>>>>>>>> I set phoenix.query.maxServerCacheBytes to 1 GB both in client >>>>>>>>>>>> and server hbase-site.xml's. However, it does not take effect. >>>>>>>>>>>> >>>>>>>>>>>> Phoenix 3.1 >>>>>>>>>>>> HBase .94 >>>>>>>>>>>> >>>>>>>>>>>> Thanks, >>>>>>>>>>>> ~Ashish >>>>>>>>>>>> >>>>>>>>>>>> On Fri, Sep 26, 2014 at 2:56 PM, Maryann Xue < >>>>>>>>>>>> [email protected]> wrote: >>>>>>>>>>>> >>>>>>>>>>>>> Yes, you should make your modification on each region server, >>>>>>>>>>>>> since this is a server-side configuration. >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> On Thu, Sep 25, 2014 at 4:15 AM, G.S.Vijay Raajaa < >>>>>>>>>>>>> [email protected]> wrote: >>>>>>>>>>>>> >>>>>>>>>>>>>> Hi Xue, >>>>>>>>>>>>>> >>>>>>>>>>>>>> Thanks for replying. I did modify the >>>>>>>>>>>>>> hbase-site.xml by increasing the default value of >>>>>>>>>>>>>> phoenix.query.maxGlobalMemoryPercentage . Also increased the >>>>>>>>>>>>>> Region server heap space memory . The change didn't get >>>>>>>>>>>>>> reflected and I >>>>>>>>>>>>>> still get the error with an indication that "global pool of >>>>>>>>>>>>>> 319507660 bytes" is present. Should I modify the hbase-site.xml >>>>>>>>>>>>>> in every >>>>>>>>>>>>>> region server or just the file present in the class path of >>>>>>>>>>>>>> Phoenix client? >>>>>>>>>>>>>> >>>>>>>>>>>>>> Regards, >>>>>>>>>>>>>> Vijay Raajaa G S >>>>>>>>>>>>>> >>>>>>>>>>>>>> On Thu, Sep 25, 2014 at 1:47 AM, Maryann Xue < >>>>>>>>>>>>>> [email protected]> wrote: >>>>>>>>>>>>>> >>>>>>>>>>>>>>> Hi Vijay, >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> I think here the query plan is scanning table *CUSTOMER_30000 >>>>>>>>>>>>>>> *while joining the other two tables at the same time, which >>>>>>>>>>>>>>> means the region server memory for Phoenix should be large >>>>>>>>>>>>>>> enough to hold 2 >>>>>>>>>>>>>>> tables together and you also need to expect some memory >>>>>>>>>>>>>>> expansion for java >>>>>>>>>>>>>>> objects. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Do you mean that after you had modified the parameters you >>>>>>>>>>>>>>> mentioned, you were still getting the same error message with >>>>>>>>>>>>>>> exactly the >>>>>>>>>>>>>>> same numbers as "global pool of 319507660 bytes"? Did you >>>>>>>>>>>>>>> make sure that the parameters actually took effect after >>>>>>>>>>>>>>> modification? >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Thanks, >>>>>>>>>>>>>>> Maryann >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> On Tue, Sep 23, 2014 at 1:43 AM, G.S.Vijay Raajaa < >>>>>>>>>>>>>>> [email protected]> wrote: >>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Hi, >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> I am trying to do a join of three tables usng the >>>>>>>>>>>>>>>> following query: >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> *select c.c_first_name, ca.ca_city, cd.cd_education_status >>>>>>>>>>>>>>>> from CUSTOMER_30000 c join CUSTOMER_DEMOGRAPHICS_1 cd on >>>>>>>>>>>>>>>> c.c_current_cdemo_sk = cd.cd_demo_sk join CUSTOMER_ADDRESS_1 >>>>>>>>>>>>>>>> ca on >>>>>>>>>>>>>>>> c.c_current_addr_sk = ca.ca_address_sk group by ca.ca_city, >>>>>>>>>>>>>>>> cd.cd_education_status, c.c_first_name;* >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> *The size of CUSTOMER_30000 is 4.1 GB with 30million >>>>>>>>>>>>>>>> records.* >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> *I get the following error:* >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> ./psql.py 10.10.5.55 test.sql >>>>>>>>>>>>>>>> java.sql.SQLException: Encountered exception in hash plan >>>>>>>>>>>>>>>> [0] execution. >>>>>>>>>>>>>>>> at >>>>>>>>>>>>>>>> org.apache.phoenix.execute.HashJoinPlan.iterator(HashJoinPlan.java:146) >>>>>>>>>>>>>>>> at >>>>>>>>>>>>>>>> org.apache.phoenix.jdbc.PhoenixStatement$1.call(PhoenixStatement.java:211) >>>>>>>>>>>>>>>> at >>>>>>>>>>>>>>>> org.apache.phoenix.jdbc.PhoenixStatement$1.call(PhoenixStatement.java:204) >>>>>>>>>>>>>>>> at >>>>>>>>>>>>>>>> org.apache.phoenix.util.PhoenixContextExecutor.call(PhoenixContextExecutor.java:54) >>>>>>>>>>>>>>>> at >>>>>>>>>>>>>>>> org.apache.phoenix.jdbc.PhoenixStatement.executeQuery(PhoenixStatement.java:204) >>>>>>>>>>>>>>>> at >>>>>>>>>>>>>>>> org.apache.phoenix.jdbc.PhoenixStatement.execute(PhoenixStatement.java:193) >>>>>>>>>>>>>>>> at >>>>>>>>>>>>>>>> org.apache.phoenix.jdbc.PhoenixPreparedStatement.execute(PhoenixPreparedStatement.java:147) >>>>>>>>>>>>>>>> at >>>>>>>>>>>>>>>> org.apache.phoenix.jdbc.PhoenixPreparedStatement.execute(PhoenixPreparedStatement.java:152) >>>>>>>>>>>>>>>> at >>>>>>>>>>>>>>>> org.apache.phoenix.jdbc.PhoenixConnection.executeStatements(PhoenixConnection.java:220) >>>>>>>>>>>>>>>> at >>>>>>>>>>>>>>>> org.apache.phoenix.util.PhoenixRuntime.executeStatements(PhoenixRuntime.java:193) >>>>>>>>>>>>>>>> at >>>>>>>>>>>>>>>> org.apache.phoenix.util.PhoenixRuntime.main(PhoenixRuntime.java:140) >>>>>>>>>>>>>>>> Caused by: java.sql.SQLException: >>>>>>>>>>>>>>>> java.util.concurrent.ExecutionException: >>>>>>>>>>>>>>>> java.lang.reflect.UndeclaredThrowableException >>>>>>>>>>>>>>>> at >>>>>>>>>>>>>>>> org.apache.phoenix.cache.ServerCacheClient.addServerCache(ServerCacheClient.java:199) >>>>>>>>>>>>>>>> at >>>>>>>>>>>>>>>> org.apache.phoenix.join.HashCacheClient.addHashCache(HashCacheClient.java:78) >>>>>>>>>>>>>>>> at >>>>>>>>>>>>>>>> org.apache.phoenix.execute.HashJoinPlan$1.call(HashJoinPlan.java:119) >>>>>>>>>>>>>>>> at >>>>>>>>>>>>>>>> org.apache.phoenix.execute.HashJoinPlan$1.call(HashJoinPlan.java:114) >>>>>>>>>>>>>>>> at >>>>>>>>>>>>>>>> java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) >>>>>>>>>>>>>>>> at java.util.concurrent.FutureTask.run(FutureTask.java:138) >>>>>>>>>>>>>>>> at >>>>>>>>>>>>>>>> java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) >>>>>>>>>>>>>>>> at >>>>>>>>>>>>>>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) >>>>>>>>>>>>>>>> at java.lang.Thread.run(Thread.java:662) >>>>>>>>>>>>>>>> Caused by: java.util.concurrent.ExecutionException: >>>>>>>>>>>>>>>> java.lang.reflect.UndeclaredThrowableException >>>>>>>>>>>>>>>> at >>>>>>>>>>>>>>>> java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:232) >>>>>>>>>>>>>>>> at java.util.concurrent.FutureTask.get(FutureTask.java:91) >>>>>>>>>>>>>>>> at >>>>>>>>>>>>>>>> org.apache.phoenix.cache.ServerCacheClient.addServerCache(ServerCacheClient.java:191) >>>>>>>>>>>>>>>> ... 8 more >>>>>>>>>>>>>>>> Caused by: java.lang.reflect.UndeclaredThrowableException >>>>>>>>>>>>>>>> at $Proxy10.addServerCache(Unknown Source) >>>>>>>>>>>>>>>> at >>>>>>>>>>>>>>>> org.apache.phoenix.cache.ServerCacheClient$1.call(ServerCacheClient.java:169) >>>>>>>>>>>>>>>> at >>>>>>>>>>>>>>>> org.apache.phoenix.cache.ServerCacheClient$1.call(ServerCacheClient.java:164) >>>>>>>>>>>>>>>> ... 5 more >>>>>>>>>>>>>>>> Caused by: >>>>>>>>>>>>>>>> org.apache.hadoop.hbase.client.RetriesExhaustedException: >>>>>>>>>>>>>>>> Failed after >>>>>>>>>>>>>>>> attempts=14, exceptions: >>>>>>>>>>>>>>>> Tue Sep 23 00:25:53 CDT 2014, >>>>>>>>>>>>>>>> org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@100e398, >>>>>>>>>>>>>>>> java.io.IOException: java.io.IOException: >>>>>>>>>>>>>>>> org.apache.phoenix.memory.InsufficientMemoryException: >>>>>>>>>>>>>>>> Requested memory of >>>>>>>>>>>>>>>> 446623727 bytes is larger than global pool of 319507660 bytes. >>>>>>>>>>>>>>>> Tue Sep 23 00:26:02 CDT 2014, >>>>>>>>>>>>>>>> org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@100e398, >>>>>>>>>>>>>>>> java.io.IOException: java.io.IOException: >>>>>>>>>>>>>>>> org.apache.phoenix.memory.InsufficientMemoryException: >>>>>>>>>>>>>>>> Requested memory of >>>>>>>>>>>>>>>> 446623727 bytes is larger than global pool of 319507660 bytes. >>>>>>>>>>>>>>>> Tue Sep 23 00:26:18 CDT 2014, >>>>>>>>>>>>>>>> org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@100e398, >>>>>>>>>>>>>>>> java.io.IOException: java.io.IOException: >>>>>>>>>>>>>>>> org.apache.phoenix.memory.InsufficientMemoryException: >>>>>>>>>>>>>>>> Requested memory of >>>>>>>>>>>>>>>> 446623727 bytes is larger than global pool of 319507660 bytes. >>>>>>>>>>>>>>>> Tue Sep 23 00:26:43 CDT 2014, >>>>>>>>>>>>>>>> org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@100e398, >>>>>>>>>>>>>>>> java.io.IOException: java.io.IOException: >>>>>>>>>>>>>>>> org.apache.phoenix.memory.InsufficientMemoryException: >>>>>>>>>>>>>>>> Requested memory of >>>>>>>>>>>>>>>> 446623727 bytes is larger than global pool of 319507660 bytes. >>>>>>>>>>>>>>>> Tue Sep 23 00:27:01 CDT 2014, >>>>>>>>>>>>>>>> org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@100e398, >>>>>>>>>>>>>>>> java.io.IOException: java.io.IOException: >>>>>>>>>>>>>>>> org.apache.phoenix.memory.InsufficientMemoryException: >>>>>>>>>>>>>>>> Requested memory of >>>>>>>>>>>>>>>> 446623727 bytes is larger than global pool of 319507660 bytes. >>>>>>>>>>>>>>>> Tue Sep 23 00:27:10 CDT 2014, >>>>>>>>>>>>>>>> org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@100e398, >>>>>>>>>>>>>>>> java.io.IOException: java.io.IOException: >>>>>>>>>>>>>>>> org.apache.phoenix.memory.InsufficientMemoryException: >>>>>>>>>>>>>>>> Requested memory of >>>>>>>>>>>>>>>> 446623727 bytes is larger than global pool of 319507660 bytes. >>>>>>>>>>>>>>>> Tue Sep 23 00:27:24 CDT 2014, >>>>>>>>>>>>>>>> org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@100e398, >>>>>>>>>>>>>>>> java.io.IOException: java.io.IOException: >>>>>>>>>>>>>>>> org.apache.phoenix.memory.InsufficientMemoryException: >>>>>>>>>>>>>>>> Requested memory of >>>>>>>>>>>>>>>> 446623727 bytes is larger than global pool of 319507660 bytes. >>>>>>>>>>>>>>>> Tue Sep 23 00:28:16 CDT 2014, >>>>>>>>>>>>>>>> org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@100e398, >>>>>>>>>>>>>>>> java.io.IOException: java.io.IOException: >>>>>>>>>>>>>>>> org.apache.phoenix.memory.InsufficientMemoryException: >>>>>>>>>>>>>>>> Requested memory of >>>>>>>>>>>>>>>> 446623727 bytes is larger than global pool of 319507660 bytes. >>>>>>>>>>>>>>>> Tue Sep 23 00:28:35 CDT 2014, >>>>>>>>>>>>>>>> org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@100e398, >>>>>>>>>>>>>>>> java.io.IOException: java.io.IOException: >>>>>>>>>>>>>>>> org.apache.phoenix.memory.InsufficientMemoryException: >>>>>>>>>>>>>>>> Requested memory of >>>>>>>>>>>>>>>> 446623727 bytes is larger than global pool of 319507660 bytes. >>>>>>>>>>>>>>>> Tue Sep 23 00:29:09 CDT 2014, >>>>>>>>>>>>>>>> org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@100e398, >>>>>>>>>>>>>>>> java.io.IOException: java.io.IOException: >>>>>>>>>>>>>>>> org.apache.phoenix.memory.InsufficientMemoryException: >>>>>>>>>>>>>>>> Requested memory of >>>>>>>>>>>>>>>> 446623727 bytes is larger than global pool of 319507660 bytes. >>>>>>>>>>>>>>>> Tue Sep 23 00:30:16 CDT 2014, >>>>>>>>>>>>>>>> org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@100e398, >>>>>>>>>>>>>>>> java.io.IOException: java.io.IOException: >>>>>>>>>>>>>>>> org.apache.phoenix.memory.InsufficientMemoryException: >>>>>>>>>>>>>>>> Requested memory of >>>>>>>>>>>>>>>> 446623727 bytes is larger than global pool of 319507660 bytes. >>>>>>>>>>>>>>>> Tue Sep 23 00:31:22 CDT 2014, >>>>>>>>>>>>>>>> org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@100e398, >>>>>>>>>>>>>>>> java.io.IOException: java.io.IOException: >>>>>>>>>>>>>>>> org.apache.phoenix.memory.InsufficientMemoryException: >>>>>>>>>>>>>>>> Requested memory of >>>>>>>>>>>>>>>> 446623727 bytes is larger than global pool of 319507660 bytes. >>>>>>>>>>>>>>>> Tue Sep 23 00:32:29 CDT 2014, >>>>>>>>>>>>>>>> org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@100e398, >>>>>>>>>>>>>>>> java.io.IOException: java.io.IOException: >>>>>>>>>>>>>>>> org.apache.phoenix.memory.InsufficientMemoryException: >>>>>>>>>>>>>>>> Requested memory of >>>>>>>>>>>>>>>> 446623727 bytes is larger than global pool of 319507660 bytes. >>>>>>>>>>>>>>>> Tue Sep 23 00:33:35 CDT 2014, >>>>>>>>>>>>>>>> org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@100e398, >>>>>>>>>>>>>>>> java.io.IOException: java.io.IOException: >>>>>>>>>>>>>>>> org.apache.phoenix.memory.InsufficientMemoryException: >>>>>>>>>>>>>>>> Requested memory of >>>>>>>>>>>>>>>> 446623727 bytes is larger than global pool of 319507660 bytes. >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> at >>>>>>>>>>>>>>>> org.apache.hadoop.hbase.client.ServerCallable.withRetries(ServerCallable.java:187) >>>>>>>>>>>>>>>> at >>>>>>>>>>>>>>>> org.apache.hadoop.hbase.ipc.ExecRPCInvoker.invoke(ExecRPCInvoker.java:79) >>>>>>>>>>>>>>>> ... 8 more >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Trials: >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> I tried to increase the Region Server Heap space , >>>>>>>>>>>>>>>> modified phoenix.query.maxGlobalMemoryPercentage as well. >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> I am not able to increase the global memory . >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Regards, >>>>>>>>>>>>>>>> Vijay Raajaa >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>> Thanks, >>>>>>>>>>>>>>> Maryann >>>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> -- >>>>>>>>>>>>> Thanks, >>>>>>>>>>>>> Maryann >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> -- >>>>>>>>>>> Thanks, >>>>>>>>>>> Maryann >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> -- >>>>>>>>> Thanks, >>>>>>>>> Maryann >>>>>>>>> >>>>>>>> >>>>>>>> >>>>>>> >>>>>>> >>>>>>> -- >>>>>>> Thanks, >>>>>>> Maryann >>>>>>> >>>>>> >>>>>> >>>>> >>>>> >>>>> -- >>>>> Thanks, >>>>> Maryann >>>>> >>>> >>>> >>> >>> >>> -- >>> Thanks, >>> Maryann >>> >> >> > -- Thanks, Maryann
