@James: Could you point me to a place where i can find tar file of Phoenix-4.0.0-incubating release? All the links on this page are broken: http://www.apache.org/dyn/closer.cgi/incubator/phoenix/
On Thu, Mar 5, 2015 at 5:04 PM, anil gupta <[email protected]> wrote: > I have tried to disable the table but since none of the RS are coming up. > I am unable to do it. Am i missing something? > On the server side, we were using the "4.0.0-incubating". It seems like my > only option is to upgrade the server to 4.1. At-least, the HBase cluster > to be UP. I just want my cluster to come and then i will disable the table > that has a Phoenix view. > What would be the possible side effects of using Phoenix 4.1 with > HDP2.1.5. > Even after updating to Phoenix4.1, if the problem is not fixed. What is > the next alternative? > > > On Thu, Mar 5, 2015 at 4:54 PM, Nick Dimiduk <[email protected]> wrote: > >> Hi Anil, >> >> HDP-2.1.5 ships with Phoenix [0]. Are you using the version shipped, or >> trying out a newer version? As James says, the upgrade must be servers >> first, then client. Also, Phoenix versions tend to be picky about their >> underlying HBase version. >> >> You can also try altering the now-broken phoenix tables via HBase shell, >> removing the phoenix coprocessor. I've tried this in the past with other >> coprocessor-loading woes and had mixed results. Try: disable table, alter >> table, enable table. There's still sharp edges around coprocessor-based >> deployment. >> >> Keep us posted, and sorry for the mess. >> >> -n >> >> [0]: >> http://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.1.7/bk_releasenotes_hdp_2.1/content/ch_relnotes-hdp-2.1.5-product.html >> >> On Thu, Mar 5, 2015 at 4:34 PM, anil gupta <[email protected]> wrote: >> >>> Unfortunately, we ran out of luck on this one because we are not running >>> the latest version of HBase. This property was introduced recently: >>> https://issues.apache.org/jira/browse/HBASE-13044 :( >>> Thanks, Vladimir. >>> >>> On Thu, Mar 5, 2015 at 3:44 PM, Vladimir Rodionov < >>> [email protected]> wrote: >>> >>>> Try the following: >>>> >>>> Update hbase-site.xml config, set >>>> >>>> hbase.coprocessor.enabed=false >>>> >>>> or: >>>> >>>> hbase.coprocessor.user.enabed=false >>>> >>>> sync config across cluster. >>>> >>>> restart the cluster >>>> >>>> than update your table's settings in hbase shell >>>> >>>> -Vlad >>>> >>>> >>>> On Thu, Mar 5, 2015 at 3:32 PM, anil gupta <[email protected]> >>>> wrote: >>>> >>>>> Hi All, >>>>> >>>>> I am using HDP2.1.5, Phoenix4-0.0 was installed on RS. I was running >>>>> Phoenix4.1 client because i could not find tar file for >>>>> "Phoenix4-0.0-incubating". >>>>> I tried to create a view on existing table and then my entire cluster >>>>> went down(all the RS went down. MAster is still up). >>>>> >>>>> >>>>> This is the exception i am seeing: >>>>> >>>>> 2015-03-05 14:30:53,296 FATAL [RS_OPEN_REGION-hdpslave8:60020-2] >>>>> regionserver.HRegionServer: ABORTING region server >>>>> bigdatabox.com,60020,1423589420136: The coprocessor >>>>> org.apache.hadoop.hbase.regionserver.LocalIndexSplitter threw an >>>>> unexpected exception >>>>> java.io.IOException: No jar path specified for >>>>> org.apache.hadoop.hbase.regionserver.LocalIndexSplitter >>>>> at >>>>> org.apache.hadoop.hbase.coprocessor.CoprocessorHost.load(CoprocessorHost.java:177) >>>>> at >>>>> org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.loadTableCoprocessors(RegionCoprocessorHost.java:207) >>>>> at >>>>> org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.<init>(RegionCoprocessorHost.java:163) >>>>> at >>>>> org.apache.hadoop.hbase.regionserver.HRegion.<init>(HRegion.java:555) >>>>> at >>>>> org.apache.hadoop.hbase.regionserver.HRegion.<init>(HRegion.java:462) >>>>> at sun.reflect.GeneratedConstructorAccessor33.newInstance(Unknown >>>>> Source) >>>>> at >>>>> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) >>>>> at java.lang.reflect.Constructor.newInstance(Constructor.java:526) >>>>> at >>>>> org.apache.hadoop.hbase.regionserver.HRegion.newHRegion(HRegion.java:4119) >>>>> at >>>>> org.apache.hadoop.hbase.regionserver.HRegion.openHRegion(HRegion.java:4430) >>>>> at >>>>> org.apache.hadoop.hbase.regionserver.HRegion.openHRegion(HRegion.java:4403) >>>>> at >>>>> org.apache.hadoop.hbase.regionserver.HRegion.openHRegion(HRegion.java:4359) >>>>> at >>>>> org.apache.hadoop.hbase.regionserver.HRegion.openHRegion(HRegion.java:4310) >>>>> at >>>>> org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler.openRegion(OpenRegionHandler.java:465) >>>>> at >>>>> org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler.process(OpenRegionHandler.java:139) >>>>> at >>>>> org.apache.hadoop.hbase.executor.EventHandler.run(EventHandler.java:128) >>>>> at >>>>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) >>>>> at >>>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) >>>>> at java.lang.Thread.run(Thread.java:744) >>>>> >>>>> >>>>> We tried to restart the cluster. It died again. It seems, its stucks at >>>>> this point looking for >>>>> >>>>> LocalIndexSplitter class. How can i resolve this error? We cant do >>>>> anything in the cluster until we fix it. >>>>> >>>>> I was thinking of disabling those tables but none of the RS is coming up. >>>>> Can anyone suggest me how can i bail out of this BAD situation. >>>>> >>>>> >>>>> -- >>>>> Thanks & Regards, >>>>> Anil Gupta >>>>> >>>> >>>> >>> >>> >>> -- >>> Thanks & Regards, >>> Anil Gupta >>> >> >> > > > -- > Thanks & Regards, > Anil Gupta > -- Thanks & Regards, Anil Gupta
