[ 
https://issues.apache.org/jira/browse/PHOENIX-5047?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ievgen Nekrashevych updated PHOENIX-5047:
-----------------------------------------
    Description: 
The upgrade scenario as following:
install phoenix 4.13 on top of hbase 1.2.0-cdh5.13.0. Run simple script to make 
sure some data is there:
{code}
-- system tables are created on the first connection
create schema if not exists TS
create table if not exists TS.TEST (STR varchar not null,INTCOL bigint not 
null, STARTTIME integer, DUMMY integer default 0 CONSTRAINT PK PRIMARY KEY 
(STR, INTCOL))
create local index if not exists "TEST_INDEX" on TS.TEST (STR,STARTTIME)
upsert into TS.TEST(STR,INTCOL,STARTTIME,DUMMY) values ('TEST',4,1,3)
-- made sure there is a data
select * from TS.TEST
{code}

then I shut down everything (queryserver, regionserver, master and zookeeper), 
install hbase 1.2.0-cdh5.14.2, replace phoenix libs with 4.14.1 and start 
servers. Trying to connect to the server and run:
{code}
select * from TS.TEST
{code}

I get:
{code}
2018-11-28 07:53:03,088 ERROR 
[RpcServer.FifoWFPBQ.default.handler=29,queue=2,port=60020] 
coprocessor.MetaDataEndpointImpl: Add column failed: 
org.apache.hadoop.hbase.DoNotRetryIOException: SYSTEM:CATALOG: 63
        at 
org.apache.phoenix.util.ServerUtil.createIOException(ServerUtil.java:120)
        at 
org.apache.phoenix.util.ServerUtil.throwIOException(ServerUtil.java:86)
        at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.mutateColumn(MetaDataEndpointImpl.java:2368)
        at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.addColumn(MetaDataEndpointImpl.java:3242)
        at 
org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService.callMethod(MetaDataProtos.java:16402)
        at 
org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:7931)
        at 
org.apache.hadoop.hbase.regionserver.RSRpcServices.execServiceOnRegion(RSRpcServices.java:1969)
        at 
org.apache.hadoop.hbase.regionserver.RSRpcServices.execService(RSRpcServices.java:1951)
        at 
org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:33652)
        at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2191)
        at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:112)
        at 
org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:183)
        at 
org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:163)
Caused by: java.lang.ArrayIndexOutOfBoundsException: 63
        at org.apache.phoenix.schema.PTableImpl.init(PTableImpl.java:517)
        at org.apache.phoenix.schema.PTableImpl.<init>(PTableImpl.java:421)
        at org.apache.phoenix.schema.PTableImpl.makePTable(PTableImpl.java:406)
        at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.getTable(MetaDataEndpointImpl.java:1073)
        at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.buildTable(MetaDataEndpointImpl.java:614)
        at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.mutateColumn(MetaDataEndpointImpl.java:2361)
        ... 10 more
{code}

In subsequent calls I get same exception with slightly different message that 
I've got different versions of client and server jars (with 
ArrayIndexOutOfBoundsException as cause, and only 
ArrayIndexOutOfBoundsException in server logs), which is not true.

Serverside exception:
{code}
2018-11-28 08:45:00,611 ERROR 
[RpcServer.FifoWFPBQ.default.handler=29,queue=2,port=60020] 
coprocessor.MetaDataEndpointImpl: loading system catalog table inside 
getVersion failed
java.lang.ArrayIndexOutOfBoundsException: 63
        at org.apache.phoenix.schema.PTableImpl.init(PTableImpl.java:517)
        at org.apache.phoenix.schema.PTableImpl.<init>(PTableImpl.java:421)
        at org.apache.phoenix.schema.PTableImpl.makePTable(PTableImpl.java:406)
        at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.getTable(MetaDataEndpointImpl.java:1073)
        at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.buildTable(MetaDataEndpointImpl.java:614)
        at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.loadTable(MetaDataEndpointImpl.java:1339)
        at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.getVersion(MetaDataEndpointImpl.java:3721)
        at 
org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService.callMethod(MetaDataProtos.java:16422)
        at 
org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:7996)
        at 
org.apache.hadoop.hbase.regionserver.RSRpcServices.execServiceOnRegion(RSRpcServices.java:1986)
        at 
org.apache.hadoop.hbase.regionserver.RSRpcServices.execService(RSRpcServices.java:1968)
        at 
org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:33652)
        at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2191)
        at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:112)
        at 
org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:183)
        at 
org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:163)
{code}

clientside:
{code}
[2018-11-28 10:45:00] [INT08][2006] ERROR 2006 (INT08): Incompatible jars 
detected between client and server. Ensure that phoenix-[version]-server.jar is 
put on the classpath of HBase in every region server: 
org.apache.hadoop.hbase.DoNotRetryIOException: SYSTEM:CATALOG: 63
[2018-11-28 10:45:00]   at 
org.apache.phoenix.util.ServerUtil.createIOException(ServerUtil.java:120)
[2018-11-28 10:45:00]   at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.getVersion(MetaDataEndpointImpl.java:3726)
[2018-11-28 10:45:00]   at 
org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService.callMethod(MetaDataProtos.java:16422)
[2018-11-28 10:45:00]   at 
org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:7996)
[2018-11-28 10:45:00]   at 
org.apache.hadoop.hbase.regionserver.RSRpcServices.execServiceOnRegion(RSRpcServices.java:1986)
[2018-11-28 10:45:00]   at 
org.apache.hadoop.hbase.regionserver.RSRpcServices.execService(RSRpcServices.java:1968)
[2018-11-28 10:45:00]   at 
org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:33652)
[2018-11-28 10:45:00]   at 
org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2191)
[2018-11-28 10:45:00]   at 
org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:112)
[2018-11-28 10:45:00]   at 
org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:183)
[2018-11-28 10:45:00]   at 
org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:163)
[2018-11-28 10:45:00] Caused by: java.lang.ArrayIndexOutOfBoundsException: 63
[2018-11-28 10:45:00]   at 
org.apache.phoenix.schema.PTableImpl.init(PTableImpl.java:517)
[2018-11-28 10:45:00]   at 
org.apache.phoenix.schema.PTableImpl.<init>(PTableImpl.java:421)
[2018-11-28 10:45:00]   at 
org.apache.phoenix.schema.PTableImpl.makePTable(PTableImpl.java:406)
[2018-11-28 10:45:00]   at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.getTable(MetaDataEndpointImpl.java:1073)
[2018-11-28 10:45:00]   at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.buildTable(MetaDataEndpointImpl.java:614)
[2018-11-28 10:45:00]   at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.loadTable(MetaDataEndpointImpl.java:1339)
[2018-11-28 10:45:00]   at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.getVersion(MetaDataEndpointImpl.java:3721)
[2018-11-28 10:45:00]   ... 9 more
{code}

Note: phoenix.schema.isNamespaceMappingEnabled is set to true.

  was:
The upgrade scenario as following:
install phoenix 4.13 on top of hbase 1.2.0-cdh5.13.0. Run simple script to make 
sure some data is there:
{code}
-- system tables are created on the first connection
create schema if not exists TS
create table if not exists TS.TEST (STR varchar not null,INTCOL bigint not 
null, STARTTIME integer, DUMMY integer default 0 CONSTRAINT PK PRIMARY KEY 
(STR, INTCOL))
create local index if not exists "TEST_INDEX" on TS.TEST (STR,STARTTIME)
upsert into TS.TEST(STR,INTCOL,STARTTIME,DUMMY) values ('TEST',4,1,3)
-- made sure there is a data
select * from TS.TEST
{code}

then I shut down everything (queryserver, regionserver, master and zookeeper), 
install hbase 1.2.0-cdh5.14.2, replace phoenix libs with 4.14.1 and start 
servers. Trying to connect to the server and run:
{code}
select * from TS.TEST
{code}

I get:
{code}
2018-11-28 07:53:03,088 ERROR 
[RpcServer.FifoWFPBQ.default.handler=29,queue=2,port=60020] 
coprocessor.MetaDataEndpointImpl: Add column failed: 
org.apache.hadoop.hbase.DoNotRetryIOException: SYSTEM:CATALOG: 63
        at 
org.apache.phoenix.util.ServerUtil.createIOException(ServerUtil.java:120)
        at 
org.apache.phoenix.util.ServerUtil.throwIOException(ServerUtil.java:86)
        at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.mutateColumn(MetaDataEndpointImpl.java:2368)
        at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.addColumn(MetaDataEndpointImpl.java:3242)
        at 
org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService.callMethod(MetaDataProtos.java:16402)
        at 
org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:7931)
        at 
org.apache.hadoop.hbase.regionserver.RSRpcServices.execServiceOnRegion(RSRpcServices.java:1969)
        at 
org.apache.hadoop.hbase.regionserver.RSRpcServices.execService(RSRpcServices.java:1951)
        at 
org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:33652)
        at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2191)
        at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:112)
        at 
org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:183)
        at 
org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:163)
Caused by: java.lang.ArrayIndexOutOfBoundsException: 63
        at org.apache.phoenix.schema.PTableImpl.init(PTableImpl.java:517)
        at org.apache.phoenix.schema.PTableImpl.<init>(PTableImpl.java:421)
        at org.apache.phoenix.schema.PTableImpl.makePTable(PTableImpl.java:406)
        at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.getTable(MetaDataEndpointImpl.java:1073)
        at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.buildTable(MetaDataEndpointImpl.java:614)
        at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.mutateColumn(MetaDataEndpointImpl.java:2361)
        ... 10 more
{code}

In subsequent calls I get same exception with slightly different message that 
I've got different versions of client and server jars (with 
ArrayIndexOutOfBoundsException as cause, and only 
ArrayIndexOutOfBoundsException in server logs), which is not true.

Serverside exception:
{code}
2018-11-28 08:45:00,611 ERROR 
[RpcServer.FifoWFPBQ.default.handler=29,queue=2,port=60020] 
coprocessor.MetaDataEndpointImpl: loading system catalog table inside 
getVersion failed
java.lang.ArrayIndexOutOfBoundsException: 63
        at org.apache.phoenix.schema.PTableImpl.init(PTableImpl.java:517)
        at org.apache.phoenix.schema.PTableImpl.<init>(PTableImpl.java:421)
        at org.apache.phoenix.schema.PTableImpl.makePTable(PTableImpl.java:406)
        at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.getTable(MetaDataEndpointImpl.java:1073)
        at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.buildTable(MetaDataEndpointImpl.java:614)
        at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.loadTable(MetaDataEndpointImpl.java:1339)
        at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.getVersion(MetaDataEndpointImpl.java:3721)
        at 
org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService.callMethod(MetaDataProtos.java:16422)
        at 
org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:7996)
        at 
org.apache.hadoop.hbase.regionserver.RSRpcServices.execServiceOnRegion(RSRpcServices.java:1986)
        at 
org.apache.hadoop.hbase.regionserver.RSRpcServices.execService(RSRpcServices.java:1968)
        at 
org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:33652)
        at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2191)
        at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:112)
        at 
org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:183)
        at 
org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:163)
{code}

clientside:
{code}
[2018-11-28 10:45:00] [INT08][2006] ERROR 2006 (INT08): Incompatible jars 
detected between client and server. Ensure that phoenix-[version]-server.jar is 
put on the classpath of HBase in every region server: 
org.apache.hadoop.hbase.DoNotRetryIOException: SYSTEM:CATALOG: 63
[2018-11-28 10:45:00]   at 
org.apache.phoenix.util.ServerUtil.createIOException(ServerUtil.java:120)
[2018-11-28 10:45:00]   at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.getVersion(MetaDataEndpointImpl.java:3726)
[2018-11-28 10:45:00]   at 
org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService.callMethod(MetaDataProtos.java:16422)
[2018-11-28 10:45:00]   at 
org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:7996)
[2018-11-28 10:45:00]   at 
org.apache.hadoop.hbase.regionserver.RSRpcServices.execServiceOnRegion(RSRpcServices.java:1986)
[2018-11-28 10:45:00]   at 
org.apache.hadoop.hbase.regionserver.RSRpcServices.execService(RSRpcServices.java:1968)
[2018-11-28 10:45:00]   at 
org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:33652)
[2018-11-28 10:45:00]   at 
org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2191)
[2018-11-28 10:45:00]   at 
org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:112)
[2018-11-28 10:45:00]   at 
org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:183)
[2018-11-28 10:45:00]   at 
org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:163)
[2018-11-28 10:45:00] Caused by: java.lang.ArrayIndexOutOfBoundsException: 63
[2018-11-28 10:45:00]   at 
org.apache.phoenix.schema.PTableImpl.init(PTableImpl.java:517)
[2018-11-28 10:45:00]   at 
org.apache.phoenix.schema.PTableImpl.<init>(PTableImpl.java:421)
[2018-11-28 10:45:00]   at 
org.apache.phoenix.schema.PTableImpl.makePTable(PTableImpl.java:406)
[2018-11-28 10:45:00]   at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.getTable(MetaDataEndpointImpl.java:1073)
[2018-11-28 10:45:00]   at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.buildTable(MetaDataEndpointImpl.java:614)
[2018-11-28 10:45:00]   at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.loadTable(MetaDataEndpointImpl.java:1339)
[2018-11-28 10:45:00]   at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.getVersion(MetaDataEndpointImpl.java:3721)
[2018-11-28 10:45:00]   ... 9 more
{code}


> can't upgrade phoenix from 4.13 to 4.14.1
> -----------------------------------------
>
>                 Key: PHOENIX-5047
>                 URL: https://issues.apache.org/jira/browse/PHOENIX-5047
>             Project: Phoenix
>          Issue Type: Bug
>    Affects Versions: 4.14.1
>         Environment: 4.13 on top of cdh 5.13.0
> upgrading to 4.14.1 on top of hbase cdh 5.14.2
>            Reporter: Ievgen Nekrashevych
>            Priority: Major
>              Labels: cdh
>
> The upgrade scenario as following:
> install phoenix 4.13 on top of hbase 1.2.0-cdh5.13.0. Run simple script to 
> make sure some data is there:
> {code}
> -- system tables are created on the first connection
> create schema if not exists TS
> create table if not exists TS.TEST (STR varchar not null,INTCOL bigint not 
> null, STARTTIME integer, DUMMY integer default 0 CONSTRAINT PK PRIMARY KEY 
> (STR, INTCOL))
> create local index if not exists "TEST_INDEX" on TS.TEST (STR,STARTTIME)
> upsert into TS.TEST(STR,INTCOL,STARTTIME,DUMMY) values ('TEST',4,1,3)
> -- made sure there is a data
> select * from TS.TEST
> {code}
> then I shut down everything (queryserver, regionserver, master and 
> zookeeper), install hbase 1.2.0-cdh5.14.2, replace phoenix libs with 4.14.1 
> and start servers. Trying to connect to the server and run:
> {code}
> select * from TS.TEST
> {code}
> I get:
> {code}
> 2018-11-28 07:53:03,088 ERROR 
> [RpcServer.FifoWFPBQ.default.handler=29,queue=2,port=60020] 
> coprocessor.MetaDataEndpointImpl: Add column failed: 
> org.apache.hadoop.hbase.DoNotRetryIOException: SYSTEM:CATALOG: 63
>         at 
> org.apache.phoenix.util.ServerUtil.createIOException(ServerUtil.java:120)
>         at 
> org.apache.phoenix.util.ServerUtil.throwIOException(ServerUtil.java:86)
>         at 
> org.apache.phoenix.coprocessor.MetaDataEndpointImpl.mutateColumn(MetaDataEndpointImpl.java:2368)
>         at 
> org.apache.phoenix.coprocessor.MetaDataEndpointImpl.addColumn(MetaDataEndpointImpl.java:3242)
>         at 
> org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService.callMethod(MetaDataProtos.java:16402)
>         at 
> org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:7931)
>         at 
> org.apache.hadoop.hbase.regionserver.RSRpcServices.execServiceOnRegion(RSRpcServices.java:1969)
>         at 
> org.apache.hadoop.hbase.regionserver.RSRpcServices.execService(RSRpcServices.java:1951)
>         at 
> org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:33652)
>         at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2191)
>         at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:112)
>         at 
> org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:183)
>         at 
> org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:163)
> Caused by: java.lang.ArrayIndexOutOfBoundsException: 63
>         at org.apache.phoenix.schema.PTableImpl.init(PTableImpl.java:517)
>         at org.apache.phoenix.schema.PTableImpl.<init>(PTableImpl.java:421)
>         at 
> org.apache.phoenix.schema.PTableImpl.makePTable(PTableImpl.java:406)
>         at 
> org.apache.phoenix.coprocessor.MetaDataEndpointImpl.getTable(MetaDataEndpointImpl.java:1073)
>         at 
> org.apache.phoenix.coprocessor.MetaDataEndpointImpl.buildTable(MetaDataEndpointImpl.java:614)
>         at 
> org.apache.phoenix.coprocessor.MetaDataEndpointImpl.mutateColumn(MetaDataEndpointImpl.java:2361)
>         ... 10 more
> {code}
> In subsequent calls I get same exception with slightly different message that 
> I've got different versions of client and server jars (with 
> ArrayIndexOutOfBoundsException as cause, and only 
> ArrayIndexOutOfBoundsException in server logs), which is not true.
> Serverside exception:
> {code}
> 2018-11-28 08:45:00,611 ERROR 
> [RpcServer.FifoWFPBQ.default.handler=29,queue=2,port=60020] 
> coprocessor.MetaDataEndpointImpl: loading system catalog table inside 
> getVersion failed
> java.lang.ArrayIndexOutOfBoundsException: 63
>         at org.apache.phoenix.schema.PTableImpl.init(PTableImpl.java:517)
>         at org.apache.phoenix.schema.PTableImpl.<init>(PTableImpl.java:421)
>         at 
> org.apache.phoenix.schema.PTableImpl.makePTable(PTableImpl.java:406)
>         at 
> org.apache.phoenix.coprocessor.MetaDataEndpointImpl.getTable(MetaDataEndpointImpl.java:1073)
>         at 
> org.apache.phoenix.coprocessor.MetaDataEndpointImpl.buildTable(MetaDataEndpointImpl.java:614)
>         at 
> org.apache.phoenix.coprocessor.MetaDataEndpointImpl.loadTable(MetaDataEndpointImpl.java:1339)
>         at 
> org.apache.phoenix.coprocessor.MetaDataEndpointImpl.getVersion(MetaDataEndpointImpl.java:3721)
>         at 
> org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService.callMethod(MetaDataProtos.java:16422)
>         at 
> org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:7996)
>         at 
> org.apache.hadoop.hbase.regionserver.RSRpcServices.execServiceOnRegion(RSRpcServices.java:1986)
>         at 
> org.apache.hadoop.hbase.regionserver.RSRpcServices.execService(RSRpcServices.java:1968)
>         at 
> org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:33652)
>         at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2191)
>         at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:112)
>         at 
> org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:183)
>         at 
> org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:163)
> {code}
> clientside:
> {code}
> [2018-11-28 10:45:00] [INT08][2006] ERROR 2006 (INT08): Incompatible jars 
> detected between client and server. Ensure that phoenix-[version]-server.jar 
> is put on the classpath of HBase in every region server: 
> org.apache.hadoop.hbase.DoNotRetryIOException: SYSTEM:CATALOG: 63
> [2018-11-28 10:45:00]         at 
> org.apache.phoenix.util.ServerUtil.createIOException(ServerUtil.java:120)
> [2018-11-28 10:45:00]         at 
> org.apache.phoenix.coprocessor.MetaDataEndpointImpl.getVersion(MetaDataEndpointImpl.java:3726)
> [2018-11-28 10:45:00]         at 
> org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService.callMethod(MetaDataProtos.java:16422)
> [2018-11-28 10:45:00]         at 
> org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:7996)
> [2018-11-28 10:45:00]         at 
> org.apache.hadoop.hbase.regionserver.RSRpcServices.execServiceOnRegion(RSRpcServices.java:1986)
> [2018-11-28 10:45:00]         at 
> org.apache.hadoop.hbase.regionserver.RSRpcServices.execService(RSRpcServices.java:1968)
> [2018-11-28 10:45:00]         at 
> org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:33652)
> [2018-11-28 10:45:00]         at 
> org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2191)
> [2018-11-28 10:45:00]         at 
> org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:112)
> [2018-11-28 10:45:00]         at 
> org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:183)
> [2018-11-28 10:45:00]         at 
> org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:163)
> [2018-11-28 10:45:00] Caused by: java.lang.ArrayIndexOutOfBoundsException: 63
> [2018-11-28 10:45:00]         at 
> org.apache.phoenix.schema.PTableImpl.init(PTableImpl.java:517)
> [2018-11-28 10:45:00]         at 
> org.apache.phoenix.schema.PTableImpl.<init>(PTableImpl.java:421)
> [2018-11-28 10:45:00]         at 
> org.apache.phoenix.schema.PTableImpl.makePTable(PTableImpl.java:406)
> [2018-11-28 10:45:00]         at 
> org.apache.phoenix.coprocessor.MetaDataEndpointImpl.getTable(MetaDataEndpointImpl.java:1073)
> [2018-11-28 10:45:00]         at 
> org.apache.phoenix.coprocessor.MetaDataEndpointImpl.buildTable(MetaDataEndpointImpl.java:614)
> [2018-11-28 10:45:00]         at 
> org.apache.phoenix.coprocessor.MetaDataEndpointImpl.loadTable(MetaDataEndpointImpl.java:1339)
> [2018-11-28 10:45:00]         at 
> org.apache.phoenix.coprocessor.MetaDataEndpointImpl.getVersion(MetaDataEndpointImpl.java:3721)
> [2018-11-28 10:45:00]         ... 9 more
> {code}
> Note: phoenix.schema.isNamespaceMappingEnabled is set to true.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

Reply via email to