[jira] [Updated] (PHOENIX-5863) Upsert into view against a table with index throws exception when 4.14.3 client connects to 4.16 server

2020-05-19 Thread Abhishek Singh Chouhan (Jira)


 [ 
https://issues.apache.org/jira/browse/PHOENIX-5863?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Abhishek Singh Chouhan updated PHOENIX-5863:

Fix Version/s: 4.16.0

> Upsert into view against a table with index throws exception when 4.14.3 
> client connects to 4.16 server
> ---
>
> Key: PHOENIX-5863
> URL: https://issues.apache.org/jira/browse/PHOENIX-5863
> Project: Phoenix
>  Issue Type: Bug
>Affects Versions: 4.16.0
>Reporter: Sandeep Guggilam
>Assignee: Sandeep Guggilam
>Priority: Blocker
> Fix For: 4.16.0
>
> Attachments: PHOENIX-5863.4.x.v1.patch
>
>
> Steps to reproduce:
>  # Bring up a 4.16 server ( current 4.x branch)
>  # Connect a 4.14.3 client to the server and execute below queries
>  # CREATE TABLE IF NOT EXISTS T1 (ID INTEGER NOT NULL PRIMARY KEY, V1 
> VARCHAR(10));
>  # CREATE INDEX T1_IDX ON T1(V1);
>  # CREATE VIEW T1_V1 (V2 INTEGER) AS SELECT * FROM T1 WHERE ID=1;
>  # UPSERT INTO T1_V1(ID, V1, V2) VALUES (1, 'abc', 1); //*This throws below 
> exception***
> *Error: org.apache.hadoop.hbase.client.RetriesExhaustedWithDetailsException: 
> Failed 1 action: org.apache.hadoop.hbase.DoNotRetryIOException: At least one 
> index write failed after retries,serverTimestamp=1587513404084,*
> *at 
> org.apache.phoenix.util.ServerUtil.wrapInDoNotRetryIOException(ServerUtil.java:267)*
> *at 
> org.apache.phoenix.hbase.index.write.TrackingParallelWriterIndexCommitter.write(TrackingParallelWriterIndexCommitter.java:249)*
> *at 
> org.apache.phoenix.hbase.index.write.IndexWriter.write(IndexWriter.java:203)*
> *at 
> org.apache.phoenix.hbase.index.IndexRegionObserver.doIndexWritesWithExceptions(IndexRegionObserver.java:1025)*
> *at 
> org.apache.phoenix.hbase.index.IndexRegionObserver.doPre(IndexRegionObserver.java:1053)*
> *at 
> org.apache.phoenix.hbase.index.IndexRegionObserver.preBatchMutateWithExceptions(IndexRegionObserver.java:938)*
> *at 
> org.apache.phoenix.hbase.index.IndexRegionObserver.preBatchMutate(IndexRegionObserver.java:351)*
> *at 
> org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost$35.call(RegionCoprocessorHost.java:1024)*
> *at 
> org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost$RegionOperation.call(RegionCoprocessorHost.java:1742)*
> *at 
> org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.execOperation(RegionCoprocessorHost.java:1827)*
> *at 
> org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.execOperation(RegionCoprocessorHost.java:1783)*
> *at 
> org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.preBatchMutate(RegionCoprocessorHost.java:1020)*
> *at 
> org.apache.hadoop.hbase.regionserver.HRegion.doMiniBatchMutation(HRegion.java:3420)*
> *at 
> org.apache.hadoop.hbase.regionserver.HRegion.batchMutate(HRegion.java:3158)*
> *at 
> org.apache.hadoop.hbase.regionserver.HRegion.batchMutate(HRegion.java:3100)*
> *at 
> org.apache.hadoop.hbase.regionserver.RSRpcServices.doBatchOp(RSRpcServices.java:944)*
> *at 
> org.apache.hadoop.hbase.regionserver.RSRpcServices.doNonAtomicRegionMutation(RSRpcServices.java:872)*
> *at 
> org.apache.hadoop.hbase.regionserver.RSRpcServices.multi(RSRpcServices.java:2472)*
> *at 
> org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:36812)*
> *at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2399)*
> *at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124)*
> *at org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:311)*
> *at org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:291)*
> *Caused by: 
> org.apache.phoenix.hbase.index.exception.MultiIndexWriteFailureException:  
> disableIndexOnFailure=false, Failed to write to multiple index tables: 
> [V2#I2|#I2]*
> *at 
> org.apache.phoenix.hbase.index.write.TrackingParallelWriterIndexCommitter.write(TrackingParallelWriterIndexCommitter.java:245)*
> *... 21 more*



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (PHOENIX-5863) Upsert into view against a table with index throws exception when 4.14.3 client connects to 4.16 server

2020-04-28 Thread Sandeep Guggilam (Jira)


 [ 
https://issues.apache.org/jira/browse/PHOENIX-5863?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sandeep Guggilam updated PHOENIX-5863:
--
Attachment: PHOENIX-5863.4.x.v1.patch

> Upsert into view against a table with index throws exception when 4.14.3 
> client connects to 4.16 server
> ---
>
> Key: PHOENIX-5863
> URL: https://issues.apache.org/jira/browse/PHOENIX-5863
> Project: Phoenix
>  Issue Type: Bug
>Affects Versions: 4.16.0
>Reporter: Sandeep Guggilam
>Assignee: Sandeep Guggilam
>Priority: Blocker
> Attachments: PHOENIX-5863.4.x.v1.patch
>
>
> Steps to reproduce:
>  # Bring up a 4.16 server ( current 4.x branch)
>  # Connect a 4.14.3 client to the server and execute below queries
>  # CREATE TABLE IF NOT EXISTS T1 (ID INTEGER NOT NULL PRIMARY KEY, V1 
> VARCHAR(10));
>  # CREATE INDEX T1_IDX ON T1(V1);
>  # CREATE VIEW T1_V1 (V2 INTEGER) AS SELECT * FROM T1 WHERE ID=1;
>  # UPSERT INTO T1_V1(ID, V1, V2) VALUES (1, 'abc', 1); //*This throws below 
> exception***
> *Error: org.apache.hadoop.hbase.client.RetriesExhaustedWithDetailsException: 
> Failed 1 action: org.apache.hadoop.hbase.DoNotRetryIOException: At least one 
> index write failed after retries,serverTimestamp=1587513404084,*
> *at 
> org.apache.phoenix.util.ServerUtil.wrapInDoNotRetryIOException(ServerUtil.java:267)*
> *at 
> org.apache.phoenix.hbase.index.write.TrackingParallelWriterIndexCommitter.write(TrackingParallelWriterIndexCommitter.java:249)*
> *at 
> org.apache.phoenix.hbase.index.write.IndexWriter.write(IndexWriter.java:203)*
> *at 
> org.apache.phoenix.hbase.index.IndexRegionObserver.doIndexWritesWithExceptions(IndexRegionObserver.java:1025)*
> *at 
> org.apache.phoenix.hbase.index.IndexRegionObserver.doPre(IndexRegionObserver.java:1053)*
> *at 
> org.apache.phoenix.hbase.index.IndexRegionObserver.preBatchMutateWithExceptions(IndexRegionObserver.java:938)*
> *at 
> org.apache.phoenix.hbase.index.IndexRegionObserver.preBatchMutate(IndexRegionObserver.java:351)*
> *at 
> org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost$35.call(RegionCoprocessorHost.java:1024)*
> *at 
> org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost$RegionOperation.call(RegionCoprocessorHost.java:1742)*
> *at 
> org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.execOperation(RegionCoprocessorHost.java:1827)*
> *at 
> org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.execOperation(RegionCoprocessorHost.java:1783)*
> *at 
> org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.preBatchMutate(RegionCoprocessorHost.java:1020)*
> *at 
> org.apache.hadoop.hbase.regionserver.HRegion.doMiniBatchMutation(HRegion.java:3420)*
> *at 
> org.apache.hadoop.hbase.regionserver.HRegion.batchMutate(HRegion.java:3158)*
> *at 
> org.apache.hadoop.hbase.regionserver.HRegion.batchMutate(HRegion.java:3100)*
> *at 
> org.apache.hadoop.hbase.regionserver.RSRpcServices.doBatchOp(RSRpcServices.java:944)*
> *at 
> org.apache.hadoop.hbase.regionserver.RSRpcServices.doNonAtomicRegionMutation(RSRpcServices.java:872)*
> *at 
> org.apache.hadoop.hbase.regionserver.RSRpcServices.multi(RSRpcServices.java:2472)*
> *at 
> org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:36812)*
> *at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2399)*
> *at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124)*
> *at org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:311)*
> *at org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:291)*
> *Caused by: 
> org.apache.phoenix.hbase.index.exception.MultiIndexWriteFailureException:  
> disableIndexOnFailure=false, Failed to write to multiple index tables: 
> [V2#I2|#I2]*
> *at 
> org.apache.phoenix.hbase.index.write.TrackingParallelWriterIndexCommitter.write(TrackingParallelWriterIndexCommitter.java:245)*
> *... 21 more*



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (PHOENIX-5863) Upsert into view against a table with index throws exception when 4.14.3 client connects to 4.16 server

2020-04-22 Thread Sandeep Guggilam (Jira)


 [ 
https://issues.apache.org/jira/browse/PHOENIX-5863?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sandeep Guggilam updated PHOENIX-5863:
--
Description: 
Steps to reproduce:
 # Bring up a 4.16 server ( current 4.x branch)
 # Connect a 4.14.3 client to the server and execute below queries
 # CREATE TABLE IF NOT EXISTS T1 (ID INTEGER NOT NULL PRIMARY KEY, V1 
VARCHAR(10));
 # CREATE INDEX T1_IDX ON T1(V1);
 # CREATE VIEW T1_V1 (V2 INTEGER) AS SELECT * FROM T1 WHERE ID=1;
 # UPSERT INTO T1_V1(ID, V1, V2) VALUES (1, 'abc', 1); //*This throws below 
exception***

*Error: org.apache.hadoop.hbase.client.RetriesExhaustedWithDetailsException: 
Failed 1 action: org.apache.hadoop.hbase.DoNotRetryIOException: At least one 
index write failed after retries,serverTimestamp=1587513404084,*

*at 
org.apache.phoenix.util.ServerUtil.wrapInDoNotRetryIOException(ServerUtil.java:267)*

*at 
org.apache.phoenix.hbase.index.write.TrackingParallelWriterIndexCommitter.write(TrackingParallelWriterIndexCommitter.java:249)*

*at 
org.apache.phoenix.hbase.index.write.IndexWriter.write(IndexWriter.java:203)*

*at 
org.apache.phoenix.hbase.index.IndexRegionObserver.doIndexWritesWithExceptions(IndexRegionObserver.java:1025)*

*at 
org.apache.phoenix.hbase.index.IndexRegionObserver.doPre(IndexRegionObserver.java:1053)*

*at 
org.apache.phoenix.hbase.index.IndexRegionObserver.preBatchMutateWithExceptions(IndexRegionObserver.java:938)*

*at 
org.apache.phoenix.hbase.index.IndexRegionObserver.preBatchMutate(IndexRegionObserver.java:351)*

*at 
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost$35.call(RegionCoprocessorHost.java:1024)*

*at 
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost$RegionOperation.call(RegionCoprocessorHost.java:1742)*

*at 
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.execOperation(RegionCoprocessorHost.java:1827)*

*at 
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.execOperation(RegionCoprocessorHost.java:1783)*

*at 
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.preBatchMutate(RegionCoprocessorHost.java:1020)*

*at 
org.apache.hadoop.hbase.regionserver.HRegion.doMiniBatchMutation(HRegion.java:3420)*

*at org.apache.hadoop.hbase.regionserver.HRegion.batchMutate(HRegion.java:3158)*

*at org.apache.hadoop.hbase.regionserver.HRegion.batchMutate(HRegion.java:3100)*

*at 
org.apache.hadoop.hbase.regionserver.RSRpcServices.doBatchOp(RSRpcServices.java:944)*

*at 
org.apache.hadoop.hbase.regionserver.RSRpcServices.doNonAtomicRegionMutation(RSRpcServices.java:872)*

*at 
org.apache.hadoop.hbase.regionserver.RSRpcServices.multi(RSRpcServices.java:2472)*

*at 
org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:36812)*

*at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2399)*

*at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124)*

*at org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:311)*

*at org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:291)*

*Caused by: 
org.apache.phoenix.hbase.index.exception.MultiIndexWriteFailureException:  
disableIndexOnFailure=false, Failed to write to multiple index tables: 
[V2#I2|#I2]*

*at 
org.apache.phoenix.hbase.index.write.TrackingParallelWriterIndexCommitter.write(TrackingParallelWriterIndexCommitter.java:245)*

*... 21 more*

  was:
Steps to reproduce:
 # Bring up a 4.16 server ( current 4.x branch)
 # Connect a 4.14.3 client to the server and execute below queries
 # CREATE TABLE IF NOT EXISTS T1 (ID INTEGER NOT NULL PRIMARY KEY, V1 
VARCHAR(10));
 # CREATE INDEX T1_IDX ON T1(V1);
 # CREATE VIEW T1_V1 (V2 INTEGER) AS SELECT * FROM T1 WHERE ID=1;
 # UPSERT INTO T2_V2(ID, V1, V2) VALUES (1, 'abc', 1); //*This throws below 
exception***

*Error: org.apache.hadoop.hbase.client.RetriesExhaustedWithDetailsException: 
Failed 1 action: org.apache.hadoop.hbase.DoNotRetryIOException: At least one 
index write failed after retries,serverTimestamp=1587513404084,*

 *at 
org.apache.phoenix.util.ServerUtil.wrapInDoNotRetryIOException(ServerUtil.java:267)*

 *at 
org.apache.phoenix.hbase.index.write.TrackingParallelWriterIndexCommitter.write(TrackingParallelWriterIndexCommitter.java:249)*

 *at 
org.apache.phoenix.hbase.index.write.IndexWriter.write(IndexWriter.java:203)*

 *at 
org.apache.phoenix.hbase.index.IndexRegionObserver.doIndexWritesWithExceptions(IndexRegionObserver.java:1025)*

 *at 
org.apache.phoenix.hbase.index.IndexRegionObserver.doPre(IndexRegionObserver.java:1053)*

 *at 
org.apache.phoenix.hbase.index.IndexRegionObserver.preBatchMutateWithExceptions(IndexRegionObserver.java:938)*

 *at 
org.apache.phoenix.hbase.index.IndexRegionObserver.preBatchMutate(IndexRegionObserver.java:351)*

 *at 
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost$35.call(RegionCoprocessorHost.java:1024)*

 *at