[jira] [Updated] (IGNITE-22542) Synchronous message handling on local node

2024-06-20 Thread Nikita Sivkov (Jira)


 [ 
https://issues.apache.org/jira/browse/IGNITE-22542?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nikita Sivkov updated IGNITE-22542:
---
Description: 
{{org.apache.ignite.internal.network.DefaultMessagingService#isSelf}} - if we 
detect that we send a message to the local node, we handle it immediately in 
the same thread, which could be very bad for throughput of the system.

"send"/"invoke" themselves appear to be slow as well, we should benchmark them. 
We should remove instantiation of InetSocketAddress for sure, if it's possible, 
it takes time to resolve it. Maybe we should create it unresolved or just cache 
it like in Ignite 2.

  was:
{{org.apache.ignite.internal.network.DefaultMessagingService#isSelf}} - if we 
detect that we send a message to the local node, we handle it immediately in 
the same thread, which could be very bed for throughput of the system.

"send"/"invoke" themselves appear to be slow as well, we should benchmark them. 
We should remove instantiation of InetSocketAddress for sure, if it's possible, 
it takes time to resolve it. Maybe we should create it unresolved or just cache 
it like in Ignite 2.


> Synchronous message handling on local node
> --
>
> Key: IGNITE-22542
> URL: https://issues.apache.org/jira/browse/IGNITE-22542
> Project: Ignite
>  Issue Type: Bug
>Reporter: Ivan Bessonov
>Priority: Major
>  Labels: ignite-3
>
> {{org.apache.ignite.internal.network.DefaultMessagingService#isSelf}} - if we 
> detect that we send a message to the local node, we handle it immediately in 
> the same thread, which could be very bad for throughput of the system.
> "send"/"invoke" themselves appear to be slow as well, we should benchmark 
> them. We should remove instantiation of InetSocketAddress for sure, if it's 
> possible, it takes time to resolve it. Maybe we should create it unresolved 
> or just cache it like in Ignite 2.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (IGNITE-22424) Insert and read errors under high cluster load

2024-06-06 Thread Nikita Sivkov (Jira)


 [ 
https://issues.apache.org/jira/browse/IGNITE-22424?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nikita Sivkov updated IGNITE-22424:
---
Description: 
h2. Setup
 * 1 or 3 server nodes cluster
 * 3 client nodes
 * 64+ threads per client performing KV put/get
 * Each node (server or client) is an AWS instance of {{c5d.4xlarge}}

h2. Steps
 * Client nodes do put 15m records
 * Client nodes do get 15m records

h2. Expected result

No errors occurred, all records inserted and read after successfully.
h2. Actual result

On some point, cluster degrades and throws errors on inserting/reading.

Error examples:
 * The primary replica has changed

{code:java}
org.apache.ignite.lang.IgniteException: The primary replica has changed 
[expectedLeaseholderName=poc-tester-SERVER-192.168.1.58-id-0, 
currentLeaseholderName=null, 
expectedLeaseholderId=b4425409-2d71-482d-a5d6-468522569cd0, 
currentLeaseholderId=null, 
expectedEnlistmentConsistencyToken=112569527664115719, 
currentEnlistmentConsistencyToken=null]
at 
java.lang.invoke.MethodHandle.invokeWithArguments(MethodHandle.java:710) ~[?:?]
at 
org.apache.ignite.internal.util.ExceptionUtils$1.copy(ExceptionUtils.java:789) 
~[ignite-core-3.0.0-SNAPSHOT.jar:?]
at 
org.apache.ignite.internal.util.ExceptionUtils$ExceptionFactory.createCopy(ExceptionUtils.java:723)
 ~[ignite-core-3.0.0-SNAPSHOT.jar:?]
at 
org.apache.ignite.internal.util.ExceptionUtils.copyExceptionWithCause(ExceptionUtils.java:525)
 ~[ignite-core-3.0.0-SNAPSHOT.jar:?]
at 
org.apache.ignite.internal.client.ClientUtils.copyExceptionWithCauseIfPossible(ClientUtils.java:73)
 ~[ignite-client-3.0.0-SNAPSHOT.jar:?]
at 
org.apache.ignite.internal.client.ClientUtils.ensurePublicException(ClientUtils.java:54)
 ~[ignite-client-3.0.0-SNAPSHOT.jar:?]
at 
org.apache.ignite.internal.client.ClientUtils.sync(ClientUtils.java:97) 
~[ignite-client-3.0.0-SNAPSHOT.jar:?]
at 
org.apache.ignite.internal.client.table.ClientKeyValueBinaryView.get(ClientKeyValueBinaryView.java:78)
 ~[ignite-client-3.0.0-SNAPSHOT.jar:?]
at 
org.apache.ignite.internal.client.table.ClientKeyValueBinaryView.get(ClientKeyValueBinaryView.java:59)
 ~[ignite-client-3.0.0-SNAPSHOT.jar:?]
at site.ycsb.db.ignite3.IgniteClient.read(IgniteClient.java:90) 
[ignite3-binding-2024.11.jar:?]
at site.ycsb.DBWrapper.read(DBWrapper.java:157) [core-2024.11.jar:?]
at 
site.ycsb.workloads.CoreWorkload.doTransactionRead(CoreWorkload.java:803) 
[core-2024.11.jar:?]
at 
site.ycsb.workloads.CoreWorkload.doTransaction(CoreWorkload.java:722) 
[core-2024.11.jar:?]
at site.ycsb.ClientThread.run(ClientThread.java:145) 
[core-2024.11.jar:?]
at java.lang.Thread.run(Thread.java:829) [?:?] {code}
 
 * Replication is timed out

{code:java}
org.apache.ignite.tx.TransactionException: Replication is timed out 
[replicaGrpId=10_part_4]
at 
java.lang.invoke.MethodHandle.invokeWithArguments(MethodHandle.java:710) ~[?:?]
at 
org.apache.ignite.internal.util.ExceptionUtils$1.copy(ExceptionUtils.java:789) 
~[ignite-core-3.0.0-SNAPSHOT.jar:?]
at 
org.apache.ignite.internal.util.ExceptionUtils$ExceptionFactory.createCopy(ExceptionUtils.java:723)
 ~[ignite-core-3.0.0-SNAPSHOT.jar:?]
at 
org.apache.ignite.internal.util.ExceptionUtils.copyExceptionWithCause(ExceptionUtils.java:525)
 ~[ignite-core-3.0.0-SNAPSHOT.jar:?]
at 
org.apache.ignite.internal.client.ClientUtils.copyExceptionWithCauseIfPossible(ClientUtils.java:73)
 ~[ignite-client-3.0.0-SNAPSHOT.jar:?]
at 
org.apache.ignite.internal.client.ClientUtils.ensurePublicException(ClientUtils.java:54)
 ~[ignite-client-3.0.0-SNAPSHOT.jar:?]
at 
org.apache.ignite.internal.client.ClientUtils.sync(ClientUtils.java:97) 
~[ignite-client-3.0.0-SNAPSHOT.jar:?]
at 
org.apache.ignite.internal.client.table.ClientKeyValueBinaryView.get(ClientKeyValueBinaryView.java:78)
 ~[ignite-client-3.0.0-SNAPSHOT.jar:?]
at 
org.apache.ignite.internal.client.table.ClientKeyValueBinaryView.get(ClientKeyValueBinaryView.java:59)
 ~[ignite-client-3.0.0-SNAPSHOT.jar:?]
at site.ycsb.db.ignite3.IgniteClient.read(IgniteClient.java:90) 
[ignite3-binding-2024.11.jar:?]
at site.ycsb.DBWrapper.read(DBWrapper.java:157) [core-2024.11.jar:?]
at 
site.ycsb.workloads.CoreWorkload.doTransactionRead(CoreWorkload.java:803) 
[core-2024.11.jar:?]
at 
site.ycsb.workloads.CoreWorkload.doTransaction(CoreWorkload.java:722) 
[core-2024.11.jar:?]
at site.ycsb.ClientThread.run(ClientThread.java:145) 
[core-2024.11.jar:?]
at java.lang.Thread.run(Thread.java:829) [?:?] {code}
 
 * A critical thread is blocked for XXX ms that is more than the allowed 500 ms

{code:java}
2024-06-04 22:26:49:031 +0300 
[ERROR][%poc-tester-SERVER-192.168.1.119-id-0%common-scheduler-0][CriticalWorkerWatchdog]
 A critical thread 

[jira] [Updated] (IGNITE-22424) Insert and read errors under high cluster load

2024-06-06 Thread Nikita Sivkov (Jira)


 [ 
https://issues.apache.org/jira/browse/IGNITE-22424?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nikita Sivkov updated IGNITE-22424:
---
Description: 
h2. Setup
 * 1 or 3 server nodes cluster
 * 3 client nodes
 * 64+ threads per client performing KV put/get
 * Each node (server or client) is an AWS instance of {{c5d.4xlarge}}

h2. Steps
 * Client nodes do put 15m records
 * Client nodes do get 15m records

h2. Expected result

No errors occurred, all records inserted and read after successfully.
h2. Actual result

On some point, cluster degrades and throws errors on inserting/reading.

Error examples:
 * The primary replica has changed

{code:java}
org.apache.ignite.lang.IgniteException: The primary replica has changed 
[expectedLeaseholderName=poc-tester-SERVER-192.168.1.58-id-0, 
currentLeaseholderName=null, 
expectedLeaseholderId=b4425409-2d71-482d-a5d6-468522569cd0, 
currentLeaseholderId=null, 
expectedEnlistmentConsistencyToken=112569527664115719, 
currentEnlistmentConsistencyToken=null]
at 
java.lang.invoke.MethodHandle.invokeWithArguments(MethodHandle.java:710) ~[?:?]
at 
org.apache.ignite.internal.util.ExceptionUtils$1.copy(ExceptionUtils.java:789) 
~[ignite-core-3.0.0-SNAPSHOT.jar:?]
at 
org.apache.ignite.internal.util.ExceptionUtils$ExceptionFactory.createCopy(ExceptionUtils.java:723)
 ~[ignite-core-3.0.0-SNAPSHOT.jar:?]
at 
org.apache.ignite.internal.util.ExceptionUtils.copyExceptionWithCause(ExceptionUtils.java:525)
 ~[ignite-core-3.0.0-SNAPSHOT.jar:?]
at 
org.apache.ignite.internal.client.ClientUtils.copyExceptionWithCauseIfPossible(ClientUtils.java:73)
 ~[ignite-client-3.0.0-SNAPSHOT.jar:?]
at 
org.apache.ignite.internal.client.ClientUtils.ensurePublicException(ClientUtils.java:54)
 ~[ignite-client-3.0.0-SNAPSHOT.jar:?]
at 
org.apache.ignite.internal.client.ClientUtils.sync(ClientUtils.java:97) 
~[ignite-client-3.0.0-SNAPSHOT.jar:?]
at 
org.apache.ignite.internal.client.table.ClientKeyValueBinaryView.get(ClientKeyValueBinaryView.java:78)
 ~[ignite-client-3.0.0-SNAPSHOT.jar:?]
at 
org.apache.ignite.internal.client.table.ClientKeyValueBinaryView.get(ClientKeyValueBinaryView.java:59)
 ~[ignite-client-3.0.0-SNAPSHOT.jar:?]
at site.ycsb.db.ignite3.IgniteClient.read(IgniteClient.java:90) 
[ignite3-binding-2024.11.jar:?]
at site.ycsb.DBWrapper.read(DBWrapper.java:157) [core-2024.11.jar:?]
at 
site.ycsb.workloads.CoreWorkload.doTransactionRead(CoreWorkload.java:803) 
[core-2024.11.jar:?]
at 
site.ycsb.workloads.CoreWorkload.doTransaction(CoreWorkload.java:722) 
[core-2024.11.jar:?]
at site.ycsb.ClientThread.run(ClientThread.java:145) 
[core-2024.11.jar:?]
at java.lang.Thread.run(Thread.java:829) [?:?] {code}
 
 * A critical thread is blocked for XXX ms that is more than the allowed 500 ms

{code:java}
2024-06-04 22:26:49:031 +0300 
[ERROR][%poc-tester-SERVER-192.168.1.119-id-0%common-scheduler-0][CriticalWorkerWatchdog]
 A critical thread is blocked for 512 ms that is more than the allowed 500 ms, 
it is "poc-tester-SERVER-192.168.1.119-id-0-client-32" prio=10 Id=122 RUNNABLE 
(in native)
    at java.base@11.0.12/sun.nio.ch.EPoll.wait(Native Method)
    at 
java.base@11.0.12/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:120)
    at 
java.base@11.0.12/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:124)
    -  locked io.netty.channel.nio.SelectedSelectionKeySet@71be3d25
    -  locked sun.nio.ch.EPollSelectorImpl@5aebdd7
    at java.base@11.0.12/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141)
    at 
app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68)
    at app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879)
    at app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526)
    at 
app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994)
    at 
app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)
    at 
app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
    at java.base@11.0.12/java.lang.Thread.run(Thread.java:829){code}
 
 * Recoverable error during the request occurred (will be retried on the 
randomly selected node)

{code:java}
2024-06-04 22:26:49:306 +0300 
[WARNING][CompletableFutureDelayScheduler][RaftGroupServiceImpl] Recoverable 
error during the request occurred (will be retried on the randomly selected 
node) [request=WriteActionRequestImpl [command=[0, 9, 41, -117, -128, -40, -77, 
-3, -106, -7, -57, 1], deserializedCommand=SafeTimeSyncCommandImpl 
[safeTimeLong=112559993935757322], groupId=10_part_10], peer=Peer 
[consistentId=poc-tester-SERVER-192.168.1.119-id-0, idx=0], newPeer=Peer 
[consistentId=poc-tester-SERVER-192.168.1.119-id-0, idx=0]].
java.util.concurrent.CompletionException: 

[jira] [Created] (IGNITE-22424) Insert and read errors under high cluster load

2024-06-06 Thread Nikita Sivkov (Jira)
Nikita Sivkov created IGNITE-22424:
--

 Summary: Insert and read errors under high cluster load
 Key: IGNITE-22424
 URL: https://issues.apache.org/jira/browse/IGNITE-22424
 Project: Ignite
  Issue Type: Bug
Reporter: Nikita Sivkov


h2. Setup
 * 1 or 3 server nodes cluster
 * 3 client nodes
 * 64+ threads per client performing KV put/get
 * Each node (server or client) is an AWS instance of `c5d.4xlarge`

h2. Steps
 * Client nodes do put 15m records
 * Client nodes do get 15m records

h2. Expected result

No errors occurred, all records inserted and read after successfully.
h2. Actual result

On some point, cluster degrades and throws errors on inserting/reading.

Error examples:
 * The primary replica has changed

{code:java}
org.apache.ignite.lang.IgniteException: The primary replica has changed 
[expectedLeaseholderName=poc-tester-SERVER-192.168.1.58-id-0, 
currentLeaseholderName=null, 
expectedLeaseholderId=b4425409-2d71-482d-a5d6-468522569cd0, 
currentLeaseholderId=null, 
expectedEnlistmentConsistencyToken=112569527664115719, 
currentEnlistmentConsistencyToken=null]
at 
java.lang.invoke.MethodHandle.invokeWithArguments(MethodHandle.java:710) ~[?:?]
at 
org.apache.ignite.internal.util.ExceptionUtils$1.copy(ExceptionUtils.java:789) 
~[ignite-core-3.0.0-SNAPSHOT.jar:?]
at 
org.apache.ignite.internal.util.ExceptionUtils$ExceptionFactory.createCopy(ExceptionUtils.java:723)
 ~[ignite-core-3.0.0-SNAPSHOT.jar:?]
at 
org.apache.ignite.internal.util.ExceptionUtils.copyExceptionWithCause(ExceptionUtils.java:525)
 ~[ignite-core-3.0.0-SNAPSHOT.jar:?]
at 
org.apache.ignite.internal.client.ClientUtils.copyExceptionWithCauseIfPossible(ClientUtils.java:73)
 ~[ignite-client-3.0.0-SNAPSHOT.jar:?]
at 
org.apache.ignite.internal.client.ClientUtils.ensurePublicException(ClientUtils.java:54)
 ~[ignite-client-3.0.0-SNAPSHOT.jar:?]
at 
org.apache.ignite.internal.client.ClientUtils.sync(ClientUtils.java:97) 
~[ignite-client-3.0.0-SNAPSHOT.jar:?]
at 
org.apache.ignite.internal.client.table.ClientKeyValueBinaryView.get(ClientKeyValueBinaryView.java:78)
 ~[ignite-client-3.0.0-SNAPSHOT.jar:?]
at 
org.apache.ignite.internal.client.table.ClientKeyValueBinaryView.get(ClientKeyValueBinaryView.java:59)
 ~[ignite-client-3.0.0-SNAPSHOT.jar:?]
at site.ycsb.db.ignite3.IgniteClient.read(IgniteClient.java:90) 
[ignite3-binding-2024.11.jar:?]
at site.ycsb.DBWrapper.read(DBWrapper.java:157) [core-2024.11.jar:?]
at 
site.ycsb.workloads.CoreWorkload.doTransactionRead(CoreWorkload.java:803) 
[core-2024.11.jar:?]
at 
site.ycsb.workloads.CoreWorkload.doTransaction(CoreWorkload.java:722) 
[core-2024.11.jar:?]
at site.ycsb.ClientThread.run(ClientThread.java:145) 
[core-2024.11.jar:?]
at java.lang.Thread.run(Thread.java:829) [?:?] {code}
 
 * A critical thread is blocked for XXX ms that is more than the allowed 500 ms

{code:java}
2024-06-04 22:26:49:031 +0300 
[ERROR][%poc-tester-SERVER-192.168.1.119-id-0%common-scheduler-0][CriticalWorkerWatchdog]
 A critical thread is blocked for 512 ms that is more than the allowed 500 ms, 
it is "poc-tester-SERVER-192.168.1.119-id-0-client-32" prio=10 Id=122 RUNNABLE 
(in native)
    at java.base@11.0.12/sun.nio.ch.EPoll.wait(Native Method)
    at 
java.base@11.0.12/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:120)
    at 
java.base@11.0.12/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:124)
    -  locked io.netty.channel.nio.SelectedSelectionKeySet@71be3d25
    -  locked sun.nio.ch.EPollSelectorImpl@5aebdd7
    at java.base@11.0.12/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141)
    at 
app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68)
    at app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879)
    at app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526)
    at 
app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994)
    at 
app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)
    at 
app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
    at java.base@11.0.12/java.lang.Thread.run(Thread.java:829){code}
 * Recoverable error during the request occurred (will be retried on the 
randomly selected node)

{code:java}
2024-06-04 22:26:49:306 +0300 
[WARNING][CompletableFutureDelayScheduler][RaftGroupServiceImpl] Recoverable 
error during the request occurred (will be retried on the randomly selected 
node) [request=WriteActionRequestImpl [command=[0, 9, 41, -117, -128, -40, -77, 
-3, -106, -7, -57, 1], deserializedCommand=SafeTimeSyncCommandImpl 
[safeTimeLong=112559993935757322], groupId=10_part_10], peer=Peer 
[consistentId=poc-tester-SERVER-192.168.1.119-id-0, 

[jira] [Updated] (IGNITE-22248) Creation of new tables in 1 node cluster stuck after 850+ tables

2024-05-22 Thread Nikita Sivkov (Jira)


 [ 
https://issues.apache.org/jira/browse/IGNITE-22248?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nikita Sivkov updated IGNITE-22248:
---
Description: 
*Steps to reproduce:*
 # Single node cluster with arguments "-Xms4096m", "-Xmx4096m"
 # Create tables one by one up to 1000

*Expected:*
1000 tables are created.

*Actual:*
After 850+ tables the creation time is higher than 30 seconds or creation table 
request fails.

!image-2024-05-15-13-22-40-059.png!

In the server logs continuous errors:
{code:java}
2024-05-15 04:11:58:116 + 
[WARNING][CompletableFutureDelayScheduler][RaftGroupServiceImpl] Recoverable 
error during the request occurred (will be retried on the randomly selected 
node) [request=WriteActionRequestImpl [command=[0, 9, 41, -126, -128, -36, -49, 
-79, -50, -34, -57, 1], deserializedCommand=SafeTimeSyncCommandImpl 
[safeTimeLong=112443150482997249], groupId=950_part_21], peer=Peer 
[consistentId=TablesAmountCapacityTest_cluster_0, idx=0], newPeer=Peer 
[consistentId=TablesAmountCapacityTest_cluster_0, idx=0]].
java.util.concurrent.CompletionException: java.util.concurrent.TimeoutException
at 
java.base/java.util.concurrent.CompletableFuture.encodeRelay(CompletableFuture.java:368)
at 
java.base/java.util.concurrent.CompletableFuture.completeRelay(CompletableFuture.java:377)
at 
java.base/java.util.concurrent.CompletableFuture$UniRelay.tryFire(CompletableFuture.java:1097)
at 
java.base/java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:510)
at 
java.base/java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:2162)
at 
java.base/java.util.concurrent.CompletableFuture$Timeout.run(CompletableFuture.java:2874)
at 
java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:539)
at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
at 
java.base/java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:304)
at 
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136)
at 
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635)
at java.base/java.lang.Thread.run(Thread.java:833)
Caused by: java.util.concurrent.TimeoutException
... 7 more {code}

  was:
*Steps to reproduce:*
 # Single node cluster with arguments "-Xms4096m", "-Xmx4096m"
 # Create tables one by one up to 1000

*Expected:*
1000 tables are created.

*Actual:*
After 850+ tables the creation time is higher than 30 seconds.

!image-2024-05-15-13-22-40-059.png!

In the server logs continuous errors:
{code:java}
2024-05-15 04:11:58:116 + 
[WARNING][CompletableFutureDelayScheduler][RaftGroupServiceImpl] Recoverable 
error during the request occurred (will be retried on the randomly selected 
node) [request=WriteActionRequestImpl [command=[0, 9, 41, -126, -128, -36, -49, 
-79, -50, -34, -57, 1], deserializedCommand=SafeTimeSyncCommandImpl 
[safeTimeLong=112443150482997249], groupId=950_part_21], peer=Peer 
[consistentId=TablesAmountCapacityTest_cluster_0, idx=0], newPeer=Peer 
[consistentId=TablesAmountCapacityTest_cluster_0, idx=0]].
java.util.concurrent.CompletionException: java.util.concurrent.TimeoutException
at 
java.base/java.util.concurrent.CompletableFuture.encodeRelay(CompletableFuture.java:368)
at 
java.base/java.util.concurrent.CompletableFuture.completeRelay(CompletableFuture.java:377)
at 
java.base/java.util.concurrent.CompletableFuture$UniRelay.tryFire(CompletableFuture.java:1097)
at 
java.base/java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:510)
at 
java.base/java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:2162)
at 
java.base/java.util.concurrent.CompletableFuture$Timeout.run(CompletableFuture.java:2874)
at 
java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:539)
at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
at 
java.base/java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:304)
at 
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136)
at 
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635)
at java.base/java.lang.Thread.run(Thread.java:833)
Caused by: java.util.concurrent.TimeoutException
... 7 more {code}


> Creation of new tables in 1 node cluster stuck after 850+ tables
> 
>
> Key: IGNITE-22248
> URL: https://issues.apache.org/jira/browse/IGNITE-22248
> Project: Ignite
>  Issue Type: Bug
>  

[jira] [Resolved] (IGNITE-22111) [SQL] TPC-H q21 query hangs with sc=0.1

2024-05-07 Thread Nikita Sivkov (Jira)


 [ 
https://issues.apache.org/jira/browse/IGNITE-22111?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nikita Sivkov resolved IGNITE-22111.

Resolution: Not A Problem

Fixed with creating indexes as it done in 
[https://github.com/apache/ignite-3/tree/main/modules/sql-engine/src/testFixtures/resources/tpch/ddl]

> [SQL] TPC-H q21 query hangs with sc=0.1
> ---
>
> Key: IGNITE-22111
> URL: https://issues.apache.org/jira/browse/IGNITE-22111
> Project: Ignite
>  Issue Type: Bug
>  Components: sql
>Reporter: Nikita Sivkov
>Priority: Major
>  Labels: ignite-3, ignite3_performance
>
> Benchmark: 
> [https://github.com/cmu-db/benchbase/tree/main/src/main/java/com/oltpbenchmark/benchmarks/tpch]
>  
> h1. Setup
>  * 1 server node
>  * TPC-H with scale factor = 0.1
> h1. Steps
>  # Start an Ignite node
>  # Run benchbase with {{-s 1 --create=true --load=true --execute=false}} to 
> preload data
>  # Observe via the benchbase log that the data was successfully loaded
>  # Run {{benchbase with -s 1 --create=false --load=false --execute=true}} to 
> run the benchmark
> h1. Expected result
> The benchmark finishes after warmup + duration time
> h1. Actual result
> The benchmark hangs for hours on 
> [Q21|https://github.com/cmu-db/benchbase/blob/main/src/main/java/com/oltpbenchmark/benchmarks/tpch/procedures/Q21.java]
>  query:
>  
> {noformat}
> SELECT
>    s_name,
>    COUNT(*) AS numwait
> FROM
>    supplier,
>    lineitem l1,
>    orders,
>    nation
> WHERE
>    s_suppkey = l1.l_suppkey
>    AND o_orderkey = l1.l_orderkey
>    AND o_orderstatus = 'F'
>    AND l1.l_receiptdate > l1.l_commitdate
>    AND EXISTS
>    (
>       SELECT
>          *
>       FROM
>          lineitem l2
>       WHERE
>          l2.l_orderkey = l1.l_orderkey
>          AND l2.l_suppkey <> l1.l_suppkey
>    )
>    AND NOT EXISTS
>    (
>       SELECT
>          *
>       FROM
>          lineitem l3
>       WHERE
>          l3.l_orderkey = l1.l_orderkey
>          AND l3.l_suppkey <> l1.l_suppkey
>          AND l3.l_receiptdate > l3.l_commitdate
>    )
>    AND s_nationkey = n_nationkey
>    AND n_name = ?
> GROUP BY
>    s_name
> ORDER BY
>    numwait DESC,
>    s_name LIMIT 100{noformat}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Resolved] (IGNITE-22109) [SQL] TPC-H q4 query hangs with sc=0.1

2024-05-07 Thread Nikita Sivkov (Jira)


 [ 
https://issues.apache.org/jira/browse/IGNITE-22109?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nikita Sivkov resolved IGNITE-22109.

Resolution: Not A Problem

Fixed with creating indexes as it done in 
[https://github.com/apache/ignite-3/tree/main/modules/sql-engine/src/testFixtures/resources/tpch/ddl|https://github.com/apache/ignite-3/tree/main/modules/sql-engine/src/testFixtures/resources/tpch/ddl]

> [SQL] TPC-H q4 query hangs with sc=0.1
> --
>
> Key: IGNITE-22109
> URL: https://issues.apache.org/jira/browse/IGNITE-22109
> Project: Ignite
>  Issue Type: Bug
>  Components: sql
>Reporter: Nikita Sivkov
>Priority: Major
>  Labels: ignite-3, ignite3_performance
>
> Benchmark: 
> [https://github.com/cmu-db/benchbase/tree/main/src/main/java/com/oltpbenchmark/benchmarks/tpch]
>  
> h1. Setup
>  * 1 server node
>  * TPC-H with scale factor = 0.1
> h1. Steps
>  # Start an Ignite node
>  # Run benchbase with {{-s 1 --create=true --load=true --execute=false}} to 
> preload data
>  # Observe via the benchbase log that the data was successfully loaded
>  # Run {{benchbase with -s 1 --create=false --load=false --execute=true}} to 
> run the benchmark
> h1. Expected result
> The benchmark finishes after warmup + duration time
> h1. Actual result
> The benchmark hangs for hours on 
> [Q4|https://github.com/cmu-db/benchbase/blob/main/src/main/java/com/oltpbenchmark/benchmarks/tpch/procedures/Q4.java]
>  query:
>  
> {noformat}
> SELECT
>    o_orderpriority,
>    COUNT(*) AS order_count
> FROM
>    orders 
> WHERE
>    o_orderdate >= ?::date
>    AND o_orderdate < ?::date + INTERVAL '3' MONTH
>    AND EXISTS
>    (
>       SELECT
>          *
>       FROM
>          lineitem
>       WHERE 
>          l_orderkey = o_orderkey 
>          AND l_commitdate < l_receiptdate 
>    )
> GROUP BY 
>    o_orderpriority
> ORDER BY
>    o_orderpriority{noformat}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Resolved] (IGNITE-22110) [SQL] TPC-H q16 query hangs with sc=0.1

2024-05-07 Thread Nikita Sivkov (Jira)


 [ 
https://issues.apache.org/jira/browse/IGNITE-22110?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nikita Sivkov resolved IGNITE-22110.

Resolution: Not A Problem

Fixed with creating indexes as it done in 
[https://github.com/apache/ignite-3/tree/main/modules/sql-engine/src/testFixtures/resources/tpch/ddl]

> [SQL] TPC-H q16 query hangs with sc=0.1
> ---
>
> Key: IGNITE-22110
> URL: https://issues.apache.org/jira/browse/IGNITE-22110
> Project: Ignite
>  Issue Type: Bug
>  Components: sql
>Reporter: Nikita Sivkov
>Priority: Major
>  Labels: ignite-3, ignite3_performance
>
> Benchmark: 
> [https://github.com/cmu-db/benchbase/tree/main/src/main/java/com/oltpbenchmark/benchmarks/tpch]
>  
> h1. Setup
>  * 1 server node
>  * TPC-H with scale factor = 0.1
> h1. Steps
>  # Start an Ignite node
>  # Run benchbase with {{-s 1 --create=true --load=true --execute=false}} to 
> preload data
>  # Observe via the benchbase log that the data was successfully loaded
>  # Run {{benchbase with -s 1 --create=false --load=false --execute=true}} to 
> run the benchmark
> h1. Expected result
> The benchmark finishes after warmup + duration time
> h1. Actual result
> The benchmark hangs for hours on 
> [Q16|https://github.com/cmu-db/benchbase/blob/main/src/main/java/com/oltpbenchmark/benchmarks/tpch/procedures/Q16.java]
>  query:
>  
> {noformat}
> SELECT
>    p_brand,
>    p_type,
>    p_size,
>    COUNT(DISTINCT ps_suppkey) AS supplier_cnt
> FROM
>    partsupp,
>    part
> WHERE
>    p_partkey = ps_partkey
>    AND p_brand <> ?
>    AND p_type NOT LIKE ?
>    AND p_size IN (?, ?, ?, ?, ?, ?, ?, ?)
>    AND ps_suppkey NOT IN
>    (
>       SELECT
>          s_suppkey
>       FROM
>          supplier
>       WHERE
>          s_comment LIKE '%Customer%Complaints%'
>    )
> GROUP BY
>    p_brand,
>    p_type,
>    p_size
> ORDER BY
>    supplier_cnt DESC,
>    p_brand,
>    p_type,
>    p_size{noformat}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (IGNITE-22109) [SQL] TPC-H q4 query hangs with sc=0.1

2024-04-25 Thread Nikita Sivkov (Jira)


 [ 
https://issues.apache.org/jira/browse/IGNITE-22109?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nikita Sivkov updated IGNITE-22109:
---
Description: 
Benchmark: 
[https://github.com/cmu-db/benchbase/tree/main/src/main/java/com/oltpbenchmark/benchmarks/tpch]
 
h1. Setup
 * 1 server node
 * TPC-H with scale factor = 0.1

h1. Steps
 # Start an Ignite node
 # Run benchbase with {{-s 1 --create=true --load=true --execute=false}} to 
preload data
 # Observe via the benchbase log that the data was successfully loaded
 # Run {{benchbase with -s 1 --create=false --load=false --execute=true}} to 
run the benchmark

h1. Expected result

The benchmark finishes after warmup + duration time
h1. Actual result

The benchmark hangs for hours on 
[Q4|https://github.com/cmu-db/benchbase/blob/main/src/main/java/com/oltpbenchmark/benchmarks/tpch/procedures/Q4.java]
 query:

 
{noformat}
SELECT
   o_orderpriority,
   COUNT(*) AS order_count
FROM
   orders 
WHERE
   o_orderdate >= ?::date
   AND o_orderdate < ?::date + INTERVAL '3' MONTH
   AND EXISTS
   (
      SELECT
         *
      FROM
         lineitem
      WHERE 
         l_orderkey = o_orderkey 
         AND l_commitdate < l_receiptdate 
   )
GROUP BY 
   o_orderpriority
ORDER BY
   o_orderpriority{noformat}

  was:
Benchmark: 
[https://github.com/cmu-db/benchbase/tree/main/src/main/java/com/oltpbenchmark/benchmarks/tpch]
 
h1. Setup
 * 1 server node
 * TPC-H with scale factor = 0.1

h1. Steps
 # Start an Ignite node
 # Run benchbase with {{-s 1 --create=true --load=true --execute=false}} to 
preload data
 # Observe via the benchbase log that the data was successfully loaded
 # Run {{benchbase with -s 1 --create=false --load=false --execute=true}} to 
run the benchmark

h1. Expected result

The benchmark finishes after warmup + duration time
h1. Actual result

The benchmark hangs for hours on 
[Q4|https://github.com/cmu-db/benchbase/blob/main/src/main/java/com/oltpbenchmark/benchmarks/tpch/procedures/Q4.java]
 query:

 
{noformat}
SELECT
   o_orderpriority,
   COUNT(*) AS order_count
FROM
   orders
WHERE
   o_orderdate >= DATE ?
   AND o_orderdate < DATE ? + INTERVAL '3' MONTH
   AND EXISTS
   (
      SELECT
         *
      FROM
         lineitem
      WHERE
         l_orderkey = o_orderkey
         AND l_commitdate < l_receiptdate
   )
GROUP BY
   o_orderpriority
ORDER BY
   o_orderpriority {noformat}


> [SQL] TPC-H q4 query hangs with sc=0.1
> --
>
> Key: IGNITE-22109
> URL: https://issues.apache.org/jira/browse/IGNITE-22109
> Project: Ignite
>  Issue Type: Bug
>  Components: sql
>Reporter: Nikita Sivkov
>Priority: Major
>  Labels: ignite-3, ignite3_performance
>
> Benchmark: 
> [https://github.com/cmu-db/benchbase/tree/main/src/main/java/com/oltpbenchmark/benchmarks/tpch]
>  
> h1. Setup
>  * 1 server node
>  * TPC-H with scale factor = 0.1
> h1. Steps
>  # Start an Ignite node
>  # Run benchbase with {{-s 1 --create=true --load=true --execute=false}} to 
> preload data
>  # Observe via the benchbase log that the data was successfully loaded
>  # Run {{benchbase with -s 1 --create=false --load=false --execute=true}} to 
> run the benchmark
> h1. Expected result
> The benchmark finishes after warmup + duration time
> h1. Actual result
> The benchmark hangs for hours on 
> [Q4|https://github.com/cmu-db/benchbase/blob/main/src/main/java/com/oltpbenchmark/benchmarks/tpch/procedures/Q4.java]
>  query:
>  
> {noformat}
> SELECT
>    o_orderpriority,
>    COUNT(*) AS order_count
> FROM
>    orders 
> WHERE
>    o_orderdate >= ?::date
>    AND o_orderdate < ?::date + INTERVAL '3' MONTH
>    AND EXISTS
>    (
>       SELECT
>          *
>       FROM
>          lineitem
>       WHERE 
>          l_orderkey = o_orderkey 
>          AND l_commitdate < l_receiptdate 
>    )
> GROUP BY 
>    o_orderpriority
> ORDER BY
>    o_orderpriority{noformat}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (IGNITE-22112) [SQL] TPC-H q9 query with sc=0.1 takes long time

2024-04-25 Thread Nikita Sivkov (Jira)


 [ 
https://issues.apache.org/jira/browse/IGNITE-22112?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nikita Sivkov updated IGNITE-22112:
---
Description: 
Benchmark: 
[https://github.com/cmu-db/benchbase/tree/main/src/main/java/com/oltpbenchmark/benchmarks/tpch]
 
h1. Setup
 * 1 server node
 * TPC-H with scale factor = 0.1

h1. Steps
 # Start an Ignite node
 # Run benchbase with {{-s 1 --create=true --load=true --execute=false}} to 
preload data
 # Observe via the benchbase log that the data was successfully loaded
 # Run {{benchbase with -s 1 --create=false --load=false --execute=true}} to 
run the benchmark

h1. Expected result

The benchmark finishes after warmup + duration time in about the same time as 
other requests.
h1. Actual result

The 
[Q9|https://github.com/cmu-db/benchbase/blob/main/src/main/java/com/oltpbenchmark/benchmarks/tpch/procedures/Q9.java]
 query has significantly longer execution time. For example, it may take 10 
minutes, while the rest queries take ~1-15 seconds.
{noformat}
SELECT
   nation,
   o_year,
   SUM(amount) AS sum_profit
FROM
   (
      SELECT
         n_name AS nation,
         EXTRACT(YEAR
      FROM
         o_orderdate) AS o_year,
         l_extendedprice * (1 - l_discount) - ps_supplycost * l_quantity AS 
amount
      FROM
         part,
         supplier,
         lineitem,
         partsupp,
         orders,
         nation
      WHERE
         s_suppkey = l_suppkey
         AND ps_suppkey = l_suppkey
         AND ps_partkey = l_partkey
         AND p_partkey = l_partkey
         AND o_orderkey = l_orderkey
         AND s_nationkey = n_nationkey
         AND p_name LIKE ?
   )
   AS profit
GROUP BY
   nation,
   o_year
ORDER BY
   nation,
   o_year DESC{noformat}

  was:
Benchmark: 
[https://github.com/cmu-db/benchbase/tree/main/src/main/java/com/oltpbenchmark/benchmarks/tpch]
 
h1. Setup
 * 1 server node
 * TPC-H with scale factor = 0.1

h1. Steps
 # Start an Ignite node
 # Run benchbase with {{-s 1 --create=true --load=true --execute=false}} to 
preload data
 # Observe via the benchbase log that the data was successfully loaded
 # Run {{benchbase with -s 1 --create=false --load=false --execute=true}} to 
run the benchmark

h1. Expected result

The benchmark finishes after warmup + duration time
h1. Actual result

The benchmark hangs for hours on 
[Q21|https://github.com/cmu-db/benchbase/blob/main/src/main/java/com/oltpbenchmark/benchmarks/tpch/procedures/Q21.java]
 query:

 
{noformat}
SELECT
   s_name,
   COUNT(*) AS numwait
FROM
   supplier,
   lineitem l1,
   orders,
   nation
WHERE
   s_suppkey = l1.l_suppkey
   AND o_orderkey = l1.l_orderkey
   AND o_orderstatus = 'F'
   AND l1.l_receiptdate > l1.l_commitdate
   AND EXISTS
   (
      SELECT
         *
      FROM
         lineitem l2
      WHERE
         l2.l_orderkey = l1.l_orderkey
         AND l2.l_suppkey <> l1.l_suppkey
   )
   AND NOT EXISTS
   (
      SELECT
         *
      FROM
         lineitem l3
      WHERE
         l3.l_orderkey = l1.l_orderkey
         AND l3.l_suppkey <> l1.l_suppkey
         AND l3.l_receiptdate > l3.l_commitdate
   )
   AND s_nationkey = n_nationkey
   AND n_name = ?
GROUP BY
   s_name
ORDER BY
   numwait DESC,
   s_name LIMIT 100{noformat}


> [SQL] TPC-H q9 query with sc=0.1 takes long time
> 
>
> Key: IGNITE-22112
> URL: https://issues.apache.org/jira/browse/IGNITE-22112
> Project: Ignite
>  Issue Type: Bug
>  Components: sql
>Reporter: Nikita Sivkov
>Priority: Major
>  Labels: ignite-3, ignite3_performance
>
> Benchmark: 
> [https://github.com/cmu-db/benchbase/tree/main/src/main/java/com/oltpbenchmark/benchmarks/tpch]
>  
> h1. Setup
>  * 1 server node
>  * TPC-H with scale factor = 0.1
> h1. Steps
>  # Start an Ignite node
>  # Run benchbase with {{-s 1 --create=true --load=true --execute=false}} to 
> preload data
>  # Observe via the benchbase log that the data was successfully loaded
>  # Run {{benchbase with -s 1 --create=false --load=false --execute=true}} to 
> run the benchmark
> h1. Expected result
> The benchmark finishes after warmup + duration time in about the same time as 
> other requests.
> h1. Actual result
> The 
> [Q9|https://github.com/cmu-db/benchbase/blob/main/src/main/java/com/oltpbenchmark/benchmarks/tpch/procedures/Q9.java]
>  query has significantly longer execution time. For example, it may take 10 
> minutes, while the rest queries take ~1-15 seconds.
> {noformat}
> SELECT
>    nation,
>    o_year,
>    SUM(amount) AS sum_profit
> FROM
>    (
>       SELECT
>          n_name AS nation,
>          EXTRACT(YEAR
>       FROM
>          o_orderdate) AS o_year,
>          l_extendedprice * (1 - l_discount) - ps_supplycost * l_quantity AS 
> amount
>       FROM
>          part,
>          supplier,
>          lineitem,
>          partsupp,
> 

[jira] [Created] (IGNITE-22112) [SQL] TPC-H q9 query with sc=0.1 takes long time

2024-04-25 Thread Nikita Sivkov (Jira)
Nikita Sivkov created IGNITE-22112:
--

 Summary: [SQL] TPC-H q9 query with sc=0.1 takes long time
 Key: IGNITE-22112
 URL: https://issues.apache.org/jira/browse/IGNITE-22112
 Project: Ignite
  Issue Type: Bug
  Components: sql
Reporter: Nikita Sivkov


Benchmark: 
[https://github.com/cmu-db/benchbase/tree/main/src/main/java/com/oltpbenchmark/benchmarks/tpch]
 
h1. Setup
 * 1 server node
 * TPC-H with scale factor = 0.1

h1. Steps
 # Start an Ignite node
 # Run benchbase with {{-s 1 --create=true --load=true --execute=false}} to 
preload data
 # Observe via the benchbase log that the data was successfully loaded
 # Run {{benchbase with -s 1 --create=false --load=false --execute=true}} to 
run the benchmark

h1. Expected result

The benchmark finishes after warmup + duration time
h1. Actual result

The benchmark hangs for hours on 
[Q21|https://github.com/cmu-db/benchbase/blob/main/src/main/java/com/oltpbenchmark/benchmarks/tpch/procedures/Q21.java]
 query:

 
{noformat}
SELECT
   s_name,
   COUNT(*) AS numwait
FROM
   supplier,
   lineitem l1,
   orders,
   nation
WHERE
   s_suppkey = l1.l_suppkey
   AND o_orderkey = l1.l_orderkey
   AND o_orderstatus = 'F'
   AND l1.l_receiptdate > l1.l_commitdate
   AND EXISTS
   (
      SELECT
         *
      FROM
         lineitem l2
      WHERE
         l2.l_orderkey = l1.l_orderkey
         AND l2.l_suppkey <> l1.l_suppkey
   )
   AND NOT EXISTS
   (
      SELECT
         *
      FROM
         lineitem l3
      WHERE
         l3.l_orderkey = l1.l_orderkey
         AND l3.l_suppkey <> l1.l_suppkey
         AND l3.l_receiptdate > l3.l_commitdate
   )
   AND s_nationkey = n_nationkey
   AND n_name = ?
GROUP BY
   s_name
ORDER BY
   numwait DESC,
   s_name LIMIT 100{noformat}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (IGNITE-22111) [SQL] TPC-H q21 query hangs with sc=0.1

2024-04-25 Thread Nikita Sivkov (Jira)


 [ 
https://issues.apache.org/jira/browse/IGNITE-22111?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nikita Sivkov updated IGNITE-22111:
---
Description: 
Benchmark: 
[https://github.com/cmu-db/benchbase/tree/main/src/main/java/com/oltpbenchmark/benchmarks/tpch]
 
h1. Setup
 * 1 server node
 * TPC-H with scale factor = 0.1

h1. Steps
 # Start an Ignite node
 # Run benchbase with {{-s 1 --create=true --load=true --execute=false}} to 
preload data
 # Observe via the benchbase log that the data was successfully loaded
 # Run {{benchbase with -s 1 --create=false --load=false --execute=true}} to 
run the benchmark

h1. Expected result

The benchmark finishes after warmup + duration time
h1. Actual result

The benchmark hangs for hours on 
[Q21|https://github.com/cmu-db/benchbase/blob/main/src/main/java/com/oltpbenchmark/benchmarks/tpch/procedures/Q21.java]
 query:

 
{noformat}
SELECT
   s_name,
   COUNT(*) AS numwait
FROM
   supplier,
   lineitem l1,
   orders,
   nation
WHERE
   s_suppkey = l1.l_suppkey
   AND o_orderkey = l1.l_orderkey
   AND o_orderstatus = 'F'
   AND l1.l_receiptdate > l1.l_commitdate
   AND EXISTS
   (
      SELECT
         *
      FROM
         lineitem l2
      WHERE
         l2.l_orderkey = l1.l_orderkey
         AND l2.l_suppkey <> l1.l_suppkey
   )
   AND NOT EXISTS
   (
      SELECT
         *
      FROM
         lineitem l3
      WHERE
         l3.l_orderkey = l1.l_orderkey
         AND l3.l_suppkey <> l1.l_suppkey
         AND l3.l_receiptdate > l3.l_commitdate
   )
   AND s_nationkey = n_nationkey
   AND n_name = ?
GROUP BY
   s_name
ORDER BY
   numwait DESC,
   s_name LIMIT 100{noformat}

  was:
Benchmark: 
[https://github.com/cmu-db/benchbase/tree/main/src/main/java/com/oltpbenchmark/benchmarks/tpch]
 
h1. Setup
 * 1 server node
 * TPC-H with scale factor = 0.1

h1. Steps
 # Start an Ignite node
 # Run benchbase with {{-s 1 --create=true --load=true --execute=false}} to 
preload data
 # Observe via the benchbase log that the data was successfully loaded
 # Run {{benchbase with -s 1 --create=false --load=false --execute=true}} to 
run the benchmark

h1. Expected result

The benchmark finishes after warmup + duration time
h1. Actual result

The benchmark hangs for hours on 
[Q16|https://github.com/cmu-db/benchbase/blob/main/src/main/java/com/oltpbenchmark/benchmarks/tpch/procedures/Q16.java]
 query:

 
{noformat}
SELECT
   p_brand,
   p_type,
   p_size,
   COUNT(DISTINCT ps_suppkey) AS supplier_cnt
FROM
   partsupp,
   part
WHERE
   p_partkey = ps_partkey
   AND p_brand <> ?
   AND p_type NOT LIKE ?
   AND p_size IN (?, ?, ?, ?, ?, ?, ?, ?)
   AND ps_suppkey NOT IN
   (
      SELECT
         s_suppkey
      FROM
         supplier
      WHERE
         s_comment LIKE '%Customer%Complaints%'
   )
GROUP BY
   p_brand,
   p_type,
   p_size
ORDER BY
   supplier_cnt DESC,
   p_brand,
   p_type,
   p_size{noformat}


> [SQL] TPC-H q21 query hangs with sc=0.1
> ---
>
> Key: IGNITE-22111
> URL: https://issues.apache.org/jira/browse/IGNITE-22111
> Project: Ignite
>  Issue Type: Bug
>  Components: sql
>Reporter: Nikita Sivkov
>Priority: Major
>  Labels: ignite-3, ignite3_performance
>
> Benchmark: 
> [https://github.com/cmu-db/benchbase/tree/main/src/main/java/com/oltpbenchmark/benchmarks/tpch]
>  
> h1. Setup
>  * 1 server node
>  * TPC-H with scale factor = 0.1
> h1. Steps
>  # Start an Ignite node
>  # Run benchbase with {{-s 1 --create=true --load=true --execute=false}} to 
> preload data
>  # Observe via the benchbase log that the data was successfully loaded
>  # Run {{benchbase with -s 1 --create=false --load=false --execute=true}} to 
> run the benchmark
> h1. Expected result
> The benchmark finishes after warmup + duration time
> h1. Actual result
> The benchmark hangs for hours on 
> [Q21|https://github.com/cmu-db/benchbase/blob/main/src/main/java/com/oltpbenchmark/benchmarks/tpch/procedures/Q21.java]
>  query:
>  
> {noformat}
> SELECT
>    s_name,
>    COUNT(*) AS numwait
> FROM
>    supplier,
>    lineitem l1,
>    orders,
>    nation
> WHERE
>    s_suppkey = l1.l_suppkey
>    AND o_orderkey = l1.l_orderkey
>    AND o_orderstatus = 'F'
>    AND l1.l_receiptdate > l1.l_commitdate
>    AND EXISTS
>    (
>       SELECT
>          *
>       FROM
>          lineitem l2
>       WHERE
>          l2.l_orderkey = l1.l_orderkey
>          AND l2.l_suppkey <> l1.l_suppkey
>    )
>    AND NOT EXISTS
>    (
>       SELECT
>          *
>       FROM
>          lineitem l3
>       WHERE
>          l3.l_orderkey = l1.l_orderkey
>          AND l3.l_suppkey <> l1.l_suppkey
>          AND l3.l_receiptdate > l3.l_commitdate
>    )
>    AND s_nationkey = n_nationkey
>    AND n_name = ?
> GROUP BY
>    s_name
> ORDER BY
>    numwait DESC,
>    s_name LIMIT 100{noformat}



--
This message was sent by 

[jira] [Updated] (IGNITE-22110) [SQL] TPC-H q16 query hangs with sc=0.1

2024-04-25 Thread Nikita Sivkov (Jira)


 [ 
https://issues.apache.org/jira/browse/IGNITE-22110?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nikita Sivkov updated IGNITE-22110:
---
Summary: [SQL] TPC-H q16 query hangs with sc=0.1  (was: CLONE - [SQL] TPC-H 
q16 query hangs with sc=0.1)

> [SQL] TPC-H q16 query hangs with sc=0.1
> ---
>
> Key: IGNITE-22110
> URL: https://issues.apache.org/jira/browse/IGNITE-22110
> Project: Ignite
>  Issue Type: Bug
>  Components: sql
>Reporter: Nikita Sivkov
>Priority: Major
>  Labels: ignite-3, ignite3_performance
>
> Benchmark: 
> [https://github.com/cmu-db/benchbase/tree/main/src/main/java/com/oltpbenchmark/benchmarks/tpch]
>  
> h1. Setup
>  * 1 server node
>  * TPC-H with scale factor = 0.1
> h1. Steps
>  # Start an Ignite node
>  # Run benchbase with {{-s 1 --create=true --load=true --execute=false}} to 
> preload data
>  # Observe via the benchbase log that the data was successfully loaded
>  # Run {{benchbase with -s 1 --create=false --load=false --execute=true}} to 
> run the benchmark
> h1. Expected result
> The benchmark finishes after warmup + duration time
> h1. Actual result
> The benchmark hangs for hours on 
> [Q16|https://github.com/cmu-db/benchbase/blob/main/src/main/java/com/oltpbenchmark/benchmarks/tpch/procedures/Q16.java]
>  query:
>  
> {noformat}
> SELECT
>    p_brand,
>    p_type,
>    p_size,
>    COUNT(DISTINCT ps_suppkey) AS supplier_cnt
> FROM
>    partsupp,
>    part
> WHERE
>    p_partkey = ps_partkey
>    AND p_brand <> ?
>    AND p_type NOT LIKE ?
>    AND p_size IN (?, ?, ?, ?, ?, ?, ?, ?)
>    AND ps_suppkey NOT IN
>    (
>       SELECT
>          s_suppkey
>       FROM
>          supplier
>       WHERE
>          s_comment LIKE '%Customer%Complaints%'
>    )
> GROUP BY
>    p_brand,
>    p_type,
>    p_size
> ORDER BY
>    supplier_cnt DESC,
>    p_brand,
>    p_type,
>    p_size{noformat}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (IGNITE-22110) CLONE - [SQL] TPC-H q16 query hangs with sc=0.1

2024-04-25 Thread Nikita Sivkov (Jira)


 [ 
https://issues.apache.org/jira/browse/IGNITE-22110?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nikita Sivkov updated IGNITE-22110:
---
Description: 
Benchmark: 
[https://github.com/cmu-db/benchbase/tree/main/src/main/java/com/oltpbenchmark/benchmarks/tpch]
 
h1. Setup
 * 1 server node
 * TPC-H with scale factor = 0.1

h1. Steps
 # Start an Ignite node
 # Run benchbase with {{-s 1 --create=true --load=true --execute=false}} to 
preload data
 # Observe via the benchbase log that the data was successfully loaded
 # Run {{benchbase with -s 1 --create=false --load=false --execute=true}} to 
run the benchmark

h1. Expected result

The benchmark finishes after warmup + duration time
h1. Actual result

The benchmark hangs for hours on 
[Q16|https://github.com/cmu-db/benchbase/blob/main/src/main/java/com/oltpbenchmark/benchmarks/tpch/procedures/Q16.java]
 query:

 
{noformat}
SELECT
   p_brand,
   p_type,
   p_size,
   COUNT(DISTINCT ps_suppkey) AS supplier_cnt
FROM
   partsupp,
   part
WHERE
   p_partkey = ps_partkey
   AND p_brand <> ?
   AND p_type NOT LIKE ?
   AND p_size IN (?, ?, ?, ?, ?, ?, ?, ?)
   AND ps_suppkey NOT IN
   (
      SELECT
         s_suppkey
      FROM
         supplier
      WHERE
         s_comment LIKE '%Customer%Complaints%'
   )
GROUP BY
   p_brand,
   p_type,
   p_size
ORDER BY
   supplier_cnt DESC,
   p_brand,
   p_type,
   p_size{noformat}

  was:
Benchmark: 
[https://github.com/cmu-db/benchbase/tree/main/src/main/java/com/oltpbenchmark/benchmarks/tpch]
 
h1. Setup
 * 1 server node
 * TPC-H with scale factor = 0.1

h1. Steps
 # Start an Ignite node
 # Run benchbase with {{-s 1 --create=true --load=true --execute=false}} to 
preload data
 # Observe via the benchbase log that the data was successfully loaded
 # Run {{benchbase with -s 1 --create=false --load=false --execute=true}} to 
run the benchmark

h1. Expected result

The benchmark finishes after warmup + duration time
h1. Actual result

The benchmark hangs for hours on 
[Q4|https://github.com/cmu-db/benchbase/blob/main/src/main/java/com/oltpbenchmark/benchmarks/tpch/procedures/Q4.java]
 query:

 
{noformat}
SELECT
   o_orderpriority,
   COUNT(*) AS order_count
FROM
   orders
WHERE
   o_orderdate >= DATE ?
   AND o_orderdate < DATE ? + INTERVAL '3' MONTH
   AND EXISTS
   (
      SELECT
         *
      FROM
         lineitem
      WHERE
         l_orderkey = o_orderkey
         AND l_commitdate < l_receiptdate
   )
GROUP BY
   o_orderpriority
ORDER BY
   o_orderpriority {noformat}


> CLONE - [SQL] TPC-H q16 query hangs with sc=0.1
> ---
>
> Key: IGNITE-22110
> URL: https://issues.apache.org/jira/browse/IGNITE-22110
> Project: Ignite
>  Issue Type: Bug
>  Components: sql
>Reporter: Nikita Sivkov
>Priority: Major
>  Labels: ignite-3, ignite3_performance
>
> Benchmark: 
> [https://github.com/cmu-db/benchbase/tree/main/src/main/java/com/oltpbenchmark/benchmarks/tpch]
>  
> h1. Setup
>  * 1 server node
>  * TPC-H with scale factor = 0.1
> h1. Steps
>  # Start an Ignite node
>  # Run benchbase with {{-s 1 --create=true --load=true --execute=false}} to 
> preload data
>  # Observe via the benchbase log that the data was successfully loaded
>  # Run {{benchbase with -s 1 --create=false --load=false --execute=true}} to 
> run the benchmark
> h1. Expected result
> The benchmark finishes after warmup + duration time
> h1. Actual result
> The benchmark hangs for hours on 
> [Q16|https://github.com/cmu-db/benchbase/blob/main/src/main/java/com/oltpbenchmark/benchmarks/tpch/procedures/Q16.java]
>  query:
>  
> {noformat}
> SELECT
>    p_brand,
>    p_type,
>    p_size,
>    COUNT(DISTINCT ps_suppkey) AS supplier_cnt
> FROM
>    partsupp,
>    part
> WHERE
>    p_partkey = ps_partkey
>    AND p_brand <> ?
>    AND p_type NOT LIKE ?
>    AND p_size IN (?, ?, ?, ?, ?, ?, ?, ?)
>    AND ps_suppkey NOT IN
>    (
>       SELECT
>          s_suppkey
>       FROM
>          supplier
>       WHERE
>          s_comment LIKE '%Customer%Complaints%'
>    )
> GROUP BY
>    p_brand,
>    p_type,
>    p_size
> ORDER BY
>    supplier_cnt DESC,
>    p_brand,
>    p_type,
>    p_size{noformat}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (IGNITE-22111) [SQL] TPC-H q21 query hangs with sc=0.1

2024-04-25 Thread Nikita Sivkov (Jira)
Nikita Sivkov created IGNITE-22111:
--

 Summary: [SQL] TPC-H q21 query hangs with sc=0.1
 Key: IGNITE-22111
 URL: https://issues.apache.org/jira/browse/IGNITE-22111
 Project: Ignite
  Issue Type: Bug
  Components: sql
Reporter: Nikita Sivkov


Benchmark: 
[https://github.com/cmu-db/benchbase/tree/main/src/main/java/com/oltpbenchmark/benchmarks/tpch]
 
h1. Setup
 * 1 server node
 * TPC-H with scale factor = 0.1

h1. Steps
 # Start an Ignite node
 # Run benchbase with {{-s 1 --create=true --load=true --execute=false}} to 
preload data
 # Observe via the benchbase log that the data was successfully loaded
 # Run {{benchbase with -s 1 --create=false --load=false --execute=true}} to 
run the benchmark

h1. Expected result

The benchmark finishes after warmup + duration time
h1. Actual result

The benchmark hangs for hours on 
[Q16|https://github.com/cmu-db/benchbase/blob/main/src/main/java/com/oltpbenchmark/benchmarks/tpch/procedures/Q16.java]
 query:

 
{noformat}
SELECT
   p_brand,
   p_type,
   p_size,
   COUNT(DISTINCT ps_suppkey) AS supplier_cnt
FROM
   partsupp,
   part
WHERE
   p_partkey = ps_partkey
   AND p_brand <> ?
   AND p_type NOT LIKE ?
   AND p_size IN (?, ?, ?, ?, ?, ?, ?, ?)
   AND ps_suppkey NOT IN
   (
      SELECT
         s_suppkey
      FROM
         supplier
      WHERE
         s_comment LIKE '%Customer%Complaints%'
   )
GROUP BY
   p_brand,
   p_type,
   p_size
ORDER BY
   supplier_cnt DESC,
   p_brand,
   p_type,
   p_size{noformat}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (IGNITE-22110) CLONE - [SQL] TPC-H q16 query hangs with sc=0.1

2024-04-25 Thread Nikita Sivkov (Jira)
Nikita Sivkov created IGNITE-22110:
--

 Summary: CLONE - [SQL] TPC-H q16 query hangs with sc=0.1
 Key: IGNITE-22110
 URL: https://issues.apache.org/jira/browse/IGNITE-22110
 Project: Ignite
  Issue Type: Bug
  Components: sql
Reporter: Nikita Sivkov


Benchmark: 
[https://github.com/cmu-db/benchbase/tree/main/src/main/java/com/oltpbenchmark/benchmarks/tpch]
 
h1. Setup
 * 1 server node
 * TPC-H with scale factor = 0.1

h1. Steps
 # Start an Ignite node
 # Run benchbase with {{-s 1 --create=true --load=true --execute=false}} to 
preload data
 # Observe via the benchbase log that the data was successfully loaded
 # Run {{benchbase with -s 1 --create=false --load=false --execute=true}} to 
run the benchmark

h1. Expected result

The benchmark finishes after warmup + duration time
h1. Actual result

The benchmark hangs for hours on 
[Q4|https://github.com/cmu-db/benchbase/blob/main/src/main/java/com/oltpbenchmark/benchmarks/tpch/procedures/Q4.java]
 query:

 
{noformat}
SELECT
   o_orderpriority,
   COUNT(*) AS order_count
FROM
   orders
WHERE
   o_orderdate >= DATE ?
   AND o_orderdate < DATE ? + INTERVAL '3' MONTH
   AND EXISTS
   (
      SELECT
         *
      FROM
         lineitem
      WHERE
         l_orderkey = o_orderkey
         AND l_commitdate < l_receiptdate
   )
GROUP BY
   o_orderpriority
ORDER BY
   o_orderpriority {noformat}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (IGNITE-22109) [SQL] TPC-H q4 query hangs with sc=0.1

2024-04-25 Thread Nikita Sivkov (Jira)
Nikita Sivkov created IGNITE-22109:
--

 Summary: [SQL] TPC-H q4 query hangs with sc=0.1
 Key: IGNITE-22109
 URL: https://issues.apache.org/jira/browse/IGNITE-22109
 Project: Ignite
  Issue Type: Bug
  Components: sql
Reporter: Nikita Sivkov


Benchmark: 
[https://github.com/cmu-db/benchbase/tree/main/src/main/java/com/oltpbenchmark/benchmarks/tpch]
 
h1. Setup
 * 1 server node
 * TPC-H with scale factor = 0.1

h1. Steps
 # Start an Ignite node
 # Run benchbase with {{-s 1 --create=true --load=true --execute=false}} to 
preload data
 # Observe via the benchbase log that the data was successfully loaded
 # Run {{benchbase with -s 1 --create=false --load=false --execute=true}} to 
run the benchmark

h1. Expected result

The benchmark finishes after warmup + duration time
h1. Actual result

The benchmark hangs for hours on 
[Q4|https://github.com/cmu-db/benchbase/blob/main/src/main/java/com/oltpbenchmark/benchmarks/tpch/procedures/Q4.java]
 query:

 
{noformat}
SELECT
   o_orderpriority,
   COUNT(*) AS order_count
FROM
   orders
WHERE
   o_orderdate >= DATE ?
   AND o_orderdate < DATE ? + INTERVAL '3' MONTH
   AND EXISTS
   (
      SELECT
         *
      FROM
         lineitem
      WHERE
         l_orderkey = o_orderkey
         AND l_commitdate < l_receiptdate
   )
GROUP BY
   o_orderpriority
ORDER BY
   o_orderpriority {noformat}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (IGNITE-22006) 'Failed to process the lease granted message' error under load with balance transfer scenario

2024-04-08 Thread Nikita Sivkov (Jira)


 [ 
https://issues.apache.org/jira/browse/IGNITE-22006?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nikita Sivkov updated IGNITE-22006:
---
Description: 
*Steps to reproduce:*

Perform a long (about 2 hours) load test with a balance transfer scenario (see 
scenario pseudo code in attachments).

*Expected result:*

No errors happen.

*Actual result:*

Get error in server logs - {{Failed to process the lease granted message}}
{code:java}
2024-04-05 17:50:39:180 +0300 
[WARNING][%poc-tester-SERVER-192.168.1.97-id-0%JRaft-Request-Processor-13][NodeImpl]
 Node <127_part_16/poc-tester-SERVER-192.168.1.97-id-0> is not in active state, 
currTerm=3.
2024-04-05 17:50:39:187 +0300 
[WARNING][CompletableFutureDelayScheduler][ReplicaManager] Failed to process 
the lease granted message [msg=LeaseGrantedMessageImpl [force=true, 
groupId=77_part_14, leaseExpirationTimeLong=112219169697759232, 
leaseStartTimeLong=112219161833439373]].
java.util.concurrent.TimeoutException
    at 
java.base/java.util.concurrent.CompletableFuture$Timeout.run(CompletableFuture.java:2792)
    at 
java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)
    at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
    at 
java.base/java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:304)
    at 
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
    at 
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
    at java.base/java.lang.Thread.run(Thread.java:829)
2024-04-05 17:50:39:190 +0300 
[WARNING][%poc-tester-SERVER-192.168.1.97-id-0%JRaft-Request-Processor-34][NodeImpl]
 Node <213_part_14/poc-tester-SERVER-192.168.1.97-id-0> is not in active state, 
currTerm=2.{code}

  was:
*Steps to reproduce:*

Perform a long (about 2 hours) load test with a balance transfer scenario (see 
scenario pseudo code in attachments).

*Expected result:*

No errors happen.

*Actual result:*

Get error in server logs - {{Failed to process replica request}}
{code:java}
2024-04-05 17:50:55:802 +0300 
[WARNING][%poc-tester-SERVER-192.168.1.97-id-0%JRaft-AppendEntries-Processor-2][NodeImpl]
 Node <193_part_15/poc-tester-SERVER-192.168.1.97-id-0> is not in active state, 
currTerm=2.
2024-04-05 17:50:55:805 +0300 
[WARNING][%poc-tester-SERVER-192.168.1.97-id-0%Raft-Group-Client-19][ReplicaManager]
 Failed to process replica request [request=TxFinishReplicaRequestImpl 
[commit=false, commitTimestampLong=0, 
enlistmentConsistencyToken=112218720633356321, groupId=123_part_21, 
groups=HashMap {141_part_13=poc-tester-SERVER-192.168.1.27-id-0, 
139_part_9=poc-tester-SERVER-192.168.1.97-id-0, 
193_part_3=poc-tester-SERVER-192.168.1.27-id-0, 
19_part_23=poc-tester-SERVER-192.168.1.27-id-0, 
117_part_17=poc-tester-SERVER-192.168.1.18-id-0, 
45_part_9=poc-tester-SERVER-192.168.1.18-id-0, 
39_part_3=poc-tester-SERVER-192.168.1.18-id-0, 
77_part_4=poc-tester-SERVER-192.168.1.18-id-0, 
105_part_4=poc-tester-SERVER-192.168.1.18-id-0, 
123_part_21=poc-tester-SERVER-192.168.1.97-id-0, 
103_part_9=poc-tester-SERVER-192.168.1.18-id-0, 
161_part_15=poc-tester-SERVER-192.168.1.27-id-0, 
103_part_22=poc-tester-SERVER-192.168.1.27-id-0, 
89_part_10=poc-tester-SERVER-192.168.1.18-id-0, 
39_part_19=poc-tester-SERVER-192.168.1.27-id-0, 
149_part_13=poc-tester-SERVER-192.168.1.27-id-0, 
97_part_24=poc-tester-SERVER-192.168.1.97-id-0, 
83_part_9=poc-tester-SERVER-192.168.1.27-id-0, 
209_part_10=poc-tester-SERVER-192.168.1.27-id-0, 
185_part_5=poc-tester-SERVER-192.168.1.18-id-0, 
117_part_9=poc-tester-SERVER-192.168.1.27-id-0, 
105_part_22=poc-tester-SERVER-192.168.1.18-id-0}, 
timestampLong=112219170129903617, txId=018eaebd-88ba-0001-606d-62250001]].
java.util.concurrent.CompletionException: 
org.apache.ignite.tx.TransactionException: IGN-TX-7 
TraceId:cb1577e6-ec35-47f0-ab7d-56a0687344ed 
java.util.concurrent.TimeoutException
    at 
java.base/java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:314)
    at 
java.base/java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:319)
    at 
java.base/java.util.concurrent.CompletableFuture.uniHandle(CompletableFuture.java:932)
    at 
java.base/java.util.concurrent.CompletableFuture$UniHandle.tryFire(CompletableFuture.java:907)
    at 
java.base/java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506)
    at 
java.base/java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:2088)
    at 
org.apache.ignite.internal.table.distributed.replicator.PartitionReplicaListener.lambda$applyCmdWithRetryOnSafeTimeReorderException$126(PartitionReplicaListener.java:2806)
    at 
java.base/java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:859)
    at 

[jira] [Created] (IGNITE-22006) 'Failed to process the lease granted message' error under load with balance transfer scenario

2024-04-08 Thread Nikita Sivkov (Jira)
Nikita Sivkov created IGNITE-22006:
--

 Summary: 'Failed to process the lease granted message' error under 
load with balance transfer scenario
 Key: IGNITE-22006
 URL: https://issues.apache.org/jira/browse/IGNITE-22006
 Project: Ignite
  Issue Type: Bug
Affects Versions: 3.0.0-beta2
 Environment: Cluster of 3 nodes
Reporter: Nikita Sivkov
 Attachments: transfer_ign3.yaml

*Steps to reproduce:*

Perform a long (about 2 hours) load test with a balance transfer scenario (see 
scenario pseudo code in attachments).

*Expected result:*

No errors happen.

*Actual result:*

Get error in server logs - {{Failed to process replica request}}
{code:java}
2024-04-05 17:50:55:802 +0300 
[WARNING][%poc-tester-SERVER-192.168.1.97-id-0%JRaft-AppendEntries-Processor-2][NodeImpl]
 Node <193_part_15/poc-tester-SERVER-192.168.1.97-id-0> is not in active state, 
currTerm=2.
2024-04-05 17:50:55:805 +0300 
[WARNING][%poc-tester-SERVER-192.168.1.97-id-0%Raft-Group-Client-19][ReplicaManager]
 Failed to process replica request [request=TxFinishReplicaRequestImpl 
[commit=false, commitTimestampLong=0, 
enlistmentConsistencyToken=112218720633356321, groupId=123_part_21, 
groups=HashMap {141_part_13=poc-tester-SERVER-192.168.1.27-id-0, 
139_part_9=poc-tester-SERVER-192.168.1.97-id-0, 
193_part_3=poc-tester-SERVER-192.168.1.27-id-0, 
19_part_23=poc-tester-SERVER-192.168.1.27-id-0, 
117_part_17=poc-tester-SERVER-192.168.1.18-id-0, 
45_part_9=poc-tester-SERVER-192.168.1.18-id-0, 
39_part_3=poc-tester-SERVER-192.168.1.18-id-0, 
77_part_4=poc-tester-SERVER-192.168.1.18-id-0, 
105_part_4=poc-tester-SERVER-192.168.1.18-id-0, 
123_part_21=poc-tester-SERVER-192.168.1.97-id-0, 
103_part_9=poc-tester-SERVER-192.168.1.18-id-0, 
161_part_15=poc-tester-SERVER-192.168.1.27-id-0, 
103_part_22=poc-tester-SERVER-192.168.1.27-id-0, 
89_part_10=poc-tester-SERVER-192.168.1.18-id-0, 
39_part_19=poc-tester-SERVER-192.168.1.27-id-0, 
149_part_13=poc-tester-SERVER-192.168.1.27-id-0, 
97_part_24=poc-tester-SERVER-192.168.1.97-id-0, 
83_part_9=poc-tester-SERVER-192.168.1.27-id-0, 
209_part_10=poc-tester-SERVER-192.168.1.27-id-0, 
185_part_5=poc-tester-SERVER-192.168.1.18-id-0, 
117_part_9=poc-tester-SERVER-192.168.1.27-id-0, 
105_part_22=poc-tester-SERVER-192.168.1.18-id-0}, 
timestampLong=112219170129903617, txId=018eaebd-88ba-0001-606d-62250001]].
java.util.concurrent.CompletionException: 
org.apache.ignite.tx.TransactionException: IGN-TX-7 
TraceId:cb1577e6-ec35-47f0-ab7d-56a0687344ed 
java.util.concurrent.TimeoutException
    at 
java.base/java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:314)
    at 
java.base/java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:319)
    at 
java.base/java.util.concurrent.CompletableFuture.uniHandle(CompletableFuture.java:932)
    at 
java.base/java.util.concurrent.CompletableFuture$UniHandle.tryFire(CompletableFuture.java:907)
    at 
java.base/java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506)
    at 
java.base/java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:2088)
    at 
org.apache.ignite.internal.table.distributed.replicator.PartitionReplicaListener.lambda$applyCmdWithRetryOnSafeTimeReorderException$126(PartitionReplicaListener.java:2806)
    at 
java.base/java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:859)
    at 
java.base/java.util.concurrent.CompletableFuture$UniWhenComplete.tryFire(CompletableFuture.java:837)
    at 
java.base/java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506)
    at 
java.base/java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:2088)
    at 
org.apache.ignite.internal.raft.RaftGroupServiceImpl.sendWithRetry(RaftGroupServiceImpl.java:550)
    at 
org.apache.ignite.internal.raft.RaftGroupServiceImpl.lambda$handleErrorResponse$44(RaftGroupServiceImpl.java:653)
    at 
java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)
    at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
    at 
java.base/java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:304)
    at 
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
    at 
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
    at java.base/java.lang.Thread.run(Thread.java:829)
Caused by: org.apache.ignite.tx.TransactionException: IGN-TX-7 
TraceId:cb1577e6-ec35-47f0-ab7d-56a0687344ed 
java.util.concurrent.TimeoutException
    at 
org.apache.ignite.internal.table.distributed.replicator.PartitionReplicaListener.lambda$finishTransaction$70(PartitionReplicaListener.java:1867)
    at 

[jira] [Updated] (IGNITE-22005) 'Failed to process replica request' error under load with balance transfer scenario

2024-04-08 Thread Nikita Sivkov (Jira)


 [ 
https://issues.apache.org/jira/browse/IGNITE-22005?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nikita Sivkov updated IGNITE-22005:
---
Description: 
*Steps to reproduce:*

Perform a long (about 2 hours) load test with a balance transfer scenario (see 
scenario pseudo code in attachments).

*Expected result:*

No errors happen.

*Actual result:*

Get error in server logs - {{Failed to process replica request}}
{code:java}
2024-04-05 17:50:55:802 +0300 
[WARNING][%poc-tester-SERVER-192.168.1.97-id-0%JRaft-AppendEntries-Processor-2][NodeImpl]
 Node <193_part_15/poc-tester-SERVER-192.168.1.97-id-0> is not in active state, 
currTerm=2.
2024-04-05 17:50:55:805 +0300 
[WARNING][%poc-tester-SERVER-192.168.1.97-id-0%Raft-Group-Client-19][ReplicaManager]
 Failed to process replica request [request=TxFinishReplicaRequestImpl 
[commit=false, commitTimestampLong=0, 
enlistmentConsistencyToken=112218720633356321, groupId=123_part_21, 
groups=HashMap {141_part_13=poc-tester-SERVER-192.168.1.27-id-0, 
139_part_9=poc-tester-SERVER-192.168.1.97-id-0, 
193_part_3=poc-tester-SERVER-192.168.1.27-id-0, 
19_part_23=poc-tester-SERVER-192.168.1.27-id-0, 
117_part_17=poc-tester-SERVER-192.168.1.18-id-0, 
45_part_9=poc-tester-SERVER-192.168.1.18-id-0, 
39_part_3=poc-tester-SERVER-192.168.1.18-id-0, 
77_part_4=poc-tester-SERVER-192.168.1.18-id-0, 
105_part_4=poc-tester-SERVER-192.168.1.18-id-0, 
123_part_21=poc-tester-SERVER-192.168.1.97-id-0, 
103_part_9=poc-tester-SERVER-192.168.1.18-id-0, 
161_part_15=poc-tester-SERVER-192.168.1.27-id-0, 
103_part_22=poc-tester-SERVER-192.168.1.27-id-0, 
89_part_10=poc-tester-SERVER-192.168.1.18-id-0, 
39_part_19=poc-tester-SERVER-192.168.1.27-id-0, 
149_part_13=poc-tester-SERVER-192.168.1.27-id-0, 
97_part_24=poc-tester-SERVER-192.168.1.97-id-0, 
83_part_9=poc-tester-SERVER-192.168.1.27-id-0, 
209_part_10=poc-tester-SERVER-192.168.1.27-id-0, 
185_part_5=poc-tester-SERVER-192.168.1.18-id-0, 
117_part_9=poc-tester-SERVER-192.168.1.27-id-0, 
105_part_22=poc-tester-SERVER-192.168.1.18-id-0}, 
timestampLong=112219170129903617, txId=018eaebd-88ba-0001-606d-62250001]].
java.util.concurrent.CompletionException: 
org.apache.ignite.tx.TransactionException: IGN-TX-7 
TraceId:cb1577e6-ec35-47f0-ab7d-56a0687344ed 
java.util.concurrent.TimeoutException
    at 
java.base/java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:314)
    at 
java.base/java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:319)
    at 
java.base/java.util.concurrent.CompletableFuture.uniHandle(CompletableFuture.java:932)
    at 
java.base/java.util.concurrent.CompletableFuture$UniHandle.tryFire(CompletableFuture.java:907)
    at 
java.base/java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506)
    at 
java.base/java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:2088)
    at 
org.apache.ignite.internal.table.distributed.replicator.PartitionReplicaListener.lambda$applyCmdWithRetryOnSafeTimeReorderException$126(PartitionReplicaListener.java:2806)
    at 
java.base/java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:859)
    at 
java.base/java.util.concurrent.CompletableFuture$UniWhenComplete.tryFire(CompletableFuture.java:837)
    at 
java.base/java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506)
    at 
java.base/java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:2088)
    at 
org.apache.ignite.internal.raft.RaftGroupServiceImpl.sendWithRetry(RaftGroupServiceImpl.java:550)
    at 
org.apache.ignite.internal.raft.RaftGroupServiceImpl.lambda$handleErrorResponse$44(RaftGroupServiceImpl.java:653)
    at 
java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)
    at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
    at 
java.base/java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:304)
    at 
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
    at 
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
    at java.base/java.lang.Thread.run(Thread.java:829)
Caused by: org.apache.ignite.tx.TransactionException: IGN-TX-7 
TraceId:cb1577e6-ec35-47f0-ab7d-56a0687344ed 
java.util.concurrent.TimeoutException
    at 
org.apache.ignite.internal.table.distributed.replicator.PartitionReplicaListener.lambda$finishTransaction$70(PartitionReplicaListener.java:1867)
    at 
java.base/java.util.concurrent.CompletableFuture.uniHandle(CompletableFuture.java:930)
    ... 16 more
Caused by: java.util.concurrent.CompletionException: 
java.util.concurrent.TimeoutException
    at 
java.base/java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:331)
    at 

[jira] [Created] (IGNITE-22005) 'Failed to process replica request' error under load with balance transfer scenario

2024-04-08 Thread Nikita Sivkov (Jira)
Nikita Sivkov created IGNITE-22005:
--

 Summary: 'Failed to process replica request' error under load with 
balance transfer scenario
 Key: IGNITE-22005
 URL: https://issues.apache.org/jira/browse/IGNITE-22005
 Project: Ignite
  Issue Type: Bug
Affects Versions: 3.0.0-beta2
 Environment: Cluster of 3 nodes
Reporter: Nikita Sivkov
 Attachments: transfer_ign3.yaml

*Steps to reproduce:*

Perform a long (about 2 hours) load test with a balance transfer scenario (see 
scenario pseudo code in attachments).

*Expected result:*

No errors happen.

*Actual result:*

Get error in server logs - {{Failed to process delayed response}}
{code:java}
2024-04-05 17:50:50:776 +0300 
[WARNING][%poc-tester-SERVER-192.168.1.97-id-0%JRaft-Request-Processor-1][NodeImpl]
 Node <27_part_23/poc-tester-SERVER-192.168.1.97-id-0> is not in active state, 
currTerm=2.
2024-04-05 17:50:50:778 +0300 
[WARNING][%poc-tester-SERVER-192.168.1.97-id-0%Raft-Group-Client-5][ReplicaManager]
 Failed to process delayed response 
[request=ReadWriteSingleRowReplicaRequestImpl 
[commitPartitionId=TablePartitionIdMessageImpl [partitionId=21, tableId=123], 
coordinatorId=3de6f999-7ab9-4405-aff0-ee0c7e4886ce, 
enlistmentConsistencyToken=112218720633356321, full=false, groupId=123_part_21, 
requestType=RW_UPSERT, schemaVersion=1, timestampLong=112219169796915211, 
transactionId=018eaebd-88ba-0001-606d-62250001]]
java.util.concurrent.CompletionException: java.util.concurrent.TimeoutException
    at 
java.base/java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:331)
    at 
java.base/java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:346)
    at 
java.base/java.util.concurrent.CompletableFuture$UniApply.tryFire(CompletableFuture.java:632)
    at 
java.base/java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506)
    at 
java.base/java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:2088)
    at 
org.apache.ignite.internal.raft.RaftGroupServiceImpl.sendWithRetry(RaftGroupServiceImpl.java:550)
    at 
org.apache.ignite.internal.raft.RaftGroupServiceImpl.lambda$handleErrorResponse$44(RaftGroupServiceImpl.java:653)
    at 
java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)
    at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
    at 
java.base/java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:304)
    at 
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
    at 
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
    at java.base/java.lang.Thread.run(Thread.java:829)
Caused by: java.util.concurrent.TimeoutException
    ... 8 more
2024-04-05 17:50:50:780 +0300 
[WARNING][%poc-tester-SERVER-192.168.1.97-id-0%JRaft-Request-Processor-27][NodeImpl]
 Node <99_part_6/poc-tester-SERVER-192.168.1.97-id-0> is not in active state, 
currTerm=3. {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (IGNITE-22004) 'Failed to process delayed response' error under load with balance transfer scenario

2024-04-08 Thread Nikita Sivkov (Jira)


 [ 
https://issues.apache.org/jira/browse/IGNITE-22004?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nikita Sivkov updated IGNITE-22004:
---
Attachment: transfer_ign3.yaml

> 'Failed to process delayed response' error under load with balance transfer 
> scenario
> 
>
> Key: IGNITE-22004
> URL: https://issues.apache.org/jira/browse/IGNITE-22004
> Project: Ignite
>  Issue Type: Bug
>Affects Versions: 3.0.0-beta2
> Environment: Cluster of 3 nodes
>Reporter: Nikita Sivkov
>Priority: Major
>  Labels: ignite-3
> Attachments: transfer_ign3.yaml
>
>
> *Steps to reproduce:*
> Perform a long (about 2 hours) load test with a balance transfer scenario 
> (see scenario pseudo code in attachments).
> *Expected result:*
> No errors happen.
> *Actual result:*
> Get error in server logs - {{Failed to process delayed response}}
> {code:java}
> 2024-04-05 17:50:50:776 +0300 
> [WARNING][%poc-tester-SERVER-192.168.1.97-id-0%JRaft-Request-Processor-1][NodeImpl]
>  Node <27_part_23/poc-tester-SERVER-192.168.1.97-id-0> is not in active 
> state, currTerm=2.
> 2024-04-05 17:50:50:778 +0300 
> [WARNING][%poc-tester-SERVER-192.168.1.97-id-0%Raft-Group-Client-5][ReplicaManager]
>  Failed to process delayed response 
> [request=ReadWriteSingleRowReplicaRequestImpl 
> [commitPartitionId=TablePartitionIdMessageImpl [partitionId=21, tableId=123], 
> coordinatorId=3de6f999-7ab9-4405-aff0-ee0c7e4886ce, 
> enlistmentConsistencyToken=112218720633356321, full=false, 
> groupId=123_part_21, requestType=RW_UPSERT, schemaVersion=1, 
> timestampLong=112219169796915211, 
> transactionId=018eaebd-88ba-0001-606d-62250001]]
> java.util.concurrent.CompletionException: 
> java.util.concurrent.TimeoutException
>     at 
> java.base/java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:331)
>     at 
> java.base/java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:346)
>     at 
> java.base/java.util.concurrent.CompletableFuture$UniApply.tryFire(CompletableFuture.java:632)
>     at 
> java.base/java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506)
>     at 
> java.base/java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:2088)
>     at 
> org.apache.ignite.internal.raft.RaftGroupServiceImpl.sendWithRetry(RaftGroupServiceImpl.java:550)
>     at 
> org.apache.ignite.internal.raft.RaftGroupServiceImpl.lambda$handleErrorResponse$44(RaftGroupServiceImpl.java:653)
>     at 
> java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)
>     at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
>     at 
> java.base/java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:304)
>     at 
> java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
>     at 
> java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
>     at java.base/java.lang.Thread.run(Thread.java:829)
> Caused by: java.util.concurrent.TimeoutException
>     ... 8 more
> 2024-04-05 17:50:50:780 +0300 
> [WARNING][%poc-tester-SERVER-192.168.1.97-id-0%JRaft-Request-Processor-27][NodeImpl]
>  Node <99_part_6/poc-tester-SERVER-192.168.1.97-id-0> is not in active state, 
> currTerm=3. {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (IGNITE-22004) 'Failed to process delayed response' error under load with balance transfer scenario

2024-04-08 Thread Nikita Sivkov (Jira)


 [ 
https://issues.apache.org/jira/browse/IGNITE-22004?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nikita Sivkov updated IGNITE-22004:
---
Description: 
*Steps to reproduce:*

Perform a long (about 2 hours) load test with a balance transfer scenario (see 
scenario pseudo code in attachments).

*Expected result:*

No errors happen.

*Actual result:*

Get error in server logs - {{Failed to process delayed response}}
{code:java}
2024-04-05 17:50:50:776 +0300 
[WARNING][%poc-tester-SERVER-192.168.1.97-id-0%JRaft-Request-Processor-1][NodeImpl]
 Node <27_part_23/poc-tester-SERVER-192.168.1.97-id-0> is not in active state, 
currTerm=2.
2024-04-05 17:50:50:778 +0300 
[WARNING][%poc-tester-SERVER-192.168.1.97-id-0%Raft-Group-Client-5][ReplicaManager]
 Failed to process delayed response 
[request=ReadWriteSingleRowReplicaRequestImpl 
[commitPartitionId=TablePartitionIdMessageImpl [partitionId=21, tableId=123], 
coordinatorId=3de6f999-7ab9-4405-aff0-ee0c7e4886ce, 
enlistmentConsistencyToken=112218720633356321, full=false, groupId=123_part_21, 
requestType=RW_UPSERT, schemaVersion=1, timestampLong=112219169796915211, 
transactionId=018eaebd-88ba-0001-606d-62250001]]
java.util.concurrent.CompletionException: java.util.concurrent.TimeoutException
    at 
java.base/java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:331)
    at 
java.base/java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:346)
    at 
java.base/java.util.concurrent.CompletableFuture$UniApply.tryFire(CompletableFuture.java:632)
    at 
java.base/java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506)
    at 
java.base/java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:2088)
    at 
org.apache.ignite.internal.raft.RaftGroupServiceImpl.sendWithRetry(RaftGroupServiceImpl.java:550)
    at 
org.apache.ignite.internal.raft.RaftGroupServiceImpl.lambda$handleErrorResponse$44(RaftGroupServiceImpl.java:653)
    at 
java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)
    at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
    at 
java.base/java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:304)
    at 
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
    at 
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
    at java.base/java.lang.Thread.run(Thread.java:829)
Caused by: java.util.concurrent.TimeoutException
    ... 8 more
2024-04-05 17:50:50:780 +0300 
[WARNING][%poc-tester-SERVER-192.168.1.97-id-0%JRaft-Request-Processor-27][NodeImpl]
 Node <99_part_6/poc-tester-SERVER-192.168.1.97-id-0> is not in active state, 
currTerm=3. {code}

  was:
*Steps to reproduce:*

Perform a long (about 2 hours) load test with a balance transfer scenario.

*Expected result:*

No errors happen.

*Actual result:*

Get error in server logs - {{Failed to process delayed response}}
{code:java}
2024-04-05 17:50:50:776 +0300 
[WARNING][%poc-tester-SERVER-192.168.1.97-id-0%JRaft-Request-Processor-1][NodeImpl]
 Node <27_part_23/poc-tester-SERVER-192.168.1.97-id-0> is not in active state, 
currTerm=2.
2024-04-05 17:50:50:778 +0300 
[WARNING][%poc-tester-SERVER-192.168.1.97-id-0%Raft-Group-Client-5][ReplicaManager]
 Failed to process delayed response 
[request=ReadWriteSingleRowReplicaRequestImpl 
[commitPartitionId=TablePartitionIdMessageImpl [partitionId=21, tableId=123], 
coordinatorId=3de6f999-7ab9-4405-aff0-ee0c7e4886ce, 
enlistmentConsistencyToken=112218720633356321, full=false, groupId=123_part_21, 
requestType=RW_UPSERT, schemaVersion=1, timestampLong=112219169796915211, 
transactionId=018eaebd-88ba-0001-606d-62250001]]
java.util.concurrent.CompletionException: java.util.concurrent.TimeoutException
    at 
java.base/java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:331)
    at 
java.base/java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:346)
    at 
java.base/java.util.concurrent.CompletableFuture$UniApply.tryFire(CompletableFuture.java:632)
    at 
java.base/java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506)
    at 
java.base/java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:2088)
    at 
org.apache.ignite.internal.raft.RaftGroupServiceImpl.sendWithRetry(RaftGroupServiceImpl.java:550)
    at 
org.apache.ignite.internal.raft.RaftGroupServiceImpl.lambda$handleErrorResponse$44(RaftGroupServiceImpl.java:653)
    at 
java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)
    at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
    at 
java.base/java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:304)
    at 

[jira] [Created] (IGNITE-22004) 'Failed to process delayed response' error under load with balance transfer scenario

2024-04-08 Thread Nikita Sivkov (Jira)
Nikita Sivkov created IGNITE-22004:
--

 Summary: 'Failed to process delayed response' error under load 
with balance transfer scenario
 Key: IGNITE-22004
 URL: https://issues.apache.org/jira/browse/IGNITE-22004
 Project: Ignite
  Issue Type: Bug
Affects Versions: 3.0.0-beta2
 Environment: Cluster of 3 nodes
Reporter: Nikita Sivkov


*Steps to reproduce:*

Perform a long (about 2 hours) load test with a balance transfer scenario.

*Expected result:*

No errors happen.

*Actual result:*

Get error in server logs - {{Failed to process delayed response}}
{code:java}
2024-04-05 17:50:50:776 +0300 
[WARNING][%poc-tester-SERVER-192.168.1.97-id-0%JRaft-Request-Processor-1][NodeImpl]
 Node <27_part_23/poc-tester-SERVER-192.168.1.97-id-0> is not in active state, 
currTerm=2.
2024-04-05 17:50:50:778 +0300 
[WARNING][%poc-tester-SERVER-192.168.1.97-id-0%Raft-Group-Client-5][ReplicaManager]
 Failed to process delayed response 
[request=ReadWriteSingleRowReplicaRequestImpl 
[commitPartitionId=TablePartitionIdMessageImpl [partitionId=21, tableId=123], 
coordinatorId=3de6f999-7ab9-4405-aff0-ee0c7e4886ce, 
enlistmentConsistencyToken=112218720633356321, full=false, groupId=123_part_21, 
requestType=RW_UPSERT, schemaVersion=1, timestampLong=112219169796915211, 
transactionId=018eaebd-88ba-0001-606d-62250001]]
java.util.concurrent.CompletionException: java.util.concurrent.TimeoutException
    at 
java.base/java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:331)
    at 
java.base/java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:346)
    at 
java.base/java.util.concurrent.CompletableFuture$UniApply.tryFire(CompletableFuture.java:632)
    at 
java.base/java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506)
    at 
java.base/java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:2088)
    at 
org.apache.ignite.internal.raft.RaftGroupServiceImpl.sendWithRetry(RaftGroupServiceImpl.java:550)
    at 
org.apache.ignite.internal.raft.RaftGroupServiceImpl.lambda$handleErrorResponse$44(RaftGroupServiceImpl.java:653)
    at 
java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)
    at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
    at 
java.base/java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:304)
    at 
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
    at 
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
    at java.base/java.lang.Thread.run(Thread.java:829)
Caused by: java.util.concurrent.TimeoutException
    ... 8 more
2024-04-05 17:50:50:780 +0300 
[WARNING][%poc-tester-SERVER-192.168.1.97-id-0%JRaft-Request-Processor-27][NodeImpl]
 Node <99_part_6/poc-tester-SERVER-192.168.1.97-id-0> is not in active state, 
currTerm=3. {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Resolved] (IGNITE-21777) 'Failed to get the primary replica' or 'Replication is timed out' or hangs with 'aimem' storage engine

2024-03-28 Thread Nikita Sivkov (Jira)


 [ 
https://issues.apache.org/jira/browse/IGNITE-21777?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nikita Sivkov resolved IGNITE-21777.

Resolution: Cannot Reproduce

> 'Failed to get the primary replica' or 'Replication is timed out' or hangs 
> with 'aimem' storage engine
> --
>
> Key: IGNITE-21777
> URL: https://issues.apache.org/jira/browse/IGNITE-21777
> Project: Ignite
>  Issue Type: Bug
>Affects Versions: 3.0.0-beta2
> Environment: Cluster of 2 nodes.
> Storage engine - aimem.
>Reporter: Nikita Sivkov
>Priority: Major
>  Labels: ignite-3
>
> *Steps to reproduce:*
>  # Create a cluster with 2 nodes.
>  # Connect via JDBC.
>  # Repeat the following SQL statements in a loop (for example, 100 times):
>  ** {{drop table if exists tags}}
>  ** {{create zone if not exists "AIMEM" engine aimem}}
>  ** {{create table tags(tagId integer not null, tag varchar(100) not null, 
> primary key (tagId)) with PRIMARY_ZONE='AIMEM'}}
>  ** {{insert into tags(tagId, tag) values (1,'unit'), (2,'integration'), 
> (3,'smoke'), (4,'sanity'), (5,'regression')}}
> *Expected result:*
> No errors or hangs happen.
> *Actual result:*
> Hangs on {{Create table}} or {{Insert into}} statement.
> _*OR*_
> Get the error {{Replication is timed out}}
> {code:java}
> Replication is timed out [replicaGrpId=34_part_16]
> java.sql.SQLException: Replication is timed out [replicaGrpId=34_part_16]
>     at 
> org.apache.ignite.internal.jdbc.proto.IgniteQueryErrorCode.createJdbcSqlException(IgniteQueryErrorCode.java:57)
>     at 
> org.apache.ignite.internal.jdbc.JdbcStatement.execute0(JdbcStatement.java:154)
>     at 
> org.apache.ignite.internal.jdbc.JdbcStatement.executeUpdate(JdbcStatement.java:181)
>     at 
> org.gridgain.ai3tests.tests.teststeps.JdbcSteps.executeUpdateQuery(JdbcSteps.java:116)
>     at 
> org.gridgain.ai3tests.tests.UpdateTests.wannaCatchTheBug(UpdateTests.java:95)
>     at java.base/java.lang.reflect.Method.invoke(Method.java:566)
>     at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
>     at 
> java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
>     at 
> java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
>     at java.base/java.lang.Thread.run(Thread.java:834) {code}
> _*OR*_
> Get the error {{Failed to get the primary replica}}
> {code:java}
> java.sql.SQLException: Failed to get the primary replica 
> [tablePartitionId=18_part_1]
> at 
> org.apache.ignite.internal.jdbc.proto.IgniteQueryErrorCode.createJdbcSqlException(IgniteQueryErrorCode.java:57)
> at 
> org.apache.ignite.internal.jdbc.JdbcStatement.execute0(JdbcStatement.java:154)
> at 
> org.apache.ignite.internal.jdbc.JdbcStatement.executeUpdate(JdbcStatement.java:181)
> at 
> org.gridgain.ai3tests.tests.teststeps.JdbcSteps.executeUpdateQuery(JdbcSteps.java:116)
> at 
> org.gridgain.ai3tests.tests.UpdateTests.createAndFillTables(UpdateTests.java:247)
> at io.qameta.allure.Allure.lambda$step$0(Allure.java:113)
> at io.qameta.allure.Allure.lambda$step$1(Allure.java:127)
> at io.qameta.allure.Allure.step(Allure.java:181)
> at io.qameta.allure.Allure.step(Allure.java:125)
> at io.qameta.allure.Allure.step(Allure.java:112)
> at 
> org.gridgain.ai3tests.tests.UpdateTests.updateTableWithConditionThatHasLinkedInnerSubQueries(UpdateTests.java:147)
> at java.base/java.lang.reflect.Method.invoke(Method.java:566)
> at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
> at 
> java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
> at 
> java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
> at java.base/java.lang.Thread.run(Thread.java:834) {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Resolved] (IGNITE-21451) RocksDB: repeat of create table and drop column leads to freeze of client

2024-03-22 Thread Nikita Sivkov (Jira)


 [ 
https://issues.apache.org/jira/browse/IGNITE-21451?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nikita Sivkov resolved IGNITE-21451.

Resolution: Cannot Reproduce

Didn't reproduce on the recent builds.

> RocksDB: repeat of create table and drop column leads to freeze of client
> -
>
> Key: IGNITE-21451
> URL: https://issues.apache.org/jira/browse/IGNITE-21451
> Project: Ignite
>  Issue Type: Bug
>  Components: persistence
>Affects Versions: 3.0.0-beta1
> Environment: 2 nodes cluster running locally.
>Reporter: Igor
>Priority: Major
>  Labels: ignite-3
>
> h3. Steps to reproduce:
> Run the next queries in cycle with 50 repeats in single connection:
> {code:java}
> drop table if exists dropNoMoreIndexedColumn
> create zone if not exists "rocksdb" engine rocksdb
> create table dropNoMoreIndexedColumn(k1 TIMESTAMP not null, k2 INTEGER not 
> null, v0 TINYINT not null, v1 SMALLINT not null, v2 INTEGER not null, v3 
> BIGINT not null, v4 VARCHAR not null, v5 TIMESTAMP not null, primary key (k1, 
> k2)) with PRIMARY_ZONE='rocksdb'
> create index dropNoMoreIndexedColumn_v1idx on dropNoMoreIndexedColumn using 
> TREE (v1)
> drop index dropNoMoreIndexedColumn_v1idx
> alter table dropNoMoreIndexedColumn drop column v1 {code}
> h3. Expected:
> All queries are executed.
> h3. Actual:
> On repeat 31 the client freeze for infinite amount of time.
> h3. Analysis:
> With storages aimem and aipersist the issue is not happen.
> The servers contain repeated error in logs:
> {code:java}
> 2024-02-05 13:47:24:812 +0100 
> [ERROR][%DropColumnsTest_cluster_0%JRaft-FSMCaller-Disruptor-metastorage-_stripe_0-0][WatchProcessor]
>  Error occurred when notifying safe time advanced callback
> java.util.concurrent.CompletionException: 
> java.lang.UnsupportedOperationException: Update log is not supported in 
> RocksDB storage.
>   at 
> java.base/java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:314)
>   at 
> java.base/java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:319)
>   at 
> java.base/java.util.concurrent.CompletableFuture$UniApply.tryFire(CompletableFuture.java:645)
>   at 
> java.base/java.util.concurrent.CompletableFuture$Completion.run(CompletableFuture.java:478)
>   at 
> java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
>   at 
> java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
>   at java.base/java.lang.Thread.run(Thread.java:834)
> Caused by: java.lang.UnsupportedOperationException: Update log is not 
> supported in RocksDB storage.
>   at 
> org.apache.ignite.internal.storage.rocksdb.RocksDbMvPartitionStorage.trimUpdateLog(RocksDbMvPartitionStorage.java:908)
>   at 
> org.apache.ignite.internal.table.distributed.raft.snapshot.outgoing.SnapshotAwarePartitionDataStorage.trimUpdateLog(SnapshotAwarePartitionDataStorage.java:244)
>   at 
> org.apache.ignite.internal.table.distributed.gc.GcUpdateHandler.lambda$vacuumBatch$0(GcUpdateHandler.java:81)
>   at 
> org.apache.ignite.internal.storage.rocksdb.RocksDbMvPartitionStorage.lambda$runConsistently$2(RocksDbMvPartitionStorage.java:228)
>   at 
> org.apache.ignite.internal.storage.rocksdb.RocksDbMvPartitionStorage.busy(RocksDbMvPartitionStorage.java:1431)
>   at 
> org.apache.ignite.internal.storage.rocksdb.RocksDbMvPartitionStorage.runConsistently(RocksDbMvPartitionStorage.java:213)
>   at 
> org.apache.ignite.internal.table.distributed.raft.snapshot.outgoing.SnapshotAwarePartitionDataStorage.runConsistently(SnapshotAwarePartitionDataStorage.java:80)
>   at 
> org.apache.ignite.internal.table.distributed.gc.GcUpdateHandler.vacuumBatch(GcUpdateHandler.java:80)
>   at 
> org.apache.ignite.internal.table.distributed.gc.MvGc.lambda$scheduleGcForStorage$7(MvGc.java:242)
>   at 
> java.base/java.util.concurrent.CompletableFuture$UniApply.tryFire(CompletableFuture.java:642)
>   ... 4 more  {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (IGNITE-21777) 'Failed to get the primary replica' or 'Replication is timed out' or hangs with 'aimem' storage engine

2024-03-19 Thread Nikita Sivkov (Jira)


 [ 
https://issues.apache.org/jira/browse/IGNITE-21777?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nikita Sivkov updated IGNITE-21777:
---
Description: 
*Steps to reproduce:*
 # Create a cluster with 2 nodes.
 # Connect via JDBC.
 # Repeat the following SQL statements in a loop (for example, 100 times):
 ** {{drop table if exists tags}}
 ** {{create zone if not exists "AIMEM" engine aimem}}
 ** {{create table tags(tagId integer not null, tag varchar(100) not null, 
primary key (tagId)) with PRIMARY_ZONE='AIMEM'}}
 ** {{insert into tags(tagId, tag) values (1,'unit'), (2,'integration'), 
(3,'smoke'), (4,'sanity'), (5,'regression')}}

*Expected result:*

No errors or hangs happen.

*Actual result:*

Hangs on {{Create table}} or {{Insert into}} statement.

_*OR*_

Get the error {{Replication is timed out}}
{code:java}
Replication is timed out [replicaGrpId=34_part_16]
java.sql.SQLException: Replication is timed out [replicaGrpId=34_part_16]
    at 
org.apache.ignite.internal.jdbc.proto.IgniteQueryErrorCode.createJdbcSqlException(IgniteQueryErrorCode.java:57)
    at 
org.apache.ignite.internal.jdbc.JdbcStatement.execute0(JdbcStatement.java:154)
    at 
org.apache.ignite.internal.jdbc.JdbcStatement.executeUpdate(JdbcStatement.java:181)
    at 
org.gridgain.ai3tests.tests.teststeps.JdbcSteps.executeUpdateQuery(JdbcSteps.java:116)
    at 
org.gridgain.ai3tests.tests.UpdateTests.wannaCatchTheBug(UpdateTests.java:95)
    at java.base/java.lang.reflect.Method.invoke(Method.java:566)
    at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
    at 
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
    at 
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
    at java.base/java.lang.Thread.run(Thread.java:834) {code}
_*OR*_

Get the error {{Failed to get the primary replica}}
{code:java}
java.sql.SQLException: Failed to get the primary replica 
[tablePartitionId=18_part_1]
at 
org.apache.ignite.internal.jdbc.proto.IgniteQueryErrorCode.createJdbcSqlException(IgniteQueryErrorCode.java:57)
at 
org.apache.ignite.internal.jdbc.JdbcStatement.execute0(JdbcStatement.java:154)
at 
org.apache.ignite.internal.jdbc.JdbcStatement.executeUpdate(JdbcStatement.java:181)
at 
org.gridgain.ai3tests.tests.teststeps.JdbcSteps.executeUpdateQuery(JdbcSteps.java:116)
at 
org.gridgain.ai3tests.tests.UpdateTests.createAndFillTables(UpdateTests.java:247)
at io.qameta.allure.Allure.lambda$step$0(Allure.java:113)
at io.qameta.allure.Allure.lambda$step$1(Allure.java:127)
at io.qameta.allure.Allure.step(Allure.java:181)
at io.qameta.allure.Allure.step(Allure.java:125)
at io.qameta.allure.Allure.step(Allure.java:112)
at 
org.gridgain.ai3tests.tests.UpdateTests.updateTableWithConditionThatHasLinkedInnerSubQueries(UpdateTests.java:147)
at java.base/java.lang.reflect.Method.invoke(Method.java:566)
at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
at 
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
at 
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
at java.base/java.lang.Thread.run(Thread.java:834) {code}

  was:
*Steps to reproduce:* # Create a cluster with 2 nodes.
 # Connect via JDBC.
 # Repeat the following SQL statements in a loop (for example, 100 times):
 ** {{drop table if exists tags}}
 ** {{create zone if not exists "AIMEM" engine aimem}}
 ** {{create table tags(tagId integer not null, tag varchar(100) not null, 
primary key (tagId)) with PRIMARY_ZONE='AIMEM'}}
 ** {{insert into tags(tagId, tag) values (1,'unit'), (2,'integration'), 
(3,'smoke'), (4,'sanity'), (5,'regression')}}

*Expected result:*

No errors or hangs happen.

*Actual result:*

Hangs on {{Create table}} or {{Insert into}} statement.

OR

Get the error {{Replication is timed out}}
{code:java}
Replication is timed out [replicaGrpId=34_part_16]
java.sql.SQLException: Replication is timed out [replicaGrpId=34_part_16]
    at 
org.apache.ignite.internal.jdbc.proto.IgniteQueryErrorCode.createJdbcSqlException(IgniteQueryErrorCode.java:57)
    at 
org.apache.ignite.internal.jdbc.JdbcStatement.execute0(JdbcStatement.java:154)
    at 
org.apache.ignite.internal.jdbc.JdbcStatement.executeUpdate(JdbcStatement.java:181)
    at 
org.gridgain.ai3tests.tests.teststeps.JdbcSteps.executeUpdateQuery(JdbcSteps.java:116)
    at 
org.gridgain.ai3tests.tests.UpdateTests.wannaCatchTheBug(UpdateTests.java:95)
    at java.base/java.lang.reflect.Method.invoke(Method.java:566)
    at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
    at 
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
    at 

[jira] [Updated] (IGNITE-21777) 'Failed to get the primary replica' or 'Replication is timed out' or hangs with 'aimem' storage engine

2024-03-19 Thread Nikita Sivkov (Jira)


 [ 
https://issues.apache.org/jira/browse/IGNITE-21777?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nikita Sivkov updated IGNITE-21777:
---
Summary: 'Failed to get the primary replica' or 'Replication is timed out' 
or hangs with 'aimem' storage engine  (was: 'Replication is timed out' or hangs 
with 'aimem' storage engine)

> 'Failed to get the primary replica' or 'Replication is timed out' or hangs 
> with 'aimem' storage engine
> --
>
> Key: IGNITE-21777
> URL: https://issues.apache.org/jira/browse/IGNITE-21777
> Project: Ignite
>  Issue Type: Bug
>Affects Versions: 3.0.0-beta2
> Environment: Cluster of 2 nodes.
> Storage engine - aimem.
>Reporter: Nikita Sivkov
>Priority: Major
>  Labels: ignite-3
>
> *Steps to reproduce:* # Create a cluster with 2 nodes.
>  # Connect via JDBC.
>  # Repeat the following SQL statements in a loop (for example, 100 times):
>  ** {{drop table if exists tags}}
>  ** {{create zone if not exists "AIMEM" engine aimem}}
>  ** {{create table tags(tagId integer not null, tag varchar(100) not null, 
> primary key (tagId)) with PRIMARY_ZONE='AIMEM'}}
>  ** {{insert into tags(tagId, tag) values (1,'unit'), (2,'integration'), 
> (3,'smoke'), (4,'sanity'), (5,'regression')}}
> *Expected result:*
> No errors or hangs happen.
> *Actual result:*
> Hangs on {{Create table}} or {{Insert into}} statement.
> OR
> Get the error {{Replication is timed out}}
> {code:java}
> Replication is timed out [replicaGrpId=34_part_16]
> java.sql.SQLException: Replication is timed out [replicaGrpId=34_part_16]
>     at 
> org.apache.ignite.internal.jdbc.proto.IgniteQueryErrorCode.createJdbcSqlException(IgniteQueryErrorCode.java:57)
>     at 
> org.apache.ignite.internal.jdbc.JdbcStatement.execute0(JdbcStatement.java:154)
>     at 
> org.apache.ignite.internal.jdbc.JdbcStatement.executeUpdate(JdbcStatement.java:181)
>     at 
> org.gridgain.ai3tests.tests.teststeps.JdbcSteps.executeUpdateQuery(JdbcSteps.java:116)
>     at 
> org.gridgain.ai3tests.tests.UpdateTests.wannaCatchTheBug(UpdateTests.java:95)
>     at java.base/java.lang.reflect.Method.invoke(Method.java:566)
>     at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
>     at 
> java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
>     at 
> java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
>     at java.base/java.lang.Thread.run(Thread.java:834) {code}
> OR
> Get the error {{Failed to get the primary replica}}
> {code:java}
> java.sql.SQLException: Failed to get the primary replica 
> [tablePartitionId=18_part_1]
> at 
> org.apache.ignite.internal.jdbc.proto.IgniteQueryErrorCode.createJdbcSqlException(IgniteQueryErrorCode.java:57)
> at 
> org.apache.ignite.internal.jdbc.JdbcStatement.execute0(JdbcStatement.java:154)
> at 
> org.apache.ignite.internal.jdbc.JdbcStatement.executeUpdate(JdbcStatement.java:181)
> at 
> org.gridgain.ai3tests.tests.teststeps.JdbcSteps.executeUpdateQuery(JdbcSteps.java:116)
> at 
> org.gridgain.ai3tests.tests.UpdateTests.createAndFillTables(UpdateTests.java:247)
> at io.qameta.allure.Allure.lambda$step$0(Allure.java:113)
> at io.qameta.allure.Allure.lambda$step$1(Allure.java:127)
> at io.qameta.allure.Allure.step(Allure.java:181)
> at io.qameta.allure.Allure.step(Allure.java:125)
> at io.qameta.allure.Allure.step(Allure.java:112)
> at 
> org.gridgain.ai3tests.tests.UpdateTests.updateTableWithConditionThatHasLinkedInnerSubQueries(UpdateTests.java:147)
> at java.base/java.lang.reflect.Method.invoke(Method.java:566)
> at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
> at 
> java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
> at 
> java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
> at java.base/java.lang.Thread.run(Thread.java:834) {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (IGNITE-21777) 'Replication is timed out' or hangs with 'aimem' storage engine

2024-03-19 Thread Nikita Sivkov (Jira)


 [ 
https://issues.apache.org/jira/browse/IGNITE-21777?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nikita Sivkov updated IGNITE-21777:
---
Description: 
*Steps to reproduce:* # Create a cluster with 2 nodes.
 # Connect via JDBC.
 # Repeat the following SQL statements in a loop (for example, 100 times):
 ** {{drop table if exists tags}}
 ** {{create zone if not exists "AIMEM" engine aimem}}
 ** {{create table tags(tagId integer not null, tag varchar(100) not null, 
primary key (tagId)) with PRIMARY_ZONE='AIMEM'}}
 ** {{insert into tags(tagId, tag) values (1,'unit'), (2,'integration'), 
(3,'smoke'), (4,'sanity'), (5,'regression')}}

*Expected result:*

No errors or hangs happen.

*Actual result:*

Hangs on {{Create table}} or {{Insert into}} statement.

OR

Get the error {{Replication is timed out}}
{code:java}
Replication is timed out [replicaGrpId=34_part_16]
java.sql.SQLException: Replication is timed out [replicaGrpId=34_part_16]
    at 
org.apache.ignite.internal.jdbc.proto.IgniteQueryErrorCode.createJdbcSqlException(IgniteQueryErrorCode.java:57)
    at 
org.apache.ignite.internal.jdbc.JdbcStatement.execute0(JdbcStatement.java:154)
    at 
org.apache.ignite.internal.jdbc.JdbcStatement.executeUpdate(JdbcStatement.java:181)
    at 
org.gridgain.ai3tests.tests.teststeps.JdbcSteps.executeUpdateQuery(JdbcSteps.java:116)
    at 
org.gridgain.ai3tests.tests.UpdateTests.wannaCatchTheBug(UpdateTests.java:95)
    at java.base/java.lang.reflect.Method.invoke(Method.java:566)
    at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
    at 
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
    at 
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
    at java.base/java.lang.Thread.run(Thread.java:834) {code}
OR

Get the error {{Failed to get the primary replica}}
{code:java}
java.sql.SQLException: Failed to get the primary replica 
[tablePartitionId=18_part_1]
at 
org.apache.ignite.internal.jdbc.proto.IgniteQueryErrorCode.createJdbcSqlException(IgniteQueryErrorCode.java:57)
at 
org.apache.ignite.internal.jdbc.JdbcStatement.execute0(JdbcStatement.java:154)
at 
org.apache.ignite.internal.jdbc.JdbcStatement.executeUpdate(JdbcStatement.java:181)
at 
org.gridgain.ai3tests.tests.teststeps.JdbcSteps.executeUpdateQuery(JdbcSteps.java:116)
at 
org.gridgain.ai3tests.tests.UpdateTests.createAndFillTables(UpdateTests.java:247)
at io.qameta.allure.Allure.lambda$step$0(Allure.java:113)
at io.qameta.allure.Allure.lambda$step$1(Allure.java:127)
at io.qameta.allure.Allure.step(Allure.java:181)
at io.qameta.allure.Allure.step(Allure.java:125)
at io.qameta.allure.Allure.step(Allure.java:112)
at 
org.gridgain.ai3tests.tests.UpdateTests.updateTableWithConditionThatHasLinkedInnerSubQueries(UpdateTests.java:147)
at java.base/java.lang.reflect.Method.invoke(Method.java:566)
at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
at 
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
at 
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
at java.base/java.lang.Thread.run(Thread.java:834) {code}

  was:
*Steps to reproduce:* # Create a cluster with 2 nodes.
 # Connect via JDBC.
 # Repeat the following SQL statements in a loop (for example, 100 times):
 ** {{drop table if exists tags}}
 ** {{create zone if not exists "AIMEM" engine aimem}}
 ** {{create table tags(tagId integer not null, tag varchar(100) not null, 
primary key (tagId)) with PRIMARY_ZONE='AIMEM'}}
 ** {{insert into tags(tagId, tag) values (1,'unit'), (2,'integration'), 
(3,'smoke'), (4,'sanity'), (5,'regression')}}

*Expected result:*

No errors or hangs happen.

*Actual result:*

Hangs on {{Create table}} or {{Insert into}} statement.

OR

Get the error {{Replication is timed out}}
{code:java}
Replication is timed out [replicaGrpId=34_part_16]
java.sql.SQLException: Replication is timed out [replicaGrpId=34_part_16]
    at 
org.apache.ignite.internal.jdbc.proto.IgniteQueryErrorCode.createJdbcSqlException(IgniteQueryErrorCode.java:57)
    at 
org.apache.ignite.internal.jdbc.JdbcStatement.execute0(JdbcStatement.java:154)
    at 
org.apache.ignite.internal.jdbc.JdbcStatement.executeUpdate(JdbcStatement.java:181)
    at 
org.gridgain.ai3tests.tests.teststeps.JdbcSteps.executeUpdateQuery(JdbcSteps.java:116)
    at 
org.gridgain.ai3tests.tests.UpdateTests.wannaCatchTheBug(UpdateTests.java:95)
    at java.base/java.lang.reflect.Method.invoke(Method.java:566)
    at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
    at 
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
    at 

[jira] [Created] (IGNITE-21777) 'Replication is timed out' or hangs with 'aimem' storage engine

2024-03-18 Thread Nikita Sivkov (Jira)
Nikita Sivkov created IGNITE-21777:
--

 Summary: 'Replication is timed out' or hangs with 'aimem' storage 
engine
 Key: IGNITE-21777
 URL: https://issues.apache.org/jira/browse/IGNITE-21777
 Project: Ignite
  Issue Type: Bug
Affects Versions: 3.0.0-beta2
 Environment: Cluster of 2 nodes.

Storage engine - aimem.
Reporter: Nikita Sivkov


*Steps to reproduce:* # Create a cluster with 2 nodes.
 # Connect via JDBC.
 # Repeat the following SQL statements in a loop (for example, 100 times):
 ** {{drop table if exists tags}}
 ** {{create zone if not exists "AIMEM" engine aimem}}
 ** {{create table tags(tagId integer not null, tag varchar(100) not null, 
primary key (tagId)) with PRIMARY_ZONE='AIMEM'}}
 ** {{insert into tags(tagId, tag) values (1,'unit'), (2,'integration'), 
(3,'smoke'), (4,'sanity'), (5,'regression')}}

*Expected result:*

No errors or hangs happen.

*Actual result:*

Hangs on {{Create table}} or {{Insert into}} statement.

OR

Get the error {{Replication is timed out}}
{code:java}
Replication is timed out [replicaGrpId=34_part_16]
java.sql.SQLException: Replication is timed out [replicaGrpId=34_part_16]
    at 
org.apache.ignite.internal.jdbc.proto.IgniteQueryErrorCode.createJdbcSqlException(IgniteQueryErrorCode.java:57)
    at 
org.apache.ignite.internal.jdbc.JdbcStatement.execute0(JdbcStatement.java:154)
    at 
org.apache.ignite.internal.jdbc.JdbcStatement.executeUpdate(JdbcStatement.java:181)
    at 
org.gridgain.ai3tests.tests.teststeps.JdbcSteps.executeUpdateQuery(JdbcSteps.java:116)
    at 
org.gridgain.ai3tests.tests.UpdateTests.wannaCatchTheBug(UpdateTests.java:95)
    at java.base/java.lang.reflect.Method.invoke(Method.java:566)
    at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
    at 
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
    at 
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
    at java.base/java.lang.Thread.run(Thread.java:834) {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Resolved] (IGNITE-21739) JDBC connection to a multi-node cluster doesn't take into account clientConnector.port from each node

2024-03-14 Thread Nikita Sivkov (Jira)


 [ 
https://issues.apache.org/jira/browse/IGNITE-21739?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nikita Sivkov resolved IGNITE-21739.

Resolution: Invalid

Actually, the connection string was like 
{{jdbc:ignite:thin://172.24.1.2,172.24.1.3,172.24.1.4 and the actual issue is 
related to [IGNITE-21577|https://issues.apache.org/jira/browse/IGNITE-21577]}}

> JDBC connection to a multi-node cluster doesn't take into account 
> clientConnector.port from each node
> -
>
> Key: IGNITE-21739
> URL: https://issues.apache.org/jira/browse/IGNITE-21739
> Project: Ignite
>  Issue Type: Bug
>  Components: jdbc
>Affects Versions: 3.0.0-beta2
> Environment: * multi-node cluster
>  * different `{{{}clientConnector.port{}}}` on each cluster node
>Reporter: Nikita Sivkov
>Priority: Major
>  Labels: ignite-3
> Attachments: exception.log
>
>
> *WHEN* you create a multi-node cluster
> *AND* specify different {color:#de350b}{{clientConnector.port}}{color} on 
> each cluster node
> (for example, 
> node1 (172.24.1.2) - {color:#de350b}{{clientConnector.port=10800}}{color}
> node2 (172.24.1.3) - {color:#de350b}{{clientConnector.port=10801}}{color}
> node3 (172.24.1.4) - {color:#de350b}{{clientConnector.port=10802}}{color})
> *AND* connect to cluster like 
> {color:#de350b}{{{}jdbc:ignite:thin://{node1address{{color} (for example, 
> {{{color:#de350b}jdbc:ignite:thin://172.24.1.2{color})}}
> *AND* try to insert a couple of records
> *THEN* you will get an error like
> {code:java}
> Mar 12, 2024 7:37:21 PM org.apache.ignite.internal.logger.IgniteLogger 
> warnWARNING: Failed to establish connection to 172.24.1.3:10800: 
> org.apache.ignite.client.IgniteClientConnectionException: IGN-CLIENT-1 
> TraceId:df21d718-d40c-4506-84e7-6ec141de9ab5 Client failed to connect: 
> Connection refused: no further information: 
> /172.24.1.3:10800java.util.concurrent.CompletionException: 
> org.apache.ignite.client.IgniteClientConnectionException: IGN-CLIENT-1 
> TraceId:df21d718-d40c-4506-84e7-6ec141de9ab5 Client failed to connect: 
> Connection refused: no further information: /172.24.1.3:10800  at 
> java.base/java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:331)
>  at 
> java.base/java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:346)
>at 
> java.base/java.util.concurrent.CompletableFuture$UniCompose.tryFire(CompletableFuture.java:1063)
>  at 
> java.base/java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506)
> at 
> java.base/java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:2088)
>   at 
> org.apache.ignite.internal.client.io.netty.NettyClientConnectionMultiplexer.lambda$openAsync$1(NettyClientConnectionMultiplexer.java:197)
> at 
> io.netty.util.concurrent.DefaultPromise.notifyListener0(DefaultPromise.java:590)
>  at 
> io.netty.util.concurrent.DefaultPromise.notifyListeners0(DefaultPromise.java:583)
> at 
> io.netty.util.concurrent.DefaultPromise.notifyListenersNow(DefaultPromise.java:559)
>   at 
> io.netty.util.concurrent.DefaultPromise.notifyListeners(DefaultPromise.java:492)
>  at 
> io.netty.util.concurrent.DefaultPromise.setValue0(DefaultPromise.java:636)   
> at 
> io.netty.util.concurrent.DefaultPromise.setFailure0(DefaultPromise.java:629) 
> at 
> io.netty.util.concurrent.DefaultPromise.tryFailure(DefaultPromise.java:118)  
> at 
> io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.fulfillConnectPromise(AbstractNioChannel.java:322)
>  at 
> io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.finishConnect(AbstractNioChannel.java:338)
>  at 
> io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:776)  
> at 
> io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:724)
> at 
> io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:650) 
> at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:562) at 
> io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997)
>  at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) 
>at 
> io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
> at java.base/java.lang.Thread.run(Thread.java:834)Caused by: 
> org.apache.ignite.client.IgniteClientConnectionException: IGN-CLIENT-1 
> TraceId:df21d718-d40c-4506-84e7-6ec141de9ab5 Client failed to connect: 
> Connection refused: no further information: /172.24.1.3:10800at 
> org.apache.ignite.internal.client.io.netty.NettyClientConnectionMultiplexer.lambda$openAsync$1(NettyClientConnectionMultiplexer.java:194)
> ... 17 

[jira] [Updated] (IGNITE-21739) JDBC connection to a multi-node cluster doesn't take into account clientConnector.port from each node

2024-03-12 Thread Nikita Sivkov (Jira)


 [ 
https://issues.apache.org/jira/browse/IGNITE-21739?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nikita Sivkov updated IGNITE-21739:
---
Affects Version/s: 3.0.0-beta2

> JDBC connection to a multi-node cluster doesn't take into account 
> clientConnector.port from each node
> -
>
> Key: IGNITE-21739
> URL: https://issues.apache.org/jira/browse/IGNITE-21739
> Project: Ignite
>  Issue Type: Bug
>  Components: jdbc
>Affects Versions: 3.0.0-beta2
> Environment: * multi-node cluster
>  * different `{{{}clientConnector.port{}}}` on each cluster node
>Reporter: Nikita Sivkov
>Priority: Major
>  Labels: ignite-3
> Attachments: exception.log
>
>
> *WHEN* you create a multi-node cluster
> *AND* specify different {color:#de350b}{{clientConnector.port}}{color} on 
> each cluster node
> (for example, 
> node1 (172.24.1.2) - {color:#de350b}{{clientConnector.port=10800}}{color}
> node2 (172.24.1.3) - {color:#de350b}{{clientConnector.port=10801}}{color}
> node3 (172.24.1.4) - {color:#de350b}{{clientConnector.port=10802}}{color})
> *AND* connect to cluster like 
> {color:#de350b}{{{}jdbc:ignite:thin://{node1address{{color} (for example, 
> {{{color:#de350b}jdbc:ignite:thin://172.24.1.2{color})}}
> *AND* try to insert a couple of records
> *THEN* you will get an error like
> {code:java}
> Mar 12, 2024 7:37:21 PM org.apache.ignite.internal.logger.IgniteLogger 
> warnWARNING: Failed to establish connection to 172.24.1.3:10800: 
> org.apache.ignite.client.IgniteClientConnectionException: IGN-CLIENT-1 
> TraceId:df21d718-d40c-4506-84e7-6ec141de9ab5 Client failed to connect: 
> Connection refused: no further information: 
> /172.24.1.3:10800java.util.concurrent.CompletionException: 
> org.apache.ignite.client.IgniteClientConnectionException: IGN-CLIENT-1 
> TraceId:df21d718-d40c-4506-84e7-6ec141de9ab5 Client failed to connect: 
> Connection refused: no further information: /172.24.1.3:10800  at 
> java.base/java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:331)
>  at 
> java.base/java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:346)
>at 
> java.base/java.util.concurrent.CompletableFuture$UniCompose.tryFire(CompletableFuture.java:1063)
>  at 
> java.base/java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506)
> at 
> java.base/java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:2088)
>   at 
> org.apache.ignite.internal.client.io.netty.NettyClientConnectionMultiplexer.lambda$openAsync$1(NettyClientConnectionMultiplexer.java:197)
> at 
> io.netty.util.concurrent.DefaultPromise.notifyListener0(DefaultPromise.java:590)
>  at 
> io.netty.util.concurrent.DefaultPromise.notifyListeners0(DefaultPromise.java:583)
> at 
> io.netty.util.concurrent.DefaultPromise.notifyListenersNow(DefaultPromise.java:559)
>   at 
> io.netty.util.concurrent.DefaultPromise.notifyListeners(DefaultPromise.java:492)
>  at 
> io.netty.util.concurrent.DefaultPromise.setValue0(DefaultPromise.java:636)   
> at 
> io.netty.util.concurrent.DefaultPromise.setFailure0(DefaultPromise.java:629) 
> at 
> io.netty.util.concurrent.DefaultPromise.tryFailure(DefaultPromise.java:118)  
> at 
> io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.fulfillConnectPromise(AbstractNioChannel.java:322)
>  at 
> io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.finishConnect(AbstractNioChannel.java:338)
>  at 
> io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:776)  
> at 
> io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:724)
> at 
> io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:650) 
> at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:562) at 
> io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997)
>  at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) 
>at 
> io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
> at java.base/java.lang.Thread.run(Thread.java:834)Caused by: 
> org.apache.ignite.client.IgniteClientConnectionException: IGN-CLIENT-1 
> TraceId:df21d718-d40c-4506-84e7-6ec141de9ab5 Client failed to connect: 
> Connection refused: no further information: /172.24.1.3:10800at 
> org.apache.ignite.internal.client.io.netty.NettyClientConnectionMultiplexer.lambda$openAsync$1(NettyClientConnectionMultiplexer.java:194)
> ... 17 moreCaused by: 
> io.netty.channel.AbstractChannel$AnnotatedConnectException: Connection 
> refused: no further information: /172.24.1.3:10800Caused by: 
> java.net.ConnectException: 

[jira] [Updated] (IGNITE-21739) JDBC connection to a multi-node cluster doesn't take into account clientConnector.port from each node

2024-03-12 Thread Nikita Sivkov (Jira)


 [ 
https://issues.apache.org/jira/browse/IGNITE-21739?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nikita Sivkov updated IGNITE-21739:
---
Labels: ignite-3  (was: )

> JDBC connection to a multi-node cluster doesn't take into account 
> clientConnector.port from each node
> -
>
> Key: IGNITE-21739
> URL: https://issues.apache.org/jira/browse/IGNITE-21739
> Project: Ignite
>  Issue Type: Bug
>  Components: jdbc
> Environment: * multi-node cluster
>  * different `{{{}clientConnector.port{}}}` on each cluster node
>Reporter: Nikita Sivkov
>Priority: Major
>  Labels: ignite-3
> Attachments: exception.log
>
>
> *WHEN* you create a multi-node cluster
> *AND* specify different {color:#de350b}{{clientConnector.port}}{color} on 
> each cluster node
> (for example, 
> node1 (172.24.1.2) - {color:#de350b}{{clientConnector.port=10800}}{color}
> node2 (172.24.1.3) - {color:#de350b}{{clientConnector.port=10801}}{color}
> node3 (172.24.1.4) - {color:#de350b}{{clientConnector.port=10802}}{color})
> *AND* connect to cluster like 
> {color:#de350b}{{{}jdbc:ignite:thin://{node1address{{color} (for example, 
> {{{color:#de350b}jdbc:ignite:thin://172.24.1.2{color})}}
> *AND* try to insert a couple of records
> *THEN* you will get an error like
> {code:java}
> Mar 12, 2024 7:37:21 PM org.apache.ignite.internal.logger.IgniteLogger 
> warnWARNING: Failed to establish connection to 172.24.1.3:10800: 
> org.apache.ignite.client.IgniteClientConnectionException: IGN-CLIENT-1 
> TraceId:df21d718-d40c-4506-84e7-6ec141de9ab5 Client failed to connect: 
> Connection refused: no further information: 
> /172.24.1.3:10800java.util.concurrent.CompletionException: 
> org.apache.ignite.client.IgniteClientConnectionException: IGN-CLIENT-1 
> TraceId:df21d718-d40c-4506-84e7-6ec141de9ab5 Client failed to connect: 
> Connection refused: no further information: /172.24.1.3:10800  at 
> java.base/java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:331)
>  at 
> java.base/java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:346)
>at 
> java.base/java.util.concurrent.CompletableFuture$UniCompose.tryFire(CompletableFuture.java:1063)
>  at 
> java.base/java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506)
> at 
> java.base/java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:2088)
>   at 
> org.apache.ignite.internal.client.io.netty.NettyClientConnectionMultiplexer.lambda$openAsync$1(NettyClientConnectionMultiplexer.java:197)
> at 
> io.netty.util.concurrent.DefaultPromise.notifyListener0(DefaultPromise.java:590)
>  at 
> io.netty.util.concurrent.DefaultPromise.notifyListeners0(DefaultPromise.java:583)
> at 
> io.netty.util.concurrent.DefaultPromise.notifyListenersNow(DefaultPromise.java:559)
>   at 
> io.netty.util.concurrent.DefaultPromise.notifyListeners(DefaultPromise.java:492)
>  at 
> io.netty.util.concurrent.DefaultPromise.setValue0(DefaultPromise.java:636)   
> at 
> io.netty.util.concurrent.DefaultPromise.setFailure0(DefaultPromise.java:629) 
> at 
> io.netty.util.concurrent.DefaultPromise.tryFailure(DefaultPromise.java:118)  
> at 
> io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.fulfillConnectPromise(AbstractNioChannel.java:322)
>  at 
> io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.finishConnect(AbstractNioChannel.java:338)
>  at 
> io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:776)  
> at 
> io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:724)
> at 
> io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:650) 
> at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:562) at 
> io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997)
>  at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) 
>at 
> io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
> at java.base/java.lang.Thread.run(Thread.java:834)Caused by: 
> org.apache.ignite.client.IgniteClientConnectionException: IGN-CLIENT-1 
> TraceId:df21d718-d40c-4506-84e7-6ec141de9ab5 Client failed to connect: 
> Connection refused: no further information: /172.24.1.3:10800at 
> org.apache.ignite.internal.client.io.netty.NettyClientConnectionMultiplexer.lambda$openAsync$1(NettyClientConnectionMultiplexer.java:194)
> ... 17 moreCaused by: 
> io.netty.channel.AbstractChannel$AnnotatedConnectException: Connection 
> refused: no further information: /172.24.1.3:10800Caused by: 
> java.net.ConnectException: Connection refused: no further information at 
> 

[jira] [Updated] (IGNITE-21739) JDBC connection to a multi-node cluster doesn't take into account clientConnector.port from each node

2024-03-12 Thread Nikita Sivkov (Jira)


 [ 
https://issues.apache.org/jira/browse/IGNITE-21739?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nikita Sivkov updated IGNITE-21739:
---
Attachment: exception.log

> JDBC connection to a multi-node cluster doesn't take into account 
> clientConnector.port from each node
> -
>
> Key: IGNITE-21739
> URL: https://issues.apache.org/jira/browse/IGNITE-21739
> Project: Ignite
>  Issue Type: Bug
>  Components: jdbc
> Environment: * multi-node cluster
>  * different `{{{}clientConnector.port{}}}` on each cluster node
>Reporter: Nikita Sivkov
>Priority: Major
> Attachments: exception.log
>
>
> *WHEN* you create a multi-node cluster
> *AND* specify different {color:#de350b}{{clientConnector.port}}{color} on 
> each cluster node
> (for example, 
> node1 (172.24.1.2) - {color:#de350b}{{clientConnector.port=10800}}{color}
> node2 (172.24.1.3) - {color:#de350b}{{clientConnector.port=10801}}{color}
> node3 (172.24.1.4) - {color:#de350b}{{clientConnector.port=10802}}{color})
> *AND* connect to cluster like 
> {color:#de350b}{{{}jdbc:ignite:thin://{node1address{{color} (for example, 
> {{{color:#de350b}jdbc:ignite:thin://172.24.1.2{color})}}
> *AND* try to insert a couple of records
> *THEN* you will get an error like
> {code:java}
> Mar 12, 2024 7:37:21 PM org.apache.ignite.internal.logger.IgniteLogger 
> warnWARNING: Failed to establish connection to 172.24.1.3:10800: 
> org.apache.ignite.client.IgniteClientConnectionException: IGN-CLIENT-1 
> TraceId:df21d718-d40c-4506-84e7-6ec141de9ab5 Client failed to connect: 
> Connection refused: no further information: 
> /172.24.1.3:10800java.util.concurrent.CompletionException: 
> org.apache.ignite.client.IgniteClientConnectionException: IGN-CLIENT-1 
> TraceId:df21d718-d40c-4506-84e7-6ec141de9ab5 Client failed to connect: 
> Connection refused: no further information: /172.24.1.3:10800  at 
> java.base/java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:331)
>  at 
> java.base/java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:346)
>at 
> java.base/java.util.concurrent.CompletableFuture$UniCompose.tryFire(CompletableFuture.java:1063)
>  at 
> java.base/java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506)
> at 
> java.base/java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:2088)
>   at 
> org.apache.ignite.internal.client.io.netty.NettyClientConnectionMultiplexer.lambda$openAsync$1(NettyClientConnectionMultiplexer.java:197)
> at 
> io.netty.util.concurrent.DefaultPromise.notifyListener0(DefaultPromise.java:590)
>  at 
> io.netty.util.concurrent.DefaultPromise.notifyListeners0(DefaultPromise.java:583)
> at 
> io.netty.util.concurrent.DefaultPromise.notifyListenersNow(DefaultPromise.java:559)
>   at 
> io.netty.util.concurrent.DefaultPromise.notifyListeners(DefaultPromise.java:492)
>  at 
> io.netty.util.concurrent.DefaultPromise.setValue0(DefaultPromise.java:636)   
> at 
> io.netty.util.concurrent.DefaultPromise.setFailure0(DefaultPromise.java:629) 
> at 
> io.netty.util.concurrent.DefaultPromise.tryFailure(DefaultPromise.java:118)  
> at 
> io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.fulfillConnectPromise(AbstractNioChannel.java:322)
>  at 
> io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.finishConnect(AbstractNioChannel.java:338)
>  at 
> io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:776)  
> at 
> io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:724)
> at 
> io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:650) 
> at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:562) at 
> io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997)
>  at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) 
>at 
> io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
> at java.base/java.lang.Thread.run(Thread.java:834)Caused by: 
> org.apache.ignite.client.IgniteClientConnectionException: IGN-CLIENT-1 
> TraceId:df21d718-d40c-4506-84e7-6ec141de9ab5 Client failed to connect: 
> Connection refused: no further information: /172.24.1.3:10800at 
> org.apache.ignite.internal.client.io.netty.NettyClientConnectionMultiplexer.lambda$openAsync$1(NettyClientConnectionMultiplexer.java:194)
> ... 17 moreCaused by: 
> io.netty.channel.AbstractChannel$AnnotatedConnectException: Connection 
> refused: no further information: /172.24.1.3:10800Caused by: 
> java.net.ConnectException: Connection refused: no further information at 
> 

[jira] [Updated] (IGNITE-21739) JDBC connection to a multi-node cluster doesn't take into account clientConnector.port from each node

2024-03-12 Thread Nikita Sivkov (Jira)


 [ 
https://issues.apache.org/jira/browse/IGNITE-21739?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nikita Sivkov updated IGNITE-21739:
---
Description: 
*WHEN* you create a multi-node cluster
*AND* specify different {color:#de350b}{{clientConnector.port}}{color} on each 
cluster node
(for example, 
node1 (172.24.1.2) - {color:#de350b}{{clientConnector.port=10800}}{color}
node2 (172.24.1.3) - {color:#de350b}{{clientConnector.port=10801}}{color}
node3 (172.24.1.4) - {color:#de350b}{{clientConnector.port=10802}}{color})
*AND* connect to cluster like 
{color:#de350b}{{{}jdbc:ignite:thin://{node1address{{color} (for example, 
{{{color:#de350b}jdbc:ignite:thin://172.24.1.2{color})}}
*AND* try to insert a couple of records
*THEN* you will get an error like
{code:java}
Mar 12, 2024 7:37:21 PM org.apache.ignite.internal.logger.IgniteLogger 
warnWARNING: Failed to establish connection to 172.24.1.3:10800: 
org.apache.ignite.client.IgniteClientConnectionException: IGN-CLIENT-1 
TraceId:df21d718-d40c-4506-84e7-6ec141de9ab5 Client failed to connect: 
Connection refused: no further information: 
/172.24.1.3:10800java.util.concurrent.CompletionException: 
org.apache.ignite.client.IgniteClientConnectionException: IGN-CLIENT-1 
TraceId:df21d718-d40c-4506-84e7-6ec141de9ab5 Client failed to connect: 
Connection refused: no further information: /172.24.1.3:10800at 
java.base/java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:331)
 at 
java.base/java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:346)
   at 
java.base/java.util.concurrent.CompletableFuture$UniCompose.tryFire(CompletableFuture.java:1063)
 at 
java.base/java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506)
at 
java.base/java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:2088)
  at 
org.apache.ignite.internal.client.io.netty.NettyClientConnectionMultiplexer.lambda$openAsync$1(NettyClientConnectionMultiplexer.java:197)
at 
io.netty.util.concurrent.DefaultPromise.notifyListener0(DefaultPromise.java:590)
 at 
io.netty.util.concurrent.DefaultPromise.notifyListeners0(DefaultPromise.java:583)
at 
io.netty.util.concurrent.DefaultPromise.notifyListenersNow(DefaultPromise.java:559)
  at 
io.netty.util.concurrent.DefaultPromise.notifyListeners(DefaultPromise.java:492)
 at 
io.netty.util.concurrent.DefaultPromise.setValue0(DefaultPromise.java:636)   at 
io.netty.util.concurrent.DefaultPromise.setFailure0(DefaultPromise.java:629) at 
io.netty.util.concurrent.DefaultPromise.tryFailure(DefaultPromise.java:118)  at 
io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.fulfillConnectPromise(AbstractNioChannel.java:322)
 at 
io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.finishConnect(AbstractNioChannel.java:338)
 at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:776) 
 at 
io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:724)
at 
io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:650) at 
io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:562) at 
io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997)
 at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)   
 at 
io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
at java.base/java.lang.Thread.run(Thread.java:834)Caused by: 
org.apache.ignite.client.IgniteClientConnectionException: IGN-CLIENT-1 
TraceId:df21d718-d40c-4506-84e7-6ec141de9ab5 Client failed to connect: 
Connection refused: no further information: /172.24.1.3:10800at 
org.apache.ignite.internal.client.io.netty.NettyClientConnectionMultiplexer.lambda$openAsync$1(NettyClientConnectionMultiplexer.java:194)
... 17 moreCaused by: 
io.netty.channel.AbstractChannel$AnnotatedConnectException: Connection refused: 
no further information: /172.24.1.3:10800Caused by: java.net.ConnectException: 
Connection refused: no further information at 
java.base/sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)   at 
java.base/sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:779)
 at 
io.netty.channel.socket.nio.NioSocketChannel.doFinishConnect(NioSocketChannel.java:337)
  at 
io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.finishConnect(AbstractNioChannel.java:335)
 at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:776) 
 at 
io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:724)
at 
io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:650) at 
io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:562) at 
io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997)
 at 

[jira] [Updated] (IGNITE-21739) JDBC connection to a multi-node cluster doesn't take into account clientConnector.port from each node

2024-03-12 Thread Nikita Sivkov (Jira)


 [ 
https://issues.apache.org/jira/browse/IGNITE-21739?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nikita Sivkov updated IGNITE-21739:
---
Description: 
Mar 12, 2024 7:37:21 PM org.apache.ignite.internal.logger.IgniteLogger warn
WARNING: Failed to establish connection to 172.24.1.3:10800: 
org.apache.ignite.client.IgniteClientConnectionException: IGN-CLIENT-1 
TraceId:df21d718-d40c-4506-84e7-6ec141de9ab5 Client failed to connect: 
Connection refused: no further information: /172.24.1.3:10800
java.util.concurrent.CompletionException: 
org.apache.ignite.client.IgniteClientConnectionException: IGN-CLIENT-1 
TraceId:df21d718-d40c-4506-84e7-6ec141de9ab5 Client failed to connect: 
Connection refused: no further information: /172.24.1.3:10800
at 
java.base/java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:331)
at 
java.base/java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:346)
at 
java.base/java.util.concurrent.CompletableFuture$UniCompose.tryFire(CompletableFuture.java:1063)
at 
java.base/java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506)
at 
java.base/java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:2088)
at 
org.apache.ignite.internal.client.io.netty.NettyClientConnectionMultiplexer.lambda$openAsync$1(NettyClientConnectionMultiplexer.java:197)
at 
io.netty.util.concurrent.DefaultPromise.notifyListener0(DefaultPromise.java:590)
at 
io.netty.util.concurrent.DefaultPromise.notifyListeners0(DefaultPromise.java:583)
at 
io.netty.util.concurrent.DefaultPromise.notifyListenersNow(DefaultPromise.java:559)
at 
io.netty.util.concurrent.DefaultPromise.notifyListeners(DefaultPromise.java:492)
at io.netty.util.concurrent.DefaultPromise.setValue0(DefaultPromise.java:636)
at io.netty.util.concurrent.DefaultPromise.setFailure0(DefaultPromise.java:629)
at io.netty.util.concurrent.DefaultPromise.tryFailure(DefaultPromise.java:118)
at 
io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.fulfillConnectPromise(AbstractNioChannel.java:322)
at 
io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.finishConnect(AbstractNioChannel.java:338)
at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:776)
at 
io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:724)
at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:650)
at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:562)
at 
io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997)
at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)
at 
io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
at java.base/java.lang.Thread.run(Thread.java:834)
Caused by: org.apache.ignite.client.IgniteClientConnectionException: 
IGN-CLIENT-1 TraceId:df21d718-d40c-4506-84e7-6ec141de9ab5 Client failed to 
connect: Connection refused: no further information: /172.24.1.3:10800
at 
org.apache.ignite.internal.client.io.netty.NettyClientConnectionMultiplexer.lambda$openAsync$1(NettyClientConnectionMultiplexer.java:194)
... 17 more
Caused by: io.netty.channel.AbstractChannel$AnnotatedConnectException: 
Connection refused: no further information: /172.24.1.3:10800
Caused by: java.net.ConnectException: Connection refused: no further information
at java.base/sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
at 
java.base/sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:779)
at 
io.netty.channel.socket.nio.NioSocketChannel.doFinishConnect(NioSocketChannel.java:337)
at 
io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.finishConnect(AbstractNioChannel.java:335)
at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:776)
at 
io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:724)
at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:650)
at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:562)
at 
io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997)
at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)
at 
io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
at java.base/java.lang.Thread.run(Thread.java:834)
 
Mar 12, 2024 7:37:21 PM org.apache.ignite.internal.logger.IgniteLogger warn
WARNING: Failed to establish connection to 172.24.1.4:10800: 
org.apache.ignite.client.IgniteClientConnectionException: IGN-CLIENT-1 
TraceId:d7d08ee0-64f6-45ea-872f-1c2adb8d128a Client failed to connect: 
Connection refused: no further information: /172.24.1.4:10800
java.util.concurrent.CompletionException: 
org.apache.ignite.client.IgniteClientConnectionException: IGN-CLIENT-1 
TraceId:d7d08ee0-64f6-45ea-872f-1c2adb8d128a Client failed to connect: 
Connection refused: 

[jira] [Updated] (IGNITE-21739) JDBC connection to a multi-node cluster doesn't take into account clientConnector.port from each node

2024-03-12 Thread Nikita Sivkov (Jira)


 [ 
https://issues.apache.org/jira/browse/IGNITE-21739?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nikita Sivkov updated IGNITE-21739:
---
Description: 
*WHEN* you create a multi-node cluster
*AND* specify different {{clientConnector.port}} on each cluster node
(for example, 
node1 - 172.24.1.2 - {{clientConnector.port=10800}}
node2 - 172.24.1.3 - {{clientConnector.port=10801}}
node3 - 172.24.1.4 - {{{}clientConnector.port=10802{}}})
*AND* connect to cluster like {{{}jdbc:ignite:thin://{node1address{ (for 
example, {{jdbc:ignite:thin://172.24.1.2)}}
*AND* try to insert a couple of records
*THEN* you will get an error like

 
{code:java}
WARNING: Failed to establish connection to 172.24.1.3:10800: 
org.apache.ignite.client.IgniteClientConnectionException: IGN-CLIENT-1 
TraceId:df21d718-d40c-4506-84e7-6ec141de9ab5 Client failed to connect: 
Connection refused: no further information: /172.24.1.3:10800
java.util.concurrent.CompletionException: 
org.apache.ignite.client.IgniteClientConnectionException: IGN-CLIENT-1 
TraceId:df21d718-d40c-4506-84e7-6ec141de9ab5 Client failed to connect: 
Connection refused: no further information: /172.24.1.3:10800
    at 
java.base/java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:331)
    at 
java.base/java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:346)
    at 
java.base/java.util.concurrent.CompletableFuture$UniCompose.tryFire(CompletableFuture.java:1063)
    at 
java.base/java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506)
    at 
java.base/java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:2088)
    at 
org.apache.ignite.internal.client.io.netty.NettyClientConnectionMultiplexer.lambda$openAsync$1(NettyClientConnectionMultiplexer.java:197)
    at 
io.netty.util.concurrent.DefaultPromise.notifyListener0(DefaultPromise.java:590)
    at 
io.netty.util.concurrent.DefaultPromise.notifyListeners0(DefaultPromise.java:583)
    at 
io.netty.util.concurrent.DefaultPromise.notifyListenersNow(DefaultPromise.java:559)
    at 
io.netty.util.concurrent.DefaultPromise.notifyListeners(DefaultPromise.java:492)
    at 
io.netty.util.concurrent.DefaultPromise.setValue0(DefaultPromise.java:636)
    at 
io.netty.util.concurrent.DefaultPromise.setFailure0(DefaultPromise.java:629)
    at 
io.netty.util.concurrent.DefaultPromise.tryFailure(DefaultPromise.java:118)
    at 
io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.fulfillConnectPromise(AbstractNioChannel.java:322)
    at 
io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.finishConnect(AbstractNioChannel.java:338)
    at 
io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:776)
    at 
io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:724)
    at 
io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:650)
    at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:562)
    at 
io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997)
    at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)
    at 
io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
    at java.base/java.lang.Thread.run(Thread.java:834)
Caused by: org.apache.ignite.client.IgniteClientConnectionException: 
IGN-CLIENT-1 TraceId:df21d718-d40c-4506-84e7-6ec141de9ab5 Client failed to 
connect: Connection refused: no further information: /172.24.1.3:10800
    at 
org.apache.ignite.internal.client.io.netty.NettyClientConnectionMultiplexer.lambda$openAsync$1(NettyClientConnectionMultiplexer.java:194)
    ... 17 more
Caused by: io.netty.channel.AbstractChannel$AnnotatedConnectException: 
Connection refused: no further information: /172.24.1.3:10800
Caused by: java.net.ConnectException: Connection refused: no further information
    at java.base/sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
    at 
java.base/sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:779)
    at 
io.netty.channel.socket.nio.NioSocketChannel.doFinishConnect(NioSocketChannel.java:337)
    at 
io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.finishConnect(AbstractNioChannel.java:335)
    at 
io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:776)
    at 
io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:724)
    at 
io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:650)
    at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:562)
    at 
io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997)
    at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)
    at 
io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
    at 

[jira] [Updated] (IGNITE-21739) JDBC connection to a multi-node cluster doesn't take into account clientConnector.port from each node

2024-03-12 Thread Nikita Sivkov (Jira)


 [ 
https://issues.apache.org/jira/browse/IGNITE-21739?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nikita Sivkov updated IGNITE-21739:
---
Description: 
*WHEN* you create a multi-node cluster
*AND* specify different {{clientConnector.port}} on each cluster node
(for example, 
node1 - {{clientConnector.port=10800}}
node2 - {{clientConnector.port=10801}}
node3 - {{{}clientConnector.port=10802{}}})
*AND* connect to cluster like {{jdbc:ignite:thin://\{node1address}}}
*AND* try to insert a couple of records
*THEN* you will get an error like

 
{code:java}
WARNING: Failed to establish connection to 172.24.1.3:10800: 
org.apache.ignite.client.IgniteClientConnectionException: IGN-CLIENT-1 
TraceId:df21d718-d40c-4506-84e7-6ec141de9ab5 Client failed to connect: 
Connection refused: no further information: /172.24.1.3:10800
java.util.concurrent.CompletionException: 
org.apache.ignite.client.IgniteClientConnectionException: IGN-CLIENT-1 
TraceId:df21d718-d40c-4506-84e7-6ec141de9ab5 Client failed to connect: 
Connection refused: no further information: /172.24.1.3:10800
    at 
java.base/java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:331)
    at 
java.base/java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:346)
    at 
java.base/java.util.concurrent.CompletableFuture$UniCompose.tryFire(CompletableFuture.java:1063)
    at 
java.base/java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506)
    at 
java.base/java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:2088)
    at 
org.apache.ignite.internal.client.io.netty.NettyClientConnectionMultiplexer.lambda$openAsync$1(NettyClientConnectionMultiplexer.java:197)
    at 
io.netty.util.concurrent.DefaultPromise.notifyListener0(DefaultPromise.java:590)
    at 
io.netty.util.concurrent.DefaultPromise.notifyListeners0(DefaultPromise.java:583)
    at 
io.netty.util.concurrent.DefaultPromise.notifyListenersNow(DefaultPromise.java:559)
    at 
io.netty.util.concurrent.DefaultPromise.notifyListeners(DefaultPromise.java:492)
    at 
io.netty.util.concurrent.DefaultPromise.setValue0(DefaultPromise.java:636)
    at 
io.netty.util.concurrent.DefaultPromise.setFailure0(DefaultPromise.java:629)
    at 
io.netty.util.concurrent.DefaultPromise.tryFailure(DefaultPromise.java:118)
    at 
io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.fulfillConnectPromise(AbstractNioChannel.java:322)
    at 
io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.finishConnect(AbstractNioChannel.java:338)
    at 
io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:776)
    at 
io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:724)
    at 
io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:650)
    at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:562)
    at 
io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997)
    at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)
    at 
io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
    at java.base/java.lang.Thread.run(Thread.java:834)
Caused by: org.apache.ignite.client.IgniteClientConnectionException: 
IGN-CLIENT-1 TraceId:df21d718-d40c-4506-84e7-6ec141de9ab5 Client failed to 
connect: Connection refused: no further information: /172.24.1.3:10800
    at 
org.apache.ignite.internal.client.io.netty.NettyClientConnectionMultiplexer.lambda$openAsync$1(NettyClientConnectionMultiplexer.java:194)
    ... 17 more
Caused by: io.netty.channel.AbstractChannel$AnnotatedConnectException: 
Connection refused: no further information: /172.24.1.3:10800
Caused by: java.net.ConnectException: Connection refused: no further information
    at java.base/sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
    at 
java.base/sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:779)
    at 
io.netty.channel.socket.nio.NioSocketChannel.doFinishConnect(NioSocketChannel.java:337)
    at 
io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.finishConnect(AbstractNioChannel.java:335)
    at 
io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:776)
    at 
io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:724)
    at 
io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:650)
    at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:562)
    at 
io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997)
    at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)
    at 
io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
    at java.base/java.lang.Thread.run(Thread.java:834) {code}
 

  was:
*Issue:*

WHEN you create a multi-node cluster
AND 

[jira] [Created] (IGNITE-21739) JDBC connection to a multi-node cluster doesn't take into account clientConnector.port from each node

2024-03-12 Thread Nikita Sivkov (Jira)
Nikita Sivkov created IGNITE-21739:
--

 Summary: JDBC connection to a multi-node cluster doesn't take into 
account clientConnector.port from each node
 Key: IGNITE-21739
 URL: https://issues.apache.org/jira/browse/IGNITE-21739
 Project: Ignite
  Issue Type: Bug
  Components: jdbc
 Environment: * multi-node cluster
 * different `{{{}clientConnector.port{}}}` on each cluster node
Reporter: Nikita Sivkov


*Issue:*

WHEN you create a multi-node cluster
AND specify different `{{{}clientConnector.port{}}}` on each cluster node
(for example, 
node1 - `{{{}clientConnector.port=10800{}}}`
node2 - `{{{}clientConnector.port=10801{}}}`
node3 - `{{{}clientConnector.port=10802{}}}`)
AND connect to cluster like `{{{}jdbc:ignite:thin://\{node1address}{}}}`
AND try to insert a couple of records
THEN you will get an error like

```

WARNING: Failed to establish connection to 172.24.1.3:10800: 
org.apache.ignite.client.IgniteClientConnectionException: IGN-CLIENT-1 
TraceId:df21d718-d40c-4506-84e7-6ec141de9ab5 Client failed to connect: 
Connection refused: no further information: /172.24.1.3:10800
java.util.concurrent.CompletionException: 
org.apache.ignite.client.IgniteClientConnectionException: IGN-CLIENT-1 
TraceId:df21d718-d40c-4506-84e7-6ec141de9ab5 Client failed to connect: 
Connection refused: no further information: /172.24.1.3:10800
    at 
java.base/java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:331)
    at 
java.base/java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:346)
    at 
java.base/java.util.concurrent.CompletableFuture$UniCompose.tryFire(CompletableFuture.java:1063)
    at 
java.base/java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506)
    at 
java.base/java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:2088)
    at 
org.apache.ignite.internal.client.io.netty.NettyClientConnectionMultiplexer.lambda$openAsync$1(NettyClientConnectionMultiplexer.java:197)
    at 
io.netty.util.concurrent.DefaultPromise.notifyListener0(DefaultPromise.java:590)
    at 
io.netty.util.concurrent.DefaultPromise.notifyListeners0(DefaultPromise.java:583)
    at 
io.netty.util.concurrent.DefaultPromise.notifyListenersNow(DefaultPromise.java:559)
    at 
io.netty.util.concurrent.DefaultPromise.notifyListeners(DefaultPromise.java:492)
    at 
io.netty.util.concurrent.DefaultPromise.setValue0(DefaultPromise.java:636)
    at 
io.netty.util.concurrent.DefaultPromise.setFailure0(DefaultPromise.java:629)
    at 
io.netty.util.concurrent.DefaultPromise.tryFailure(DefaultPromise.java:118)
    at 
io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.fulfillConnectPromise(AbstractNioChannel.java:322)
    at 
io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.finishConnect(AbstractNioChannel.java:338)
    at 
io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:776)
    at 
io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:724)
    at 
io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:650)
    at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:562)
    at 
io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997)
    at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)
    at 
io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
    at java.base/java.lang.Thread.run(Thread.java:834)
Caused by: org.apache.ignite.client.IgniteClientConnectionException: 
IGN-CLIENT-1 TraceId:df21d718-d40c-4506-84e7-6ec141de9ab5 Client failed to 
connect: Connection refused: no further information: /172.24.1.3:10800
    at 
org.apache.ignite.internal.client.io.netty.NettyClientConnectionMultiplexer.lambda$openAsync$1(NettyClientConnectionMultiplexer.java:194)
    ... 17 more
Caused by: io.netty.channel.AbstractChannel$AnnotatedConnectException: 
Connection refused: no further information: /172.24.1.3:10800
Caused by: java.net.ConnectException: Connection refused: no further information
    at java.base/sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
    at 
java.base/sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:779)
    at 
io.netty.channel.socket.nio.NioSocketChannel.doFinishConnect(NioSocketChannel.java:337)
    at 
io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.finishConnect(AbstractNioChannel.java:335)
    at 
io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:776)
    at 
io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:724)
    at 
io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:650)
    at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:562)
    at 

[jira] [Updated] (IGNITE-20154) Line separator in ODBC errors text

2023-08-03 Thread Nikita Sivkov (Jira)


 [ 
https://issues.apache.org/jira/browse/IGNITE-20154?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nikita Sivkov updated IGNITE-20154:
---
Description: 
*Issue:*
Getting line separator symbols "\r\n" in OBDC error text.

*For example:*
{noformat}
ALTER TABLE PUBLIC.CAR DROP COLUMN IF EXISTS NAME
('HY000', '[HY000] org.apache.ignite.sql.SqlException: Failed to parse query: 
Encountered "" at line 1, column 36.\r\nWas expecting one of:\r\n     (262147, 
3e945522-7d71-48c0-a8cf-675bbc078bb0) (0) (SQLExecDirectW)'){noformat}
*Reproducer:*
 # Start Ignite node.
 # Execute command:
_python3 odbc_client.py -o smoke0.sql.actual localhost:10800 smoke0.sql_

*Commit id:*
2655e406b06a2605c2d5ad9402e06d81c1a168ef

 

 

  was:
*Issue:*
Getting line separator symbols "\r\n" in OBDC error text.
**

*For example:*
{noformat}

{noformat}
*ALTER TABLE PUBLIC.CAR DROP COLUMN IF EXISTS NAME
('HY000', '[HY000] org.apache.ignite.sql.SqlException: Failed to parse query: 
Encountered "" at line 1, column 36.\r\nWas expecting one of:\r\n     (262147, 
3e945522-7d71-48c0-a8cf-675bbc078bb0) (0) (SQLExecDirectW)')*

 

*Reproducer:*
 # Start Ignite node.
 # Execute command:
`python3 odbc_client.py -o smoke0.sql.actual localhost:10800 smoke0.sql`

*Commit id:*
2655e406b06a2605c2d5ad9402e06d81c1a168ef

 

 


> Line separator in ODBC errors text
> --
>
> Key: IGNITE-20154
> URL: https://issues.apache.org/jira/browse/IGNITE-20154
> Project: Ignite
>  Issue Type: Bug
>  Components: odbc
>Affects Versions: 3.0
>Reporter: Nikita Sivkov
>Priority: Major
>  Labels: ignite-3, odbc
> Attachments: odbc_client.py, smoke0.sql
>
>
> *Issue:*
> Getting line separator symbols "\r\n" in OBDC error text.
> *For example:*
> {noformat}
> ALTER TABLE PUBLIC.CAR DROP COLUMN IF EXISTS NAME
> ('HY000', '[HY000] org.apache.ignite.sql.SqlException: Failed to parse query: 
> Encountered "" at line 1, column 36.\r\nWas expecting one of:\r\n     
> (262147, 3e945522-7d71-48c0-a8cf-675bbc078bb0) (0) 
> (SQLExecDirectW)'){noformat}
> *Reproducer:*
>  # Start Ignite node.
>  # Execute command:
> _python3 odbc_client.py -o smoke0.sql.actual localhost:10800 smoke0.sql_
> *Commit id:*
> 2655e406b06a2605c2d5ad9402e06d81c1a168ef
>  
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (IGNITE-20154) Line separator in ODBC errors text

2023-08-03 Thread Nikita Sivkov (Jira)
Nikita Sivkov created IGNITE-20154:
--

 Summary: Line separator in ODBC errors text
 Key: IGNITE-20154
 URL: https://issues.apache.org/jira/browse/IGNITE-20154
 Project: Ignite
  Issue Type: Bug
  Components: odbc
Affects Versions: 3.0
Reporter: Nikita Sivkov
 Attachments: odbc_client.py, smoke0.sql

*Issue:*
Getting line separator symbols "\r\n" in OBDC error text.
**

*For example:*
{noformat}

{noformat}
*ALTER TABLE PUBLIC.CAR DROP COLUMN IF EXISTS NAME
('HY000', '[HY000] org.apache.ignite.sql.SqlException: Failed to parse query: 
Encountered "" at line 1, column 36.\r\nWas expecting one of:\r\n     (262147, 
3e945522-7d71-48c0-a8cf-675bbc078bb0) (0) (SQLExecDirectW)')*

 

*Reproducer:*
 # Start Ignite node.
 # Execute command:
`python3 odbc_client.py -o smoke0.sql.actual localhost:10800 smoke0.sql`

*Commit id:*
2655e406b06a2605c2d5ad9402e06d81c1a168ef

 

 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)