[jira] [Updated] (IGNITE-22542) Synchronous message handling on local node
[ https://issues.apache.org/jira/browse/IGNITE-22542?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Nikita Sivkov updated IGNITE-22542: --- Description: {{org.apache.ignite.internal.network.DefaultMessagingService#isSelf}} - if we detect that we send a message to the local node, we handle it immediately in the same thread, which could be very bad for throughput of the system. "send"/"invoke" themselves appear to be slow as well, we should benchmark them. We should remove instantiation of InetSocketAddress for sure, if it's possible, it takes time to resolve it. Maybe we should create it unresolved or just cache it like in Ignite 2. was: {{org.apache.ignite.internal.network.DefaultMessagingService#isSelf}} - if we detect that we send a message to the local node, we handle it immediately in the same thread, which could be very bed for throughput of the system. "send"/"invoke" themselves appear to be slow as well, we should benchmark them. We should remove instantiation of InetSocketAddress for sure, if it's possible, it takes time to resolve it. Maybe we should create it unresolved or just cache it like in Ignite 2. > Synchronous message handling on local node > -- > > Key: IGNITE-22542 > URL: https://issues.apache.org/jira/browse/IGNITE-22542 > Project: Ignite > Issue Type: Bug >Reporter: Ivan Bessonov >Priority: Major > Labels: ignite-3 > > {{org.apache.ignite.internal.network.DefaultMessagingService#isSelf}} - if we > detect that we send a message to the local node, we handle it immediately in > the same thread, which could be very bad for throughput of the system. > "send"/"invoke" themselves appear to be slow as well, we should benchmark > them. We should remove instantiation of InetSocketAddress for sure, if it's > possible, it takes time to resolve it. Maybe we should create it unresolved > or just cache it like in Ignite 2. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (IGNITE-22424) Insert and read errors under high cluster load
[ https://issues.apache.org/jira/browse/IGNITE-22424?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Nikita Sivkov updated IGNITE-22424: --- Description: h2. Setup * 1 or 3 server nodes cluster * 3 client nodes * 64+ threads per client performing KV put/get * Each node (server or client) is an AWS instance of {{c5d.4xlarge}} h2. Steps * Client nodes do put 15m records * Client nodes do get 15m records h2. Expected result No errors occurred, all records inserted and read after successfully. h2. Actual result On some point, cluster degrades and throws errors on inserting/reading. Error examples: * The primary replica has changed {code:java} org.apache.ignite.lang.IgniteException: The primary replica has changed [expectedLeaseholderName=poc-tester-SERVER-192.168.1.58-id-0, currentLeaseholderName=null, expectedLeaseholderId=b4425409-2d71-482d-a5d6-468522569cd0, currentLeaseholderId=null, expectedEnlistmentConsistencyToken=112569527664115719, currentEnlistmentConsistencyToken=null] at java.lang.invoke.MethodHandle.invokeWithArguments(MethodHandle.java:710) ~[?:?] at org.apache.ignite.internal.util.ExceptionUtils$1.copy(ExceptionUtils.java:789) ~[ignite-core-3.0.0-SNAPSHOT.jar:?] at org.apache.ignite.internal.util.ExceptionUtils$ExceptionFactory.createCopy(ExceptionUtils.java:723) ~[ignite-core-3.0.0-SNAPSHOT.jar:?] at org.apache.ignite.internal.util.ExceptionUtils.copyExceptionWithCause(ExceptionUtils.java:525) ~[ignite-core-3.0.0-SNAPSHOT.jar:?] at org.apache.ignite.internal.client.ClientUtils.copyExceptionWithCauseIfPossible(ClientUtils.java:73) ~[ignite-client-3.0.0-SNAPSHOT.jar:?] at org.apache.ignite.internal.client.ClientUtils.ensurePublicException(ClientUtils.java:54) ~[ignite-client-3.0.0-SNAPSHOT.jar:?] at org.apache.ignite.internal.client.ClientUtils.sync(ClientUtils.java:97) ~[ignite-client-3.0.0-SNAPSHOT.jar:?] at org.apache.ignite.internal.client.table.ClientKeyValueBinaryView.get(ClientKeyValueBinaryView.java:78) ~[ignite-client-3.0.0-SNAPSHOT.jar:?] at org.apache.ignite.internal.client.table.ClientKeyValueBinaryView.get(ClientKeyValueBinaryView.java:59) ~[ignite-client-3.0.0-SNAPSHOT.jar:?] at site.ycsb.db.ignite3.IgniteClient.read(IgniteClient.java:90) [ignite3-binding-2024.11.jar:?] at site.ycsb.DBWrapper.read(DBWrapper.java:157) [core-2024.11.jar:?] at site.ycsb.workloads.CoreWorkload.doTransactionRead(CoreWorkload.java:803) [core-2024.11.jar:?] at site.ycsb.workloads.CoreWorkload.doTransaction(CoreWorkload.java:722) [core-2024.11.jar:?] at site.ycsb.ClientThread.run(ClientThread.java:145) [core-2024.11.jar:?] at java.lang.Thread.run(Thread.java:829) [?:?] {code} * Replication is timed out {code:java} org.apache.ignite.tx.TransactionException: Replication is timed out [replicaGrpId=10_part_4] at java.lang.invoke.MethodHandle.invokeWithArguments(MethodHandle.java:710) ~[?:?] at org.apache.ignite.internal.util.ExceptionUtils$1.copy(ExceptionUtils.java:789) ~[ignite-core-3.0.0-SNAPSHOT.jar:?] at org.apache.ignite.internal.util.ExceptionUtils$ExceptionFactory.createCopy(ExceptionUtils.java:723) ~[ignite-core-3.0.0-SNAPSHOT.jar:?] at org.apache.ignite.internal.util.ExceptionUtils.copyExceptionWithCause(ExceptionUtils.java:525) ~[ignite-core-3.0.0-SNAPSHOT.jar:?] at org.apache.ignite.internal.client.ClientUtils.copyExceptionWithCauseIfPossible(ClientUtils.java:73) ~[ignite-client-3.0.0-SNAPSHOT.jar:?] at org.apache.ignite.internal.client.ClientUtils.ensurePublicException(ClientUtils.java:54) ~[ignite-client-3.0.0-SNAPSHOT.jar:?] at org.apache.ignite.internal.client.ClientUtils.sync(ClientUtils.java:97) ~[ignite-client-3.0.0-SNAPSHOT.jar:?] at org.apache.ignite.internal.client.table.ClientKeyValueBinaryView.get(ClientKeyValueBinaryView.java:78) ~[ignite-client-3.0.0-SNAPSHOT.jar:?] at org.apache.ignite.internal.client.table.ClientKeyValueBinaryView.get(ClientKeyValueBinaryView.java:59) ~[ignite-client-3.0.0-SNAPSHOT.jar:?] at site.ycsb.db.ignite3.IgniteClient.read(IgniteClient.java:90) [ignite3-binding-2024.11.jar:?] at site.ycsb.DBWrapper.read(DBWrapper.java:157) [core-2024.11.jar:?] at site.ycsb.workloads.CoreWorkload.doTransactionRead(CoreWorkload.java:803) [core-2024.11.jar:?] at site.ycsb.workloads.CoreWorkload.doTransaction(CoreWorkload.java:722) [core-2024.11.jar:?] at site.ycsb.ClientThread.run(ClientThread.java:145) [core-2024.11.jar:?] at java.lang.Thread.run(Thread.java:829) [?:?] {code} * A critical thread is blocked for XXX ms that is more than the allowed 500 ms {code:java} 2024-06-04 22:26:49:031 +0300 [ERROR][%poc-tester-SERVER-192.168.1.119-id-0%common-scheduler-0][CriticalWorkerWatchdog] A critical thread
[jira] [Updated] (IGNITE-22424) Insert and read errors under high cluster load
[ https://issues.apache.org/jira/browse/IGNITE-22424?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Nikita Sivkov updated IGNITE-22424: --- Description: h2. Setup * 1 or 3 server nodes cluster * 3 client nodes * 64+ threads per client performing KV put/get * Each node (server or client) is an AWS instance of {{c5d.4xlarge}} h2. Steps * Client nodes do put 15m records * Client nodes do get 15m records h2. Expected result No errors occurred, all records inserted and read after successfully. h2. Actual result On some point, cluster degrades and throws errors on inserting/reading. Error examples: * The primary replica has changed {code:java} org.apache.ignite.lang.IgniteException: The primary replica has changed [expectedLeaseholderName=poc-tester-SERVER-192.168.1.58-id-0, currentLeaseholderName=null, expectedLeaseholderId=b4425409-2d71-482d-a5d6-468522569cd0, currentLeaseholderId=null, expectedEnlistmentConsistencyToken=112569527664115719, currentEnlistmentConsistencyToken=null] at java.lang.invoke.MethodHandle.invokeWithArguments(MethodHandle.java:710) ~[?:?] at org.apache.ignite.internal.util.ExceptionUtils$1.copy(ExceptionUtils.java:789) ~[ignite-core-3.0.0-SNAPSHOT.jar:?] at org.apache.ignite.internal.util.ExceptionUtils$ExceptionFactory.createCopy(ExceptionUtils.java:723) ~[ignite-core-3.0.0-SNAPSHOT.jar:?] at org.apache.ignite.internal.util.ExceptionUtils.copyExceptionWithCause(ExceptionUtils.java:525) ~[ignite-core-3.0.0-SNAPSHOT.jar:?] at org.apache.ignite.internal.client.ClientUtils.copyExceptionWithCauseIfPossible(ClientUtils.java:73) ~[ignite-client-3.0.0-SNAPSHOT.jar:?] at org.apache.ignite.internal.client.ClientUtils.ensurePublicException(ClientUtils.java:54) ~[ignite-client-3.0.0-SNAPSHOT.jar:?] at org.apache.ignite.internal.client.ClientUtils.sync(ClientUtils.java:97) ~[ignite-client-3.0.0-SNAPSHOT.jar:?] at org.apache.ignite.internal.client.table.ClientKeyValueBinaryView.get(ClientKeyValueBinaryView.java:78) ~[ignite-client-3.0.0-SNAPSHOT.jar:?] at org.apache.ignite.internal.client.table.ClientKeyValueBinaryView.get(ClientKeyValueBinaryView.java:59) ~[ignite-client-3.0.0-SNAPSHOT.jar:?] at site.ycsb.db.ignite3.IgniteClient.read(IgniteClient.java:90) [ignite3-binding-2024.11.jar:?] at site.ycsb.DBWrapper.read(DBWrapper.java:157) [core-2024.11.jar:?] at site.ycsb.workloads.CoreWorkload.doTransactionRead(CoreWorkload.java:803) [core-2024.11.jar:?] at site.ycsb.workloads.CoreWorkload.doTransaction(CoreWorkload.java:722) [core-2024.11.jar:?] at site.ycsb.ClientThread.run(ClientThread.java:145) [core-2024.11.jar:?] at java.lang.Thread.run(Thread.java:829) [?:?] {code} * A critical thread is blocked for XXX ms that is more than the allowed 500 ms {code:java} 2024-06-04 22:26:49:031 +0300 [ERROR][%poc-tester-SERVER-192.168.1.119-id-0%common-scheduler-0][CriticalWorkerWatchdog] A critical thread is blocked for 512 ms that is more than the allowed 500 ms, it is "poc-tester-SERVER-192.168.1.119-id-0-client-32" prio=10 Id=122 RUNNABLE (in native) at java.base@11.0.12/sun.nio.ch.EPoll.wait(Native Method) at java.base@11.0.12/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:120) at java.base@11.0.12/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:124) - locked io.netty.channel.nio.SelectedSelectionKeySet@71be3d25 - locked sun.nio.ch.EPollSelectorImpl@5aebdd7 at java.base@11.0.12/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) at app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) at app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) at app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) at app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) at app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) at app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) at java.base@11.0.12/java.lang.Thread.run(Thread.java:829){code} * Recoverable error during the request occurred (will be retried on the randomly selected node) {code:java} 2024-06-04 22:26:49:306 +0300 [WARNING][CompletableFutureDelayScheduler][RaftGroupServiceImpl] Recoverable error during the request occurred (will be retried on the randomly selected node) [request=WriteActionRequestImpl [command=[0, 9, 41, -117, -128, -40, -77, -3, -106, -7, -57, 1], deserializedCommand=SafeTimeSyncCommandImpl [safeTimeLong=112559993935757322], groupId=10_part_10], peer=Peer [consistentId=poc-tester-SERVER-192.168.1.119-id-0, idx=0], newPeer=Peer [consistentId=poc-tester-SERVER-192.168.1.119-id-0, idx=0]]. java.util.concurrent.CompletionException:
[jira] [Created] (IGNITE-22424) Insert and read errors under high cluster load
Nikita Sivkov created IGNITE-22424: -- Summary: Insert and read errors under high cluster load Key: IGNITE-22424 URL: https://issues.apache.org/jira/browse/IGNITE-22424 Project: Ignite Issue Type: Bug Reporter: Nikita Sivkov h2. Setup * 1 or 3 server nodes cluster * 3 client nodes * 64+ threads per client performing KV put/get * Each node (server or client) is an AWS instance of `c5d.4xlarge` h2. Steps * Client nodes do put 15m records * Client nodes do get 15m records h2. Expected result No errors occurred, all records inserted and read after successfully. h2. Actual result On some point, cluster degrades and throws errors on inserting/reading. Error examples: * The primary replica has changed {code:java} org.apache.ignite.lang.IgniteException: The primary replica has changed [expectedLeaseholderName=poc-tester-SERVER-192.168.1.58-id-0, currentLeaseholderName=null, expectedLeaseholderId=b4425409-2d71-482d-a5d6-468522569cd0, currentLeaseholderId=null, expectedEnlistmentConsistencyToken=112569527664115719, currentEnlistmentConsistencyToken=null] at java.lang.invoke.MethodHandle.invokeWithArguments(MethodHandle.java:710) ~[?:?] at org.apache.ignite.internal.util.ExceptionUtils$1.copy(ExceptionUtils.java:789) ~[ignite-core-3.0.0-SNAPSHOT.jar:?] at org.apache.ignite.internal.util.ExceptionUtils$ExceptionFactory.createCopy(ExceptionUtils.java:723) ~[ignite-core-3.0.0-SNAPSHOT.jar:?] at org.apache.ignite.internal.util.ExceptionUtils.copyExceptionWithCause(ExceptionUtils.java:525) ~[ignite-core-3.0.0-SNAPSHOT.jar:?] at org.apache.ignite.internal.client.ClientUtils.copyExceptionWithCauseIfPossible(ClientUtils.java:73) ~[ignite-client-3.0.0-SNAPSHOT.jar:?] at org.apache.ignite.internal.client.ClientUtils.ensurePublicException(ClientUtils.java:54) ~[ignite-client-3.0.0-SNAPSHOT.jar:?] at org.apache.ignite.internal.client.ClientUtils.sync(ClientUtils.java:97) ~[ignite-client-3.0.0-SNAPSHOT.jar:?] at org.apache.ignite.internal.client.table.ClientKeyValueBinaryView.get(ClientKeyValueBinaryView.java:78) ~[ignite-client-3.0.0-SNAPSHOT.jar:?] at org.apache.ignite.internal.client.table.ClientKeyValueBinaryView.get(ClientKeyValueBinaryView.java:59) ~[ignite-client-3.0.0-SNAPSHOT.jar:?] at site.ycsb.db.ignite3.IgniteClient.read(IgniteClient.java:90) [ignite3-binding-2024.11.jar:?] at site.ycsb.DBWrapper.read(DBWrapper.java:157) [core-2024.11.jar:?] at site.ycsb.workloads.CoreWorkload.doTransactionRead(CoreWorkload.java:803) [core-2024.11.jar:?] at site.ycsb.workloads.CoreWorkload.doTransaction(CoreWorkload.java:722) [core-2024.11.jar:?] at site.ycsb.ClientThread.run(ClientThread.java:145) [core-2024.11.jar:?] at java.lang.Thread.run(Thread.java:829) [?:?] {code} * A critical thread is blocked for XXX ms that is more than the allowed 500 ms {code:java} 2024-06-04 22:26:49:031 +0300 [ERROR][%poc-tester-SERVER-192.168.1.119-id-0%common-scheduler-0][CriticalWorkerWatchdog] A critical thread is blocked for 512 ms that is more than the allowed 500 ms, it is "poc-tester-SERVER-192.168.1.119-id-0-client-32" prio=10 Id=122 RUNNABLE (in native) at java.base@11.0.12/sun.nio.ch.EPoll.wait(Native Method) at java.base@11.0.12/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:120) at java.base@11.0.12/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:124) - locked io.netty.channel.nio.SelectedSelectionKeySet@71be3d25 - locked sun.nio.ch.EPollSelectorImpl@5aebdd7 at java.base@11.0.12/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) at app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) at app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) at app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) at app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) at app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) at app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) at java.base@11.0.12/java.lang.Thread.run(Thread.java:829){code} * Recoverable error during the request occurred (will be retried on the randomly selected node) {code:java} 2024-06-04 22:26:49:306 +0300 [WARNING][CompletableFutureDelayScheduler][RaftGroupServiceImpl] Recoverable error during the request occurred (will be retried on the randomly selected node) [request=WriteActionRequestImpl [command=[0, 9, 41, -117, -128, -40, -77, -3, -106, -7, -57, 1], deserializedCommand=SafeTimeSyncCommandImpl [safeTimeLong=112559993935757322], groupId=10_part_10], peer=Peer [consistentId=poc-tester-SERVER-192.168.1.119-id-0,
[jira] [Updated] (IGNITE-22248) Creation of new tables in 1 node cluster stuck after 850+ tables
[ https://issues.apache.org/jira/browse/IGNITE-22248?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Nikita Sivkov updated IGNITE-22248: --- Description: *Steps to reproduce:* # Single node cluster with arguments "-Xms4096m", "-Xmx4096m" # Create tables one by one up to 1000 *Expected:* 1000 tables are created. *Actual:* After 850+ tables the creation time is higher than 30 seconds or creation table request fails. !image-2024-05-15-13-22-40-059.png! In the server logs continuous errors: {code:java} 2024-05-15 04:11:58:116 + [WARNING][CompletableFutureDelayScheduler][RaftGroupServiceImpl] Recoverable error during the request occurred (will be retried on the randomly selected node) [request=WriteActionRequestImpl [command=[0, 9, 41, -126, -128, -36, -49, -79, -50, -34, -57, 1], deserializedCommand=SafeTimeSyncCommandImpl [safeTimeLong=112443150482997249], groupId=950_part_21], peer=Peer [consistentId=TablesAmountCapacityTest_cluster_0, idx=0], newPeer=Peer [consistentId=TablesAmountCapacityTest_cluster_0, idx=0]]. java.util.concurrent.CompletionException: java.util.concurrent.TimeoutException at java.base/java.util.concurrent.CompletableFuture.encodeRelay(CompletableFuture.java:368) at java.base/java.util.concurrent.CompletableFuture.completeRelay(CompletableFuture.java:377) at java.base/java.util.concurrent.CompletableFuture$UniRelay.tryFire(CompletableFuture.java:1097) at java.base/java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:510) at java.base/java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:2162) at java.base/java.util.concurrent.CompletableFuture$Timeout.run(CompletableFuture.java:2874) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:539) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:304) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) at java.base/java.lang.Thread.run(Thread.java:833) Caused by: java.util.concurrent.TimeoutException ... 7 more {code} was: *Steps to reproduce:* # Single node cluster with arguments "-Xms4096m", "-Xmx4096m" # Create tables one by one up to 1000 *Expected:* 1000 tables are created. *Actual:* After 850+ tables the creation time is higher than 30 seconds. !image-2024-05-15-13-22-40-059.png! In the server logs continuous errors: {code:java} 2024-05-15 04:11:58:116 + [WARNING][CompletableFutureDelayScheduler][RaftGroupServiceImpl] Recoverable error during the request occurred (will be retried on the randomly selected node) [request=WriteActionRequestImpl [command=[0, 9, 41, -126, -128, -36, -49, -79, -50, -34, -57, 1], deserializedCommand=SafeTimeSyncCommandImpl [safeTimeLong=112443150482997249], groupId=950_part_21], peer=Peer [consistentId=TablesAmountCapacityTest_cluster_0, idx=0], newPeer=Peer [consistentId=TablesAmountCapacityTest_cluster_0, idx=0]]. java.util.concurrent.CompletionException: java.util.concurrent.TimeoutException at java.base/java.util.concurrent.CompletableFuture.encodeRelay(CompletableFuture.java:368) at java.base/java.util.concurrent.CompletableFuture.completeRelay(CompletableFuture.java:377) at java.base/java.util.concurrent.CompletableFuture$UniRelay.tryFire(CompletableFuture.java:1097) at java.base/java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:510) at java.base/java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:2162) at java.base/java.util.concurrent.CompletableFuture$Timeout.run(CompletableFuture.java:2874) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:539) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:304) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) at java.base/java.lang.Thread.run(Thread.java:833) Caused by: java.util.concurrent.TimeoutException ... 7 more {code} > Creation of new tables in 1 node cluster stuck after 850+ tables > > > Key: IGNITE-22248 > URL: https://issues.apache.org/jira/browse/IGNITE-22248 > Project: Ignite > Issue Type: Bug >
[jira] [Resolved] (IGNITE-22111) [SQL] TPC-H q21 query hangs with sc=0.1
[ https://issues.apache.org/jira/browse/IGNITE-22111?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Nikita Sivkov resolved IGNITE-22111. Resolution: Not A Problem Fixed with creating indexes as it done in [https://github.com/apache/ignite-3/tree/main/modules/sql-engine/src/testFixtures/resources/tpch/ddl] > [SQL] TPC-H q21 query hangs with sc=0.1 > --- > > Key: IGNITE-22111 > URL: https://issues.apache.org/jira/browse/IGNITE-22111 > Project: Ignite > Issue Type: Bug > Components: sql >Reporter: Nikita Sivkov >Priority: Major > Labels: ignite-3, ignite3_performance > > Benchmark: > [https://github.com/cmu-db/benchbase/tree/main/src/main/java/com/oltpbenchmark/benchmarks/tpch] > > h1. Setup > * 1 server node > * TPC-H with scale factor = 0.1 > h1. Steps > # Start an Ignite node > # Run benchbase with {{-s 1 --create=true --load=true --execute=false}} to > preload data > # Observe via the benchbase log that the data was successfully loaded > # Run {{benchbase with -s 1 --create=false --load=false --execute=true}} to > run the benchmark > h1. Expected result > The benchmark finishes after warmup + duration time > h1. Actual result > The benchmark hangs for hours on > [Q21|https://github.com/cmu-db/benchbase/blob/main/src/main/java/com/oltpbenchmark/benchmarks/tpch/procedures/Q21.java] > query: > > {noformat} > SELECT > s_name, > COUNT(*) AS numwait > FROM > supplier, > lineitem l1, > orders, > nation > WHERE > s_suppkey = l1.l_suppkey > AND o_orderkey = l1.l_orderkey > AND o_orderstatus = 'F' > AND l1.l_receiptdate > l1.l_commitdate > AND EXISTS > ( > SELECT > * > FROM > lineitem l2 > WHERE > l2.l_orderkey = l1.l_orderkey > AND l2.l_suppkey <> l1.l_suppkey > ) > AND NOT EXISTS > ( > SELECT > * > FROM > lineitem l3 > WHERE > l3.l_orderkey = l1.l_orderkey > AND l3.l_suppkey <> l1.l_suppkey > AND l3.l_receiptdate > l3.l_commitdate > ) > AND s_nationkey = n_nationkey > AND n_name = ? > GROUP BY > s_name > ORDER BY > numwait DESC, > s_name LIMIT 100{noformat} -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Resolved] (IGNITE-22109) [SQL] TPC-H q4 query hangs with sc=0.1
[ https://issues.apache.org/jira/browse/IGNITE-22109?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Nikita Sivkov resolved IGNITE-22109. Resolution: Not A Problem Fixed with creating indexes as it done in [https://github.com/apache/ignite-3/tree/main/modules/sql-engine/src/testFixtures/resources/tpch/ddl|https://github.com/apache/ignite-3/tree/main/modules/sql-engine/src/testFixtures/resources/tpch/ddl] > [SQL] TPC-H q4 query hangs with sc=0.1 > -- > > Key: IGNITE-22109 > URL: https://issues.apache.org/jira/browse/IGNITE-22109 > Project: Ignite > Issue Type: Bug > Components: sql >Reporter: Nikita Sivkov >Priority: Major > Labels: ignite-3, ignite3_performance > > Benchmark: > [https://github.com/cmu-db/benchbase/tree/main/src/main/java/com/oltpbenchmark/benchmarks/tpch] > > h1. Setup > * 1 server node > * TPC-H with scale factor = 0.1 > h1. Steps > # Start an Ignite node > # Run benchbase with {{-s 1 --create=true --load=true --execute=false}} to > preload data > # Observe via the benchbase log that the data was successfully loaded > # Run {{benchbase with -s 1 --create=false --load=false --execute=true}} to > run the benchmark > h1. Expected result > The benchmark finishes after warmup + duration time > h1. Actual result > The benchmark hangs for hours on > [Q4|https://github.com/cmu-db/benchbase/blob/main/src/main/java/com/oltpbenchmark/benchmarks/tpch/procedures/Q4.java] > query: > > {noformat} > SELECT > o_orderpriority, > COUNT(*) AS order_count > FROM > orders > WHERE > o_orderdate >= ?::date > AND o_orderdate < ?::date + INTERVAL '3' MONTH > AND EXISTS > ( > SELECT > * > FROM > lineitem > WHERE > l_orderkey = o_orderkey > AND l_commitdate < l_receiptdate > ) > GROUP BY > o_orderpriority > ORDER BY > o_orderpriority{noformat} -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Resolved] (IGNITE-22110) [SQL] TPC-H q16 query hangs with sc=0.1
[ https://issues.apache.org/jira/browse/IGNITE-22110?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Nikita Sivkov resolved IGNITE-22110. Resolution: Not A Problem Fixed with creating indexes as it done in [https://github.com/apache/ignite-3/tree/main/modules/sql-engine/src/testFixtures/resources/tpch/ddl] > [SQL] TPC-H q16 query hangs with sc=0.1 > --- > > Key: IGNITE-22110 > URL: https://issues.apache.org/jira/browse/IGNITE-22110 > Project: Ignite > Issue Type: Bug > Components: sql >Reporter: Nikita Sivkov >Priority: Major > Labels: ignite-3, ignite3_performance > > Benchmark: > [https://github.com/cmu-db/benchbase/tree/main/src/main/java/com/oltpbenchmark/benchmarks/tpch] > > h1. Setup > * 1 server node > * TPC-H with scale factor = 0.1 > h1. Steps > # Start an Ignite node > # Run benchbase with {{-s 1 --create=true --load=true --execute=false}} to > preload data > # Observe via the benchbase log that the data was successfully loaded > # Run {{benchbase with -s 1 --create=false --load=false --execute=true}} to > run the benchmark > h1. Expected result > The benchmark finishes after warmup + duration time > h1. Actual result > The benchmark hangs for hours on > [Q16|https://github.com/cmu-db/benchbase/blob/main/src/main/java/com/oltpbenchmark/benchmarks/tpch/procedures/Q16.java] > query: > > {noformat} > SELECT > p_brand, > p_type, > p_size, > COUNT(DISTINCT ps_suppkey) AS supplier_cnt > FROM > partsupp, > part > WHERE > p_partkey = ps_partkey > AND p_brand <> ? > AND p_type NOT LIKE ? > AND p_size IN (?, ?, ?, ?, ?, ?, ?, ?) > AND ps_suppkey NOT IN > ( > SELECT > s_suppkey > FROM > supplier > WHERE > s_comment LIKE '%Customer%Complaints%' > ) > GROUP BY > p_brand, > p_type, > p_size > ORDER BY > supplier_cnt DESC, > p_brand, > p_type, > p_size{noformat} -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (IGNITE-22109) [SQL] TPC-H q4 query hangs with sc=0.1
[ https://issues.apache.org/jira/browse/IGNITE-22109?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Nikita Sivkov updated IGNITE-22109: --- Description: Benchmark: [https://github.com/cmu-db/benchbase/tree/main/src/main/java/com/oltpbenchmark/benchmarks/tpch] h1. Setup * 1 server node * TPC-H with scale factor = 0.1 h1. Steps # Start an Ignite node # Run benchbase with {{-s 1 --create=true --load=true --execute=false}} to preload data # Observe via the benchbase log that the data was successfully loaded # Run {{benchbase with -s 1 --create=false --load=false --execute=true}} to run the benchmark h1. Expected result The benchmark finishes after warmup + duration time h1. Actual result The benchmark hangs for hours on [Q4|https://github.com/cmu-db/benchbase/blob/main/src/main/java/com/oltpbenchmark/benchmarks/tpch/procedures/Q4.java] query: {noformat} SELECT o_orderpriority, COUNT(*) AS order_count FROM orders WHERE o_orderdate >= ?::date AND o_orderdate < ?::date + INTERVAL '3' MONTH AND EXISTS ( SELECT * FROM lineitem WHERE l_orderkey = o_orderkey AND l_commitdate < l_receiptdate ) GROUP BY o_orderpriority ORDER BY o_orderpriority{noformat} was: Benchmark: [https://github.com/cmu-db/benchbase/tree/main/src/main/java/com/oltpbenchmark/benchmarks/tpch] h1. Setup * 1 server node * TPC-H with scale factor = 0.1 h1. Steps # Start an Ignite node # Run benchbase with {{-s 1 --create=true --load=true --execute=false}} to preload data # Observe via the benchbase log that the data was successfully loaded # Run {{benchbase with -s 1 --create=false --load=false --execute=true}} to run the benchmark h1. Expected result The benchmark finishes after warmup + duration time h1. Actual result The benchmark hangs for hours on [Q4|https://github.com/cmu-db/benchbase/blob/main/src/main/java/com/oltpbenchmark/benchmarks/tpch/procedures/Q4.java] query: {noformat} SELECT o_orderpriority, COUNT(*) AS order_count FROM orders WHERE o_orderdate >= DATE ? AND o_orderdate < DATE ? + INTERVAL '3' MONTH AND EXISTS ( SELECT * FROM lineitem WHERE l_orderkey = o_orderkey AND l_commitdate < l_receiptdate ) GROUP BY o_orderpriority ORDER BY o_orderpriority {noformat} > [SQL] TPC-H q4 query hangs with sc=0.1 > -- > > Key: IGNITE-22109 > URL: https://issues.apache.org/jira/browse/IGNITE-22109 > Project: Ignite > Issue Type: Bug > Components: sql >Reporter: Nikita Sivkov >Priority: Major > Labels: ignite-3, ignite3_performance > > Benchmark: > [https://github.com/cmu-db/benchbase/tree/main/src/main/java/com/oltpbenchmark/benchmarks/tpch] > > h1. Setup > * 1 server node > * TPC-H with scale factor = 0.1 > h1. Steps > # Start an Ignite node > # Run benchbase with {{-s 1 --create=true --load=true --execute=false}} to > preload data > # Observe via the benchbase log that the data was successfully loaded > # Run {{benchbase with -s 1 --create=false --load=false --execute=true}} to > run the benchmark > h1. Expected result > The benchmark finishes after warmup + duration time > h1. Actual result > The benchmark hangs for hours on > [Q4|https://github.com/cmu-db/benchbase/blob/main/src/main/java/com/oltpbenchmark/benchmarks/tpch/procedures/Q4.java] > query: > > {noformat} > SELECT > o_orderpriority, > COUNT(*) AS order_count > FROM > orders > WHERE > o_orderdate >= ?::date > AND o_orderdate < ?::date + INTERVAL '3' MONTH > AND EXISTS > ( > SELECT > * > FROM > lineitem > WHERE > l_orderkey = o_orderkey > AND l_commitdate < l_receiptdate > ) > GROUP BY > o_orderpriority > ORDER BY > o_orderpriority{noformat} -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (IGNITE-22112) [SQL] TPC-H q9 query with sc=0.1 takes long time
[ https://issues.apache.org/jira/browse/IGNITE-22112?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Nikita Sivkov updated IGNITE-22112: --- Description: Benchmark: [https://github.com/cmu-db/benchbase/tree/main/src/main/java/com/oltpbenchmark/benchmarks/tpch] h1. Setup * 1 server node * TPC-H with scale factor = 0.1 h1. Steps # Start an Ignite node # Run benchbase with {{-s 1 --create=true --load=true --execute=false}} to preload data # Observe via the benchbase log that the data was successfully loaded # Run {{benchbase with -s 1 --create=false --load=false --execute=true}} to run the benchmark h1. Expected result The benchmark finishes after warmup + duration time in about the same time as other requests. h1. Actual result The [Q9|https://github.com/cmu-db/benchbase/blob/main/src/main/java/com/oltpbenchmark/benchmarks/tpch/procedures/Q9.java] query has significantly longer execution time. For example, it may take 10 minutes, while the rest queries take ~1-15 seconds. {noformat} SELECT nation, o_year, SUM(amount) AS sum_profit FROM ( SELECT n_name AS nation, EXTRACT(YEAR FROM o_orderdate) AS o_year, l_extendedprice * (1 - l_discount) - ps_supplycost * l_quantity AS amount FROM part, supplier, lineitem, partsupp, orders, nation WHERE s_suppkey = l_suppkey AND ps_suppkey = l_suppkey AND ps_partkey = l_partkey AND p_partkey = l_partkey AND o_orderkey = l_orderkey AND s_nationkey = n_nationkey AND p_name LIKE ? ) AS profit GROUP BY nation, o_year ORDER BY nation, o_year DESC{noformat} was: Benchmark: [https://github.com/cmu-db/benchbase/tree/main/src/main/java/com/oltpbenchmark/benchmarks/tpch] h1. Setup * 1 server node * TPC-H with scale factor = 0.1 h1. Steps # Start an Ignite node # Run benchbase with {{-s 1 --create=true --load=true --execute=false}} to preload data # Observe via the benchbase log that the data was successfully loaded # Run {{benchbase with -s 1 --create=false --load=false --execute=true}} to run the benchmark h1. Expected result The benchmark finishes after warmup + duration time h1. Actual result The benchmark hangs for hours on [Q21|https://github.com/cmu-db/benchbase/blob/main/src/main/java/com/oltpbenchmark/benchmarks/tpch/procedures/Q21.java] query: {noformat} SELECT s_name, COUNT(*) AS numwait FROM supplier, lineitem l1, orders, nation WHERE s_suppkey = l1.l_suppkey AND o_orderkey = l1.l_orderkey AND o_orderstatus = 'F' AND l1.l_receiptdate > l1.l_commitdate AND EXISTS ( SELECT * FROM lineitem l2 WHERE l2.l_orderkey = l1.l_orderkey AND l2.l_suppkey <> l1.l_suppkey ) AND NOT EXISTS ( SELECT * FROM lineitem l3 WHERE l3.l_orderkey = l1.l_orderkey AND l3.l_suppkey <> l1.l_suppkey AND l3.l_receiptdate > l3.l_commitdate ) AND s_nationkey = n_nationkey AND n_name = ? GROUP BY s_name ORDER BY numwait DESC, s_name LIMIT 100{noformat} > [SQL] TPC-H q9 query with sc=0.1 takes long time > > > Key: IGNITE-22112 > URL: https://issues.apache.org/jira/browse/IGNITE-22112 > Project: Ignite > Issue Type: Bug > Components: sql >Reporter: Nikita Sivkov >Priority: Major > Labels: ignite-3, ignite3_performance > > Benchmark: > [https://github.com/cmu-db/benchbase/tree/main/src/main/java/com/oltpbenchmark/benchmarks/tpch] > > h1. Setup > * 1 server node > * TPC-H with scale factor = 0.1 > h1. Steps > # Start an Ignite node > # Run benchbase with {{-s 1 --create=true --load=true --execute=false}} to > preload data > # Observe via the benchbase log that the data was successfully loaded > # Run {{benchbase with -s 1 --create=false --load=false --execute=true}} to > run the benchmark > h1. Expected result > The benchmark finishes after warmup + duration time in about the same time as > other requests. > h1. Actual result > The > [Q9|https://github.com/cmu-db/benchbase/blob/main/src/main/java/com/oltpbenchmark/benchmarks/tpch/procedures/Q9.java] > query has significantly longer execution time. For example, it may take 10 > minutes, while the rest queries take ~1-15 seconds. > {noformat} > SELECT > nation, > o_year, > SUM(amount) AS sum_profit > FROM > ( > SELECT > n_name AS nation, > EXTRACT(YEAR > FROM > o_orderdate) AS o_year, > l_extendedprice * (1 - l_discount) - ps_supplycost * l_quantity AS > amount > FROM > part, > supplier, > lineitem, > partsupp, >
[jira] [Created] (IGNITE-22112) [SQL] TPC-H q9 query with sc=0.1 takes long time
Nikita Sivkov created IGNITE-22112: -- Summary: [SQL] TPC-H q9 query with sc=0.1 takes long time Key: IGNITE-22112 URL: https://issues.apache.org/jira/browse/IGNITE-22112 Project: Ignite Issue Type: Bug Components: sql Reporter: Nikita Sivkov Benchmark: [https://github.com/cmu-db/benchbase/tree/main/src/main/java/com/oltpbenchmark/benchmarks/tpch] h1. Setup * 1 server node * TPC-H with scale factor = 0.1 h1. Steps # Start an Ignite node # Run benchbase with {{-s 1 --create=true --load=true --execute=false}} to preload data # Observe via the benchbase log that the data was successfully loaded # Run {{benchbase with -s 1 --create=false --load=false --execute=true}} to run the benchmark h1. Expected result The benchmark finishes after warmup + duration time h1. Actual result The benchmark hangs for hours on [Q21|https://github.com/cmu-db/benchbase/blob/main/src/main/java/com/oltpbenchmark/benchmarks/tpch/procedures/Q21.java] query: {noformat} SELECT s_name, COUNT(*) AS numwait FROM supplier, lineitem l1, orders, nation WHERE s_suppkey = l1.l_suppkey AND o_orderkey = l1.l_orderkey AND o_orderstatus = 'F' AND l1.l_receiptdate > l1.l_commitdate AND EXISTS ( SELECT * FROM lineitem l2 WHERE l2.l_orderkey = l1.l_orderkey AND l2.l_suppkey <> l1.l_suppkey ) AND NOT EXISTS ( SELECT * FROM lineitem l3 WHERE l3.l_orderkey = l1.l_orderkey AND l3.l_suppkey <> l1.l_suppkey AND l3.l_receiptdate > l3.l_commitdate ) AND s_nationkey = n_nationkey AND n_name = ? GROUP BY s_name ORDER BY numwait DESC, s_name LIMIT 100{noformat} -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (IGNITE-22111) [SQL] TPC-H q21 query hangs with sc=0.1
[ https://issues.apache.org/jira/browse/IGNITE-22111?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Nikita Sivkov updated IGNITE-22111: --- Description: Benchmark: [https://github.com/cmu-db/benchbase/tree/main/src/main/java/com/oltpbenchmark/benchmarks/tpch] h1. Setup * 1 server node * TPC-H with scale factor = 0.1 h1. Steps # Start an Ignite node # Run benchbase with {{-s 1 --create=true --load=true --execute=false}} to preload data # Observe via the benchbase log that the data was successfully loaded # Run {{benchbase with -s 1 --create=false --load=false --execute=true}} to run the benchmark h1. Expected result The benchmark finishes after warmup + duration time h1. Actual result The benchmark hangs for hours on [Q21|https://github.com/cmu-db/benchbase/blob/main/src/main/java/com/oltpbenchmark/benchmarks/tpch/procedures/Q21.java] query: {noformat} SELECT s_name, COUNT(*) AS numwait FROM supplier, lineitem l1, orders, nation WHERE s_suppkey = l1.l_suppkey AND o_orderkey = l1.l_orderkey AND o_orderstatus = 'F' AND l1.l_receiptdate > l1.l_commitdate AND EXISTS ( SELECT * FROM lineitem l2 WHERE l2.l_orderkey = l1.l_orderkey AND l2.l_suppkey <> l1.l_suppkey ) AND NOT EXISTS ( SELECT * FROM lineitem l3 WHERE l3.l_orderkey = l1.l_orderkey AND l3.l_suppkey <> l1.l_suppkey AND l3.l_receiptdate > l3.l_commitdate ) AND s_nationkey = n_nationkey AND n_name = ? GROUP BY s_name ORDER BY numwait DESC, s_name LIMIT 100{noformat} was: Benchmark: [https://github.com/cmu-db/benchbase/tree/main/src/main/java/com/oltpbenchmark/benchmarks/tpch] h1. Setup * 1 server node * TPC-H with scale factor = 0.1 h1. Steps # Start an Ignite node # Run benchbase with {{-s 1 --create=true --load=true --execute=false}} to preload data # Observe via the benchbase log that the data was successfully loaded # Run {{benchbase with -s 1 --create=false --load=false --execute=true}} to run the benchmark h1. Expected result The benchmark finishes after warmup + duration time h1. Actual result The benchmark hangs for hours on [Q16|https://github.com/cmu-db/benchbase/blob/main/src/main/java/com/oltpbenchmark/benchmarks/tpch/procedures/Q16.java] query: {noformat} SELECT p_brand, p_type, p_size, COUNT(DISTINCT ps_suppkey) AS supplier_cnt FROM partsupp, part WHERE p_partkey = ps_partkey AND p_brand <> ? AND p_type NOT LIKE ? AND p_size IN (?, ?, ?, ?, ?, ?, ?, ?) AND ps_suppkey NOT IN ( SELECT s_suppkey FROM supplier WHERE s_comment LIKE '%Customer%Complaints%' ) GROUP BY p_brand, p_type, p_size ORDER BY supplier_cnt DESC, p_brand, p_type, p_size{noformat} > [SQL] TPC-H q21 query hangs with sc=0.1 > --- > > Key: IGNITE-22111 > URL: https://issues.apache.org/jira/browse/IGNITE-22111 > Project: Ignite > Issue Type: Bug > Components: sql >Reporter: Nikita Sivkov >Priority: Major > Labels: ignite-3, ignite3_performance > > Benchmark: > [https://github.com/cmu-db/benchbase/tree/main/src/main/java/com/oltpbenchmark/benchmarks/tpch] > > h1. Setup > * 1 server node > * TPC-H with scale factor = 0.1 > h1. Steps > # Start an Ignite node > # Run benchbase with {{-s 1 --create=true --load=true --execute=false}} to > preload data > # Observe via the benchbase log that the data was successfully loaded > # Run {{benchbase with -s 1 --create=false --load=false --execute=true}} to > run the benchmark > h1. Expected result > The benchmark finishes after warmup + duration time > h1. Actual result > The benchmark hangs for hours on > [Q21|https://github.com/cmu-db/benchbase/blob/main/src/main/java/com/oltpbenchmark/benchmarks/tpch/procedures/Q21.java] > query: > > {noformat} > SELECT > s_name, > COUNT(*) AS numwait > FROM > supplier, > lineitem l1, > orders, > nation > WHERE > s_suppkey = l1.l_suppkey > AND o_orderkey = l1.l_orderkey > AND o_orderstatus = 'F' > AND l1.l_receiptdate > l1.l_commitdate > AND EXISTS > ( > SELECT > * > FROM > lineitem l2 > WHERE > l2.l_orderkey = l1.l_orderkey > AND l2.l_suppkey <> l1.l_suppkey > ) > AND NOT EXISTS > ( > SELECT > * > FROM > lineitem l3 > WHERE > l3.l_orderkey = l1.l_orderkey > AND l3.l_suppkey <> l1.l_suppkey > AND l3.l_receiptdate > l3.l_commitdate > ) > AND s_nationkey = n_nationkey > AND n_name = ? > GROUP BY > s_name > ORDER BY > numwait DESC, > s_name LIMIT 100{noformat} -- This message was sent by
[jira] [Updated] (IGNITE-22110) [SQL] TPC-H q16 query hangs with sc=0.1
[ https://issues.apache.org/jira/browse/IGNITE-22110?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Nikita Sivkov updated IGNITE-22110: --- Summary: [SQL] TPC-H q16 query hangs with sc=0.1 (was: CLONE - [SQL] TPC-H q16 query hangs with sc=0.1) > [SQL] TPC-H q16 query hangs with sc=0.1 > --- > > Key: IGNITE-22110 > URL: https://issues.apache.org/jira/browse/IGNITE-22110 > Project: Ignite > Issue Type: Bug > Components: sql >Reporter: Nikita Sivkov >Priority: Major > Labels: ignite-3, ignite3_performance > > Benchmark: > [https://github.com/cmu-db/benchbase/tree/main/src/main/java/com/oltpbenchmark/benchmarks/tpch] > > h1. Setup > * 1 server node > * TPC-H with scale factor = 0.1 > h1. Steps > # Start an Ignite node > # Run benchbase with {{-s 1 --create=true --load=true --execute=false}} to > preload data > # Observe via the benchbase log that the data was successfully loaded > # Run {{benchbase with -s 1 --create=false --load=false --execute=true}} to > run the benchmark > h1. Expected result > The benchmark finishes after warmup + duration time > h1. Actual result > The benchmark hangs for hours on > [Q16|https://github.com/cmu-db/benchbase/blob/main/src/main/java/com/oltpbenchmark/benchmarks/tpch/procedures/Q16.java] > query: > > {noformat} > SELECT > p_brand, > p_type, > p_size, > COUNT(DISTINCT ps_suppkey) AS supplier_cnt > FROM > partsupp, > part > WHERE > p_partkey = ps_partkey > AND p_brand <> ? > AND p_type NOT LIKE ? > AND p_size IN (?, ?, ?, ?, ?, ?, ?, ?) > AND ps_suppkey NOT IN > ( > SELECT > s_suppkey > FROM > supplier > WHERE > s_comment LIKE '%Customer%Complaints%' > ) > GROUP BY > p_brand, > p_type, > p_size > ORDER BY > supplier_cnt DESC, > p_brand, > p_type, > p_size{noformat} -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (IGNITE-22110) CLONE - [SQL] TPC-H q16 query hangs with sc=0.1
[ https://issues.apache.org/jira/browse/IGNITE-22110?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Nikita Sivkov updated IGNITE-22110: --- Description: Benchmark: [https://github.com/cmu-db/benchbase/tree/main/src/main/java/com/oltpbenchmark/benchmarks/tpch] h1. Setup * 1 server node * TPC-H with scale factor = 0.1 h1. Steps # Start an Ignite node # Run benchbase with {{-s 1 --create=true --load=true --execute=false}} to preload data # Observe via the benchbase log that the data was successfully loaded # Run {{benchbase with -s 1 --create=false --load=false --execute=true}} to run the benchmark h1. Expected result The benchmark finishes after warmup + duration time h1. Actual result The benchmark hangs for hours on [Q16|https://github.com/cmu-db/benchbase/blob/main/src/main/java/com/oltpbenchmark/benchmarks/tpch/procedures/Q16.java] query: {noformat} SELECT p_brand, p_type, p_size, COUNT(DISTINCT ps_suppkey) AS supplier_cnt FROM partsupp, part WHERE p_partkey = ps_partkey AND p_brand <> ? AND p_type NOT LIKE ? AND p_size IN (?, ?, ?, ?, ?, ?, ?, ?) AND ps_suppkey NOT IN ( SELECT s_suppkey FROM supplier WHERE s_comment LIKE '%Customer%Complaints%' ) GROUP BY p_brand, p_type, p_size ORDER BY supplier_cnt DESC, p_brand, p_type, p_size{noformat} was: Benchmark: [https://github.com/cmu-db/benchbase/tree/main/src/main/java/com/oltpbenchmark/benchmarks/tpch] h1. Setup * 1 server node * TPC-H with scale factor = 0.1 h1. Steps # Start an Ignite node # Run benchbase with {{-s 1 --create=true --load=true --execute=false}} to preload data # Observe via the benchbase log that the data was successfully loaded # Run {{benchbase with -s 1 --create=false --load=false --execute=true}} to run the benchmark h1. Expected result The benchmark finishes after warmup + duration time h1. Actual result The benchmark hangs for hours on [Q4|https://github.com/cmu-db/benchbase/blob/main/src/main/java/com/oltpbenchmark/benchmarks/tpch/procedures/Q4.java] query: {noformat} SELECT o_orderpriority, COUNT(*) AS order_count FROM orders WHERE o_orderdate >= DATE ? AND o_orderdate < DATE ? + INTERVAL '3' MONTH AND EXISTS ( SELECT * FROM lineitem WHERE l_orderkey = o_orderkey AND l_commitdate < l_receiptdate ) GROUP BY o_orderpriority ORDER BY o_orderpriority {noformat} > CLONE - [SQL] TPC-H q16 query hangs with sc=0.1 > --- > > Key: IGNITE-22110 > URL: https://issues.apache.org/jira/browse/IGNITE-22110 > Project: Ignite > Issue Type: Bug > Components: sql >Reporter: Nikita Sivkov >Priority: Major > Labels: ignite-3, ignite3_performance > > Benchmark: > [https://github.com/cmu-db/benchbase/tree/main/src/main/java/com/oltpbenchmark/benchmarks/tpch] > > h1. Setup > * 1 server node > * TPC-H with scale factor = 0.1 > h1. Steps > # Start an Ignite node > # Run benchbase with {{-s 1 --create=true --load=true --execute=false}} to > preload data > # Observe via the benchbase log that the data was successfully loaded > # Run {{benchbase with -s 1 --create=false --load=false --execute=true}} to > run the benchmark > h1. Expected result > The benchmark finishes after warmup + duration time > h1. Actual result > The benchmark hangs for hours on > [Q16|https://github.com/cmu-db/benchbase/blob/main/src/main/java/com/oltpbenchmark/benchmarks/tpch/procedures/Q16.java] > query: > > {noformat} > SELECT > p_brand, > p_type, > p_size, > COUNT(DISTINCT ps_suppkey) AS supplier_cnt > FROM > partsupp, > part > WHERE > p_partkey = ps_partkey > AND p_brand <> ? > AND p_type NOT LIKE ? > AND p_size IN (?, ?, ?, ?, ?, ?, ?, ?) > AND ps_suppkey NOT IN > ( > SELECT > s_suppkey > FROM > supplier > WHERE > s_comment LIKE '%Customer%Complaints%' > ) > GROUP BY > p_brand, > p_type, > p_size > ORDER BY > supplier_cnt DESC, > p_brand, > p_type, > p_size{noformat} -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Created] (IGNITE-22111) [SQL] TPC-H q21 query hangs with sc=0.1
Nikita Sivkov created IGNITE-22111: -- Summary: [SQL] TPC-H q21 query hangs with sc=0.1 Key: IGNITE-22111 URL: https://issues.apache.org/jira/browse/IGNITE-22111 Project: Ignite Issue Type: Bug Components: sql Reporter: Nikita Sivkov Benchmark: [https://github.com/cmu-db/benchbase/tree/main/src/main/java/com/oltpbenchmark/benchmarks/tpch] h1. Setup * 1 server node * TPC-H with scale factor = 0.1 h1. Steps # Start an Ignite node # Run benchbase with {{-s 1 --create=true --load=true --execute=false}} to preload data # Observe via the benchbase log that the data was successfully loaded # Run {{benchbase with -s 1 --create=false --load=false --execute=true}} to run the benchmark h1. Expected result The benchmark finishes after warmup + duration time h1. Actual result The benchmark hangs for hours on [Q16|https://github.com/cmu-db/benchbase/blob/main/src/main/java/com/oltpbenchmark/benchmarks/tpch/procedures/Q16.java] query: {noformat} SELECT p_brand, p_type, p_size, COUNT(DISTINCT ps_suppkey) AS supplier_cnt FROM partsupp, part WHERE p_partkey = ps_partkey AND p_brand <> ? AND p_type NOT LIKE ? AND p_size IN (?, ?, ?, ?, ?, ?, ?, ?) AND ps_suppkey NOT IN ( SELECT s_suppkey FROM supplier WHERE s_comment LIKE '%Customer%Complaints%' ) GROUP BY p_brand, p_type, p_size ORDER BY supplier_cnt DESC, p_brand, p_type, p_size{noformat} -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Created] (IGNITE-22110) CLONE - [SQL] TPC-H q16 query hangs with sc=0.1
Nikita Sivkov created IGNITE-22110: -- Summary: CLONE - [SQL] TPC-H q16 query hangs with sc=0.1 Key: IGNITE-22110 URL: https://issues.apache.org/jira/browse/IGNITE-22110 Project: Ignite Issue Type: Bug Components: sql Reporter: Nikita Sivkov Benchmark: [https://github.com/cmu-db/benchbase/tree/main/src/main/java/com/oltpbenchmark/benchmarks/tpch] h1. Setup * 1 server node * TPC-H with scale factor = 0.1 h1. Steps # Start an Ignite node # Run benchbase with {{-s 1 --create=true --load=true --execute=false}} to preload data # Observe via the benchbase log that the data was successfully loaded # Run {{benchbase with -s 1 --create=false --load=false --execute=true}} to run the benchmark h1. Expected result The benchmark finishes after warmup + duration time h1. Actual result The benchmark hangs for hours on [Q4|https://github.com/cmu-db/benchbase/blob/main/src/main/java/com/oltpbenchmark/benchmarks/tpch/procedures/Q4.java] query: {noformat} SELECT o_orderpriority, COUNT(*) AS order_count FROM orders WHERE o_orderdate >= DATE ? AND o_orderdate < DATE ? + INTERVAL '3' MONTH AND EXISTS ( SELECT * FROM lineitem WHERE l_orderkey = o_orderkey AND l_commitdate < l_receiptdate ) GROUP BY o_orderpriority ORDER BY o_orderpriority {noformat} -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Created] (IGNITE-22109) [SQL] TPC-H q4 query hangs with sc=0.1
Nikita Sivkov created IGNITE-22109: -- Summary: [SQL] TPC-H q4 query hangs with sc=0.1 Key: IGNITE-22109 URL: https://issues.apache.org/jira/browse/IGNITE-22109 Project: Ignite Issue Type: Bug Components: sql Reporter: Nikita Sivkov Benchmark: [https://github.com/cmu-db/benchbase/tree/main/src/main/java/com/oltpbenchmark/benchmarks/tpch] h1. Setup * 1 server node * TPC-H with scale factor = 0.1 h1. Steps # Start an Ignite node # Run benchbase with {{-s 1 --create=true --load=true --execute=false}} to preload data # Observe via the benchbase log that the data was successfully loaded # Run {{benchbase with -s 1 --create=false --load=false --execute=true}} to run the benchmark h1. Expected result The benchmark finishes after warmup + duration time h1. Actual result The benchmark hangs for hours on [Q4|https://github.com/cmu-db/benchbase/blob/main/src/main/java/com/oltpbenchmark/benchmarks/tpch/procedures/Q4.java] query: {noformat} SELECT o_orderpriority, COUNT(*) AS order_count FROM orders WHERE o_orderdate >= DATE ? AND o_orderdate < DATE ? + INTERVAL '3' MONTH AND EXISTS ( SELECT * FROM lineitem WHERE l_orderkey = o_orderkey AND l_commitdate < l_receiptdate ) GROUP BY o_orderpriority ORDER BY o_orderpriority {noformat} -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (IGNITE-22006) 'Failed to process the lease granted message' error under load with balance transfer scenario
[ https://issues.apache.org/jira/browse/IGNITE-22006?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Nikita Sivkov updated IGNITE-22006: --- Description: *Steps to reproduce:* Perform a long (about 2 hours) load test with a balance transfer scenario (see scenario pseudo code in attachments). *Expected result:* No errors happen. *Actual result:* Get error in server logs - {{Failed to process the lease granted message}} {code:java} 2024-04-05 17:50:39:180 +0300 [WARNING][%poc-tester-SERVER-192.168.1.97-id-0%JRaft-Request-Processor-13][NodeImpl] Node <127_part_16/poc-tester-SERVER-192.168.1.97-id-0> is not in active state, currTerm=3. 2024-04-05 17:50:39:187 +0300 [WARNING][CompletableFutureDelayScheduler][ReplicaManager] Failed to process the lease granted message [msg=LeaseGrantedMessageImpl [force=true, groupId=77_part_14, leaseExpirationTimeLong=112219169697759232, leaseStartTimeLong=112219161833439373]]. java.util.concurrent.TimeoutException at java.base/java.util.concurrent.CompletableFuture$Timeout.run(CompletableFuture.java:2792) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:304) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) at java.base/java.lang.Thread.run(Thread.java:829) 2024-04-05 17:50:39:190 +0300 [WARNING][%poc-tester-SERVER-192.168.1.97-id-0%JRaft-Request-Processor-34][NodeImpl] Node <213_part_14/poc-tester-SERVER-192.168.1.97-id-0> is not in active state, currTerm=2.{code} was: *Steps to reproduce:* Perform a long (about 2 hours) load test with a balance transfer scenario (see scenario pseudo code in attachments). *Expected result:* No errors happen. *Actual result:* Get error in server logs - {{Failed to process replica request}} {code:java} 2024-04-05 17:50:55:802 +0300 [WARNING][%poc-tester-SERVER-192.168.1.97-id-0%JRaft-AppendEntries-Processor-2][NodeImpl] Node <193_part_15/poc-tester-SERVER-192.168.1.97-id-0> is not in active state, currTerm=2. 2024-04-05 17:50:55:805 +0300 [WARNING][%poc-tester-SERVER-192.168.1.97-id-0%Raft-Group-Client-19][ReplicaManager] Failed to process replica request [request=TxFinishReplicaRequestImpl [commit=false, commitTimestampLong=0, enlistmentConsistencyToken=112218720633356321, groupId=123_part_21, groups=HashMap {141_part_13=poc-tester-SERVER-192.168.1.27-id-0, 139_part_9=poc-tester-SERVER-192.168.1.97-id-0, 193_part_3=poc-tester-SERVER-192.168.1.27-id-0, 19_part_23=poc-tester-SERVER-192.168.1.27-id-0, 117_part_17=poc-tester-SERVER-192.168.1.18-id-0, 45_part_9=poc-tester-SERVER-192.168.1.18-id-0, 39_part_3=poc-tester-SERVER-192.168.1.18-id-0, 77_part_4=poc-tester-SERVER-192.168.1.18-id-0, 105_part_4=poc-tester-SERVER-192.168.1.18-id-0, 123_part_21=poc-tester-SERVER-192.168.1.97-id-0, 103_part_9=poc-tester-SERVER-192.168.1.18-id-0, 161_part_15=poc-tester-SERVER-192.168.1.27-id-0, 103_part_22=poc-tester-SERVER-192.168.1.27-id-0, 89_part_10=poc-tester-SERVER-192.168.1.18-id-0, 39_part_19=poc-tester-SERVER-192.168.1.27-id-0, 149_part_13=poc-tester-SERVER-192.168.1.27-id-0, 97_part_24=poc-tester-SERVER-192.168.1.97-id-0, 83_part_9=poc-tester-SERVER-192.168.1.27-id-0, 209_part_10=poc-tester-SERVER-192.168.1.27-id-0, 185_part_5=poc-tester-SERVER-192.168.1.18-id-0, 117_part_9=poc-tester-SERVER-192.168.1.27-id-0, 105_part_22=poc-tester-SERVER-192.168.1.18-id-0}, timestampLong=112219170129903617, txId=018eaebd-88ba-0001-606d-62250001]]. java.util.concurrent.CompletionException: org.apache.ignite.tx.TransactionException: IGN-TX-7 TraceId:cb1577e6-ec35-47f0-ab7d-56a0687344ed java.util.concurrent.TimeoutException at java.base/java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:314) at java.base/java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:319) at java.base/java.util.concurrent.CompletableFuture.uniHandle(CompletableFuture.java:932) at java.base/java.util.concurrent.CompletableFuture$UniHandle.tryFire(CompletableFuture.java:907) at java.base/java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506) at java.base/java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:2088) at org.apache.ignite.internal.table.distributed.replicator.PartitionReplicaListener.lambda$applyCmdWithRetryOnSafeTimeReorderException$126(PartitionReplicaListener.java:2806) at java.base/java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:859) at
[jira] [Created] (IGNITE-22006) 'Failed to process the lease granted message' error under load with balance transfer scenario
Nikita Sivkov created IGNITE-22006: -- Summary: 'Failed to process the lease granted message' error under load with balance transfer scenario Key: IGNITE-22006 URL: https://issues.apache.org/jira/browse/IGNITE-22006 Project: Ignite Issue Type: Bug Affects Versions: 3.0.0-beta2 Environment: Cluster of 3 nodes Reporter: Nikita Sivkov Attachments: transfer_ign3.yaml *Steps to reproduce:* Perform a long (about 2 hours) load test with a balance transfer scenario (see scenario pseudo code in attachments). *Expected result:* No errors happen. *Actual result:* Get error in server logs - {{Failed to process replica request}} {code:java} 2024-04-05 17:50:55:802 +0300 [WARNING][%poc-tester-SERVER-192.168.1.97-id-0%JRaft-AppendEntries-Processor-2][NodeImpl] Node <193_part_15/poc-tester-SERVER-192.168.1.97-id-0> is not in active state, currTerm=2. 2024-04-05 17:50:55:805 +0300 [WARNING][%poc-tester-SERVER-192.168.1.97-id-0%Raft-Group-Client-19][ReplicaManager] Failed to process replica request [request=TxFinishReplicaRequestImpl [commit=false, commitTimestampLong=0, enlistmentConsistencyToken=112218720633356321, groupId=123_part_21, groups=HashMap {141_part_13=poc-tester-SERVER-192.168.1.27-id-0, 139_part_9=poc-tester-SERVER-192.168.1.97-id-0, 193_part_3=poc-tester-SERVER-192.168.1.27-id-0, 19_part_23=poc-tester-SERVER-192.168.1.27-id-0, 117_part_17=poc-tester-SERVER-192.168.1.18-id-0, 45_part_9=poc-tester-SERVER-192.168.1.18-id-0, 39_part_3=poc-tester-SERVER-192.168.1.18-id-0, 77_part_4=poc-tester-SERVER-192.168.1.18-id-0, 105_part_4=poc-tester-SERVER-192.168.1.18-id-0, 123_part_21=poc-tester-SERVER-192.168.1.97-id-0, 103_part_9=poc-tester-SERVER-192.168.1.18-id-0, 161_part_15=poc-tester-SERVER-192.168.1.27-id-0, 103_part_22=poc-tester-SERVER-192.168.1.27-id-0, 89_part_10=poc-tester-SERVER-192.168.1.18-id-0, 39_part_19=poc-tester-SERVER-192.168.1.27-id-0, 149_part_13=poc-tester-SERVER-192.168.1.27-id-0, 97_part_24=poc-tester-SERVER-192.168.1.97-id-0, 83_part_9=poc-tester-SERVER-192.168.1.27-id-0, 209_part_10=poc-tester-SERVER-192.168.1.27-id-0, 185_part_5=poc-tester-SERVER-192.168.1.18-id-0, 117_part_9=poc-tester-SERVER-192.168.1.27-id-0, 105_part_22=poc-tester-SERVER-192.168.1.18-id-0}, timestampLong=112219170129903617, txId=018eaebd-88ba-0001-606d-62250001]]. java.util.concurrent.CompletionException: org.apache.ignite.tx.TransactionException: IGN-TX-7 TraceId:cb1577e6-ec35-47f0-ab7d-56a0687344ed java.util.concurrent.TimeoutException at java.base/java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:314) at java.base/java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:319) at java.base/java.util.concurrent.CompletableFuture.uniHandle(CompletableFuture.java:932) at java.base/java.util.concurrent.CompletableFuture$UniHandle.tryFire(CompletableFuture.java:907) at java.base/java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506) at java.base/java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:2088) at org.apache.ignite.internal.table.distributed.replicator.PartitionReplicaListener.lambda$applyCmdWithRetryOnSafeTimeReorderException$126(PartitionReplicaListener.java:2806) at java.base/java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:859) at java.base/java.util.concurrent.CompletableFuture$UniWhenComplete.tryFire(CompletableFuture.java:837) at java.base/java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506) at java.base/java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:2088) at org.apache.ignite.internal.raft.RaftGroupServiceImpl.sendWithRetry(RaftGroupServiceImpl.java:550) at org.apache.ignite.internal.raft.RaftGroupServiceImpl.lambda$handleErrorResponse$44(RaftGroupServiceImpl.java:653) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:304) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) at java.base/java.lang.Thread.run(Thread.java:829) Caused by: org.apache.ignite.tx.TransactionException: IGN-TX-7 TraceId:cb1577e6-ec35-47f0-ab7d-56a0687344ed java.util.concurrent.TimeoutException at org.apache.ignite.internal.table.distributed.replicator.PartitionReplicaListener.lambda$finishTransaction$70(PartitionReplicaListener.java:1867) at
[jira] [Updated] (IGNITE-22005) 'Failed to process replica request' error under load with balance transfer scenario
[ https://issues.apache.org/jira/browse/IGNITE-22005?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Nikita Sivkov updated IGNITE-22005: --- Description: *Steps to reproduce:* Perform a long (about 2 hours) load test with a balance transfer scenario (see scenario pseudo code in attachments). *Expected result:* No errors happen. *Actual result:* Get error in server logs - {{Failed to process replica request}} {code:java} 2024-04-05 17:50:55:802 +0300 [WARNING][%poc-tester-SERVER-192.168.1.97-id-0%JRaft-AppendEntries-Processor-2][NodeImpl] Node <193_part_15/poc-tester-SERVER-192.168.1.97-id-0> is not in active state, currTerm=2. 2024-04-05 17:50:55:805 +0300 [WARNING][%poc-tester-SERVER-192.168.1.97-id-0%Raft-Group-Client-19][ReplicaManager] Failed to process replica request [request=TxFinishReplicaRequestImpl [commit=false, commitTimestampLong=0, enlistmentConsistencyToken=112218720633356321, groupId=123_part_21, groups=HashMap {141_part_13=poc-tester-SERVER-192.168.1.27-id-0, 139_part_9=poc-tester-SERVER-192.168.1.97-id-0, 193_part_3=poc-tester-SERVER-192.168.1.27-id-0, 19_part_23=poc-tester-SERVER-192.168.1.27-id-0, 117_part_17=poc-tester-SERVER-192.168.1.18-id-0, 45_part_9=poc-tester-SERVER-192.168.1.18-id-0, 39_part_3=poc-tester-SERVER-192.168.1.18-id-0, 77_part_4=poc-tester-SERVER-192.168.1.18-id-0, 105_part_4=poc-tester-SERVER-192.168.1.18-id-0, 123_part_21=poc-tester-SERVER-192.168.1.97-id-0, 103_part_9=poc-tester-SERVER-192.168.1.18-id-0, 161_part_15=poc-tester-SERVER-192.168.1.27-id-0, 103_part_22=poc-tester-SERVER-192.168.1.27-id-0, 89_part_10=poc-tester-SERVER-192.168.1.18-id-0, 39_part_19=poc-tester-SERVER-192.168.1.27-id-0, 149_part_13=poc-tester-SERVER-192.168.1.27-id-0, 97_part_24=poc-tester-SERVER-192.168.1.97-id-0, 83_part_9=poc-tester-SERVER-192.168.1.27-id-0, 209_part_10=poc-tester-SERVER-192.168.1.27-id-0, 185_part_5=poc-tester-SERVER-192.168.1.18-id-0, 117_part_9=poc-tester-SERVER-192.168.1.27-id-0, 105_part_22=poc-tester-SERVER-192.168.1.18-id-0}, timestampLong=112219170129903617, txId=018eaebd-88ba-0001-606d-62250001]]. java.util.concurrent.CompletionException: org.apache.ignite.tx.TransactionException: IGN-TX-7 TraceId:cb1577e6-ec35-47f0-ab7d-56a0687344ed java.util.concurrent.TimeoutException at java.base/java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:314) at java.base/java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:319) at java.base/java.util.concurrent.CompletableFuture.uniHandle(CompletableFuture.java:932) at java.base/java.util.concurrent.CompletableFuture$UniHandle.tryFire(CompletableFuture.java:907) at java.base/java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506) at java.base/java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:2088) at org.apache.ignite.internal.table.distributed.replicator.PartitionReplicaListener.lambda$applyCmdWithRetryOnSafeTimeReorderException$126(PartitionReplicaListener.java:2806) at java.base/java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:859) at java.base/java.util.concurrent.CompletableFuture$UniWhenComplete.tryFire(CompletableFuture.java:837) at java.base/java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506) at java.base/java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:2088) at org.apache.ignite.internal.raft.RaftGroupServiceImpl.sendWithRetry(RaftGroupServiceImpl.java:550) at org.apache.ignite.internal.raft.RaftGroupServiceImpl.lambda$handleErrorResponse$44(RaftGroupServiceImpl.java:653) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:304) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) at java.base/java.lang.Thread.run(Thread.java:829) Caused by: org.apache.ignite.tx.TransactionException: IGN-TX-7 TraceId:cb1577e6-ec35-47f0-ab7d-56a0687344ed java.util.concurrent.TimeoutException at org.apache.ignite.internal.table.distributed.replicator.PartitionReplicaListener.lambda$finishTransaction$70(PartitionReplicaListener.java:1867) at java.base/java.util.concurrent.CompletableFuture.uniHandle(CompletableFuture.java:930) ... 16 more Caused by: java.util.concurrent.CompletionException: java.util.concurrent.TimeoutException at java.base/java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:331) at
[jira] [Created] (IGNITE-22005) 'Failed to process replica request' error under load with balance transfer scenario
Nikita Sivkov created IGNITE-22005: -- Summary: 'Failed to process replica request' error under load with balance transfer scenario Key: IGNITE-22005 URL: https://issues.apache.org/jira/browse/IGNITE-22005 Project: Ignite Issue Type: Bug Affects Versions: 3.0.0-beta2 Environment: Cluster of 3 nodes Reporter: Nikita Sivkov Attachments: transfer_ign3.yaml *Steps to reproduce:* Perform a long (about 2 hours) load test with a balance transfer scenario (see scenario pseudo code in attachments). *Expected result:* No errors happen. *Actual result:* Get error in server logs - {{Failed to process delayed response}} {code:java} 2024-04-05 17:50:50:776 +0300 [WARNING][%poc-tester-SERVER-192.168.1.97-id-0%JRaft-Request-Processor-1][NodeImpl] Node <27_part_23/poc-tester-SERVER-192.168.1.97-id-0> is not in active state, currTerm=2. 2024-04-05 17:50:50:778 +0300 [WARNING][%poc-tester-SERVER-192.168.1.97-id-0%Raft-Group-Client-5][ReplicaManager] Failed to process delayed response [request=ReadWriteSingleRowReplicaRequestImpl [commitPartitionId=TablePartitionIdMessageImpl [partitionId=21, tableId=123], coordinatorId=3de6f999-7ab9-4405-aff0-ee0c7e4886ce, enlistmentConsistencyToken=112218720633356321, full=false, groupId=123_part_21, requestType=RW_UPSERT, schemaVersion=1, timestampLong=112219169796915211, transactionId=018eaebd-88ba-0001-606d-62250001]] java.util.concurrent.CompletionException: java.util.concurrent.TimeoutException at java.base/java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:331) at java.base/java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:346) at java.base/java.util.concurrent.CompletableFuture$UniApply.tryFire(CompletableFuture.java:632) at java.base/java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506) at java.base/java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:2088) at org.apache.ignite.internal.raft.RaftGroupServiceImpl.sendWithRetry(RaftGroupServiceImpl.java:550) at org.apache.ignite.internal.raft.RaftGroupServiceImpl.lambda$handleErrorResponse$44(RaftGroupServiceImpl.java:653) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:304) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) at java.base/java.lang.Thread.run(Thread.java:829) Caused by: java.util.concurrent.TimeoutException ... 8 more 2024-04-05 17:50:50:780 +0300 [WARNING][%poc-tester-SERVER-192.168.1.97-id-0%JRaft-Request-Processor-27][NodeImpl] Node <99_part_6/poc-tester-SERVER-192.168.1.97-id-0> is not in active state, currTerm=3. {code} -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (IGNITE-22004) 'Failed to process delayed response' error under load with balance transfer scenario
[ https://issues.apache.org/jira/browse/IGNITE-22004?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Nikita Sivkov updated IGNITE-22004: --- Attachment: transfer_ign3.yaml > 'Failed to process delayed response' error under load with balance transfer > scenario > > > Key: IGNITE-22004 > URL: https://issues.apache.org/jira/browse/IGNITE-22004 > Project: Ignite > Issue Type: Bug >Affects Versions: 3.0.0-beta2 > Environment: Cluster of 3 nodes >Reporter: Nikita Sivkov >Priority: Major > Labels: ignite-3 > Attachments: transfer_ign3.yaml > > > *Steps to reproduce:* > Perform a long (about 2 hours) load test with a balance transfer scenario > (see scenario pseudo code in attachments). > *Expected result:* > No errors happen. > *Actual result:* > Get error in server logs - {{Failed to process delayed response}} > {code:java} > 2024-04-05 17:50:50:776 +0300 > [WARNING][%poc-tester-SERVER-192.168.1.97-id-0%JRaft-Request-Processor-1][NodeImpl] > Node <27_part_23/poc-tester-SERVER-192.168.1.97-id-0> is not in active > state, currTerm=2. > 2024-04-05 17:50:50:778 +0300 > [WARNING][%poc-tester-SERVER-192.168.1.97-id-0%Raft-Group-Client-5][ReplicaManager] > Failed to process delayed response > [request=ReadWriteSingleRowReplicaRequestImpl > [commitPartitionId=TablePartitionIdMessageImpl [partitionId=21, tableId=123], > coordinatorId=3de6f999-7ab9-4405-aff0-ee0c7e4886ce, > enlistmentConsistencyToken=112218720633356321, full=false, > groupId=123_part_21, requestType=RW_UPSERT, schemaVersion=1, > timestampLong=112219169796915211, > transactionId=018eaebd-88ba-0001-606d-62250001]] > java.util.concurrent.CompletionException: > java.util.concurrent.TimeoutException > at > java.base/java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:331) > at > java.base/java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:346) > at > java.base/java.util.concurrent.CompletableFuture$UniApply.tryFire(CompletableFuture.java:632) > at > java.base/java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506) > at > java.base/java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:2088) > at > org.apache.ignite.internal.raft.RaftGroupServiceImpl.sendWithRetry(RaftGroupServiceImpl.java:550) > at > org.apache.ignite.internal.raft.RaftGroupServiceImpl.lambda$handleErrorResponse$44(RaftGroupServiceImpl.java:653) > at > java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515) > at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) > at > java.base/java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:304) > at > java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) > at > java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) > at java.base/java.lang.Thread.run(Thread.java:829) > Caused by: java.util.concurrent.TimeoutException > ... 8 more > 2024-04-05 17:50:50:780 +0300 > [WARNING][%poc-tester-SERVER-192.168.1.97-id-0%JRaft-Request-Processor-27][NodeImpl] > Node <99_part_6/poc-tester-SERVER-192.168.1.97-id-0> is not in active state, > currTerm=3. {code} -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (IGNITE-22004) 'Failed to process delayed response' error under load with balance transfer scenario
[ https://issues.apache.org/jira/browse/IGNITE-22004?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Nikita Sivkov updated IGNITE-22004: --- Description: *Steps to reproduce:* Perform a long (about 2 hours) load test with a balance transfer scenario (see scenario pseudo code in attachments). *Expected result:* No errors happen. *Actual result:* Get error in server logs - {{Failed to process delayed response}} {code:java} 2024-04-05 17:50:50:776 +0300 [WARNING][%poc-tester-SERVER-192.168.1.97-id-0%JRaft-Request-Processor-1][NodeImpl] Node <27_part_23/poc-tester-SERVER-192.168.1.97-id-0> is not in active state, currTerm=2. 2024-04-05 17:50:50:778 +0300 [WARNING][%poc-tester-SERVER-192.168.1.97-id-0%Raft-Group-Client-5][ReplicaManager] Failed to process delayed response [request=ReadWriteSingleRowReplicaRequestImpl [commitPartitionId=TablePartitionIdMessageImpl [partitionId=21, tableId=123], coordinatorId=3de6f999-7ab9-4405-aff0-ee0c7e4886ce, enlistmentConsistencyToken=112218720633356321, full=false, groupId=123_part_21, requestType=RW_UPSERT, schemaVersion=1, timestampLong=112219169796915211, transactionId=018eaebd-88ba-0001-606d-62250001]] java.util.concurrent.CompletionException: java.util.concurrent.TimeoutException at java.base/java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:331) at java.base/java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:346) at java.base/java.util.concurrent.CompletableFuture$UniApply.tryFire(CompletableFuture.java:632) at java.base/java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506) at java.base/java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:2088) at org.apache.ignite.internal.raft.RaftGroupServiceImpl.sendWithRetry(RaftGroupServiceImpl.java:550) at org.apache.ignite.internal.raft.RaftGroupServiceImpl.lambda$handleErrorResponse$44(RaftGroupServiceImpl.java:653) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:304) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) at java.base/java.lang.Thread.run(Thread.java:829) Caused by: java.util.concurrent.TimeoutException ... 8 more 2024-04-05 17:50:50:780 +0300 [WARNING][%poc-tester-SERVER-192.168.1.97-id-0%JRaft-Request-Processor-27][NodeImpl] Node <99_part_6/poc-tester-SERVER-192.168.1.97-id-0> is not in active state, currTerm=3. {code} was: *Steps to reproduce:* Perform a long (about 2 hours) load test with a balance transfer scenario. *Expected result:* No errors happen. *Actual result:* Get error in server logs - {{Failed to process delayed response}} {code:java} 2024-04-05 17:50:50:776 +0300 [WARNING][%poc-tester-SERVER-192.168.1.97-id-0%JRaft-Request-Processor-1][NodeImpl] Node <27_part_23/poc-tester-SERVER-192.168.1.97-id-0> is not in active state, currTerm=2. 2024-04-05 17:50:50:778 +0300 [WARNING][%poc-tester-SERVER-192.168.1.97-id-0%Raft-Group-Client-5][ReplicaManager] Failed to process delayed response [request=ReadWriteSingleRowReplicaRequestImpl [commitPartitionId=TablePartitionIdMessageImpl [partitionId=21, tableId=123], coordinatorId=3de6f999-7ab9-4405-aff0-ee0c7e4886ce, enlistmentConsistencyToken=112218720633356321, full=false, groupId=123_part_21, requestType=RW_UPSERT, schemaVersion=1, timestampLong=112219169796915211, transactionId=018eaebd-88ba-0001-606d-62250001]] java.util.concurrent.CompletionException: java.util.concurrent.TimeoutException at java.base/java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:331) at java.base/java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:346) at java.base/java.util.concurrent.CompletableFuture$UniApply.tryFire(CompletableFuture.java:632) at java.base/java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506) at java.base/java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:2088) at org.apache.ignite.internal.raft.RaftGroupServiceImpl.sendWithRetry(RaftGroupServiceImpl.java:550) at org.apache.ignite.internal.raft.RaftGroupServiceImpl.lambda$handleErrorResponse$44(RaftGroupServiceImpl.java:653) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:304) at
[jira] [Created] (IGNITE-22004) 'Failed to process delayed response' error under load with balance transfer scenario
Nikita Sivkov created IGNITE-22004: -- Summary: 'Failed to process delayed response' error under load with balance transfer scenario Key: IGNITE-22004 URL: https://issues.apache.org/jira/browse/IGNITE-22004 Project: Ignite Issue Type: Bug Affects Versions: 3.0.0-beta2 Environment: Cluster of 3 nodes Reporter: Nikita Sivkov *Steps to reproduce:* Perform a long (about 2 hours) load test with a balance transfer scenario. *Expected result:* No errors happen. *Actual result:* Get error in server logs - {{Failed to process delayed response}} {code:java} 2024-04-05 17:50:50:776 +0300 [WARNING][%poc-tester-SERVER-192.168.1.97-id-0%JRaft-Request-Processor-1][NodeImpl] Node <27_part_23/poc-tester-SERVER-192.168.1.97-id-0> is not in active state, currTerm=2. 2024-04-05 17:50:50:778 +0300 [WARNING][%poc-tester-SERVER-192.168.1.97-id-0%Raft-Group-Client-5][ReplicaManager] Failed to process delayed response [request=ReadWriteSingleRowReplicaRequestImpl [commitPartitionId=TablePartitionIdMessageImpl [partitionId=21, tableId=123], coordinatorId=3de6f999-7ab9-4405-aff0-ee0c7e4886ce, enlistmentConsistencyToken=112218720633356321, full=false, groupId=123_part_21, requestType=RW_UPSERT, schemaVersion=1, timestampLong=112219169796915211, transactionId=018eaebd-88ba-0001-606d-62250001]] java.util.concurrent.CompletionException: java.util.concurrent.TimeoutException at java.base/java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:331) at java.base/java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:346) at java.base/java.util.concurrent.CompletableFuture$UniApply.tryFire(CompletableFuture.java:632) at java.base/java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506) at java.base/java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:2088) at org.apache.ignite.internal.raft.RaftGroupServiceImpl.sendWithRetry(RaftGroupServiceImpl.java:550) at org.apache.ignite.internal.raft.RaftGroupServiceImpl.lambda$handleErrorResponse$44(RaftGroupServiceImpl.java:653) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:304) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) at java.base/java.lang.Thread.run(Thread.java:829) Caused by: java.util.concurrent.TimeoutException ... 8 more 2024-04-05 17:50:50:780 +0300 [WARNING][%poc-tester-SERVER-192.168.1.97-id-0%JRaft-Request-Processor-27][NodeImpl] Node <99_part_6/poc-tester-SERVER-192.168.1.97-id-0> is not in active state, currTerm=3. {code} -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Resolved] (IGNITE-21777) 'Failed to get the primary replica' or 'Replication is timed out' or hangs with 'aimem' storage engine
[ https://issues.apache.org/jira/browse/IGNITE-21777?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Nikita Sivkov resolved IGNITE-21777. Resolution: Cannot Reproduce > 'Failed to get the primary replica' or 'Replication is timed out' or hangs > with 'aimem' storage engine > -- > > Key: IGNITE-21777 > URL: https://issues.apache.org/jira/browse/IGNITE-21777 > Project: Ignite > Issue Type: Bug >Affects Versions: 3.0.0-beta2 > Environment: Cluster of 2 nodes. > Storage engine - aimem. >Reporter: Nikita Sivkov >Priority: Major > Labels: ignite-3 > > *Steps to reproduce:* > # Create a cluster with 2 nodes. > # Connect via JDBC. > # Repeat the following SQL statements in a loop (for example, 100 times): > ** {{drop table if exists tags}} > ** {{create zone if not exists "AIMEM" engine aimem}} > ** {{create table tags(tagId integer not null, tag varchar(100) not null, > primary key (tagId)) with PRIMARY_ZONE='AIMEM'}} > ** {{insert into tags(tagId, tag) values (1,'unit'), (2,'integration'), > (3,'smoke'), (4,'sanity'), (5,'regression')}} > *Expected result:* > No errors or hangs happen. > *Actual result:* > Hangs on {{Create table}} or {{Insert into}} statement. > _*OR*_ > Get the error {{Replication is timed out}} > {code:java} > Replication is timed out [replicaGrpId=34_part_16] > java.sql.SQLException: Replication is timed out [replicaGrpId=34_part_16] > at > org.apache.ignite.internal.jdbc.proto.IgniteQueryErrorCode.createJdbcSqlException(IgniteQueryErrorCode.java:57) > at > org.apache.ignite.internal.jdbc.JdbcStatement.execute0(JdbcStatement.java:154) > at > org.apache.ignite.internal.jdbc.JdbcStatement.executeUpdate(JdbcStatement.java:181) > at > org.gridgain.ai3tests.tests.teststeps.JdbcSteps.executeUpdateQuery(JdbcSteps.java:116) > at > org.gridgain.ai3tests.tests.UpdateTests.wannaCatchTheBug(UpdateTests.java:95) > at java.base/java.lang.reflect.Method.invoke(Method.java:566) > at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) > at > java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) > at > java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) > at java.base/java.lang.Thread.run(Thread.java:834) {code} > _*OR*_ > Get the error {{Failed to get the primary replica}} > {code:java} > java.sql.SQLException: Failed to get the primary replica > [tablePartitionId=18_part_1] > at > org.apache.ignite.internal.jdbc.proto.IgniteQueryErrorCode.createJdbcSqlException(IgniteQueryErrorCode.java:57) > at > org.apache.ignite.internal.jdbc.JdbcStatement.execute0(JdbcStatement.java:154) > at > org.apache.ignite.internal.jdbc.JdbcStatement.executeUpdate(JdbcStatement.java:181) > at > org.gridgain.ai3tests.tests.teststeps.JdbcSteps.executeUpdateQuery(JdbcSteps.java:116) > at > org.gridgain.ai3tests.tests.UpdateTests.createAndFillTables(UpdateTests.java:247) > at io.qameta.allure.Allure.lambda$step$0(Allure.java:113) > at io.qameta.allure.Allure.lambda$step$1(Allure.java:127) > at io.qameta.allure.Allure.step(Allure.java:181) > at io.qameta.allure.Allure.step(Allure.java:125) > at io.qameta.allure.Allure.step(Allure.java:112) > at > org.gridgain.ai3tests.tests.UpdateTests.updateTableWithConditionThatHasLinkedInnerSubQueries(UpdateTests.java:147) > at java.base/java.lang.reflect.Method.invoke(Method.java:566) > at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) > at > java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) > at > java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) > at java.base/java.lang.Thread.run(Thread.java:834) {code} -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Resolved] (IGNITE-21451) RocksDB: repeat of create table and drop column leads to freeze of client
[ https://issues.apache.org/jira/browse/IGNITE-21451?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Nikita Sivkov resolved IGNITE-21451. Resolution: Cannot Reproduce Didn't reproduce on the recent builds. > RocksDB: repeat of create table and drop column leads to freeze of client > - > > Key: IGNITE-21451 > URL: https://issues.apache.org/jira/browse/IGNITE-21451 > Project: Ignite > Issue Type: Bug > Components: persistence >Affects Versions: 3.0.0-beta1 > Environment: 2 nodes cluster running locally. >Reporter: Igor >Priority: Major > Labels: ignite-3 > > h3. Steps to reproduce: > Run the next queries in cycle with 50 repeats in single connection: > {code:java} > drop table if exists dropNoMoreIndexedColumn > create zone if not exists "rocksdb" engine rocksdb > create table dropNoMoreIndexedColumn(k1 TIMESTAMP not null, k2 INTEGER not > null, v0 TINYINT not null, v1 SMALLINT not null, v2 INTEGER not null, v3 > BIGINT not null, v4 VARCHAR not null, v5 TIMESTAMP not null, primary key (k1, > k2)) with PRIMARY_ZONE='rocksdb' > create index dropNoMoreIndexedColumn_v1idx on dropNoMoreIndexedColumn using > TREE (v1) > drop index dropNoMoreIndexedColumn_v1idx > alter table dropNoMoreIndexedColumn drop column v1 {code} > h3. Expected: > All queries are executed. > h3. Actual: > On repeat 31 the client freeze for infinite amount of time. > h3. Analysis: > With storages aimem and aipersist the issue is not happen. > The servers contain repeated error in logs: > {code:java} > 2024-02-05 13:47:24:812 +0100 > [ERROR][%DropColumnsTest_cluster_0%JRaft-FSMCaller-Disruptor-metastorage-_stripe_0-0][WatchProcessor] > Error occurred when notifying safe time advanced callback > java.util.concurrent.CompletionException: > java.lang.UnsupportedOperationException: Update log is not supported in > RocksDB storage. > at > java.base/java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:314) > at > java.base/java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:319) > at > java.base/java.util.concurrent.CompletableFuture$UniApply.tryFire(CompletableFuture.java:645) > at > java.base/java.util.concurrent.CompletableFuture$Completion.run(CompletableFuture.java:478) > at > java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) > at > java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) > at java.base/java.lang.Thread.run(Thread.java:834) > Caused by: java.lang.UnsupportedOperationException: Update log is not > supported in RocksDB storage. > at > org.apache.ignite.internal.storage.rocksdb.RocksDbMvPartitionStorage.trimUpdateLog(RocksDbMvPartitionStorage.java:908) > at > org.apache.ignite.internal.table.distributed.raft.snapshot.outgoing.SnapshotAwarePartitionDataStorage.trimUpdateLog(SnapshotAwarePartitionDataStorage.java:244) > at > org.apache.ignite.internal.table.distributed.gc.GcUpdateHandler.lambda$vacuumBatch$0(GcUpdateHandler.java:81) > at > org.apache.ignite.internal.storage.rocksdb.RocksDbMvPartitionStorage.lambda$runConsistently$2(RocksDbMvPartitionStorage.java:228) > at > org.apache.ignite.internal.storage.rocksdb.RocksDbMvPartitionStorage.busy(RocksDbMvPartitionStorage.java:1431) > at > org.apache.ignite.internal.storage.rocksdb.RocksDbMvPartitionStorage.runConsistently(RocksDbMvPartitionStorage.java:213) > at > org.apache.ignite.internal.table.distributed.raft.snapshot.outgoing.SnapshotAwarePartitionDataStorage.runConsistently(SnapshotAwarePartitionDataStorage.java:80) > at > org.apache.ignite.internal.table.distributed.gc.GcUpdateHandler.vacuumBatch(GcUpdateHandler.java:80) > at > org.apache.ignite.internal.table.distributed.gc.MvGc.lambda$scheduleGcForStorage$7(MvGc.java:242) > at > java.base/java.util.concurrent.CompletableFuture$UniApply.tryFire(CompletableFuture.java:642) > ... 4 more {code} -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (IGNITE-21777) 'Failed to get the primary replica' or 'Replication is timed out' or hangs with 'aimem' storage engine
[ https://issues.apache.org/jira/browse/IGNITE-21777?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Nikita Sivkov updated IGNITE-21777: --- Description: *Steps to reproduce:* # Create a cluster with 2 nodes. # Connect via JDBC. # Repeat the following SQL statements in a loop (for example, 100 times): ** {{drop table if exists tags}} ** {{create zone if not exists "AIMEM" engine aimem}} ** {{create table tags(tagId integer not null, tag varchar(100) not null, primary key (tagId)) with PRIMARY_ZONE='AIMEM'}} ** {{insert into tags(tagId, tag) values (1,'unit'), (2,'integration'), (3,'smoke'), (4,'sanity'), (5,'regression')}} *Expected result:* No errors or hangs happen. *Actual result:* Hangs on {{Create table}} or {{Insert into}} statement. _*OR*_ Get the error {{Replication is timed out}} {code:java} Replication is timed out [replicaGrpId=34_part_16] java.sql.SQLException: Replication is timed out [replicaGrpId=34_part_16] at org.apache.ignite.internal.jdbc.proto.IgniteQueryErrorCode.createJdbcSqlException(IgniteQueryErrorCode.java:57) at org.apache.ignite.internal.jdbc.JdbcStatement.execute0(JdbcStatement.java:154) at org.apache.ignite.internal.jdbc.JdbcStatement.executeUpdate(JdbcStatement.java:181) at org.gridgain.ai3tests.tests.teststeps.JdbcSteps.executeUpdateQuery(JdbcSteps.java:116) at org.gridgain.ai3tests.tests.UpdateTests.wannaCatchTheBug(UpdateTests.java:95) at java.base/java.lang.reflect.Method.invoke(Method.java:566) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) at java.base/java.lang.Thread.run(Thread.java:834) {code} _*OR*_ Get the error {{Failed to get the primary replica}} {code:java} java.sql.SQLException: Failed to get the primary replica [tablePartitionId=18_part_1] at org.apache.ignite.internal.jdbc.proto.IgniteQueryErrorCode.createJdbcSqlException(IgniteQueryErrorCode.java:57) at org.apache.ignite.internal.jdbc.JdbcStatement.execute0(JdbcStatement.java:154) at org.apache.ignite.internal.jdbc.JdbcStatement.executeUpdate(JdbcStatement.java:181) at org.gridgain.ai3tests.tests.teststeps.JdbcSteps.executeUpdateQuery(JdbcSteps.java:116) at org.gridgain.ai3tests.tests.UpdateTests.createAndFillTables(UpdateTests.java:247) at io.qameta.allure.Allure.lambda$step$0(Allure.java:113) at io.qameta.allure.Allure.lambda$step$1(Allure.java:127) at io.qameta.allure.Allure.step(Allure.java:181) at io.qameta.allure.Allure.step(Allure.java:125) at io.qameta.allure.Allure.step(Allure.java:112) at org.gridgain.ai3tests.tests.UpdateTests.updateTableWithConditionThatHasLinkedInnerSubQueries(UpdateTests.java:147) at java.base/java.lang.reflect.Method.invoke(Method.java:566) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) at java.base/java.lang.Thread.run(Thread.java:834) {code} was: *Steps to reproduce:* # Create a cluster with 2 nodes. # Connect via JDBC. # Repeat the following SQL statements in a loop (for example, 100 times): ** {{drop table if exists tags}} ** {{create zone if not exists "AIMEM" engine aimem}} ** {{create table tags(tagId integer not null, tag varchar(100) not null, primary key (tagId)) with PRIMARY_ZONE='AIMEM'}} ** {{insert into tags(tagId, tag) values (1,'unit'), (2,'integration'), (3,'smoke'), (4,'sanity'), (5,'regression')}} *Expected result:* No errors or hangs happen. *Actual result:* Hangs on {{Create table}} or {{Insert into}} statement. OR Get the error {{Replication is timed out}} {code:java} Replication is timed out [replicaGrpId=34_part_16] java.sql.SQLException: Replication is timed out [replicaGrpId=34_part_16] at org.apache.ignite.internal.jdbc.proto.IgniteQueryErrorCode.createJdbcSqlException(IgniteQueryErrorCode.java:57) at org.apache.ignite.internal.jdbc.JdbcStatement.execute0(JdbcStatement.java:154) at org.apache.ignite.internal.jdbc.JdbcStatement.executeUpdate(JdbcStatement.java:181) at org.gridgain.ai3tests.tests.teststeps.JdbcSteps.executeUpdateQuery(JdbcSteps.java:116) at org.gridgain.ai3tests.tests.UpdateTests.wannaCatchTheBug(UpdateTests.java:95) at java.base/java.lang.reflect.Method.invoke(Method.java:566) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) at
[jira] [Updated] (IGNITE-21777) 'Failed to get the primary replica' or 'Replication is timed out' or hangs with 'aimem' storage engine
[ https://issues.apache.org/jira/browse/IGNITE-21777?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Nikita Sivkov updated IGNITE-21777: --- Summary: 'Failed to get the primary replica' or 'Replication is timed out' or hangs with 'aimem' storage engine (was: 'Replication is timed out' or hangs with 'aimem' storage engine) > 'Failed to get the primary replica' or 'Replication is timed out' or hangs > with 'aimem' storage engine > -- > > Key: IGNITE-21777 > URL: https://issues.apache.org/jira/browse/IGNITE-21777 > Project: Ignite > Issue Type: Bug >Affects Versions: 3.0.0-beta2 > Environment: Cluster of 2 nodes. > Storage engine - aimem. >Reporter: Nikita Sivkov >Priority: Major > Labels: ignite-3 > > *Steps to reproduce:* # Create a cluster with 2 nodes. > # Connect via JDBC. > # Repeat the following SQL statements in a loop (for example, 100 times): > ** {{drop table if exists tags}} > ** {{create zone if not exists "AIMEM" engine aimem}} > ** {{create table tags(tagId integer not null, tag varchar(100) not null, > primary key (tagId)) with PRIMARY_ZONE='AIMEM'}} > ** {{insert into tags(tagId, tag) values (1,'unit'), (2,'integration'), > (3,'smoke'), (4,'sanity'), (5,'regression')}} > *Expected result:* > No errors or hangs happen. > *Actual result:* > Hangs on {{Create table}} or {{Insert into}} statement. > OR > Get the error {{Replication is timed out}} > {code:java} > Replication is timed out [replicaGrpId=34_part_16] > java.sql.SQLException: Replication is timed out [replicaGrpId=34_part_16] > at > org.apache.ignite.internal.jdbc.proto.IgniteQueryErrorCode.createJdbcSqlException(IgniteQueryErrorCode.java:57) > at > org.apache.ignite.internal.jdbc.JdbcStatement.execute0(JdbcStatement.java:154) > at > org.apache.ignite.internal.jdbc.JdbcStatement.executeUpdate(JdbcStatement.java:181) > at > org.gridgain.ai3tests.tests.teststeps.JdbcSteps.executeUpdateQuery(JdbcSteps.java:116) > at > org.gridgain.ai3tests.tests.UpdateTests.wannaCatchTheBug(UpdateTests.java:95) > at java.base/java.lang.reflect.Method.invoke(Method.java:566) > at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) > at > java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) > at > java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) > at java.base/java.lang.Thread.run(Thread.java:834) {code} > OR > Get the error {{Failed to get the primary replica}} > {code:java} > java.sql.SQLException: Failed to get the primary replica > [tablePartitionId=18_part_1] > at > org.apache.ignite.internal.jdbc.proto.IgniteQueryErrorCode.createJdbcSqlException(IgniteQueryErrorCode.java:57) > at > org.apache.ignite.internal.jdbc.JdbcStatement.execute0(JdbcStatement.java:154) > at > org.apache.ignite.internal.jdbc.JdbcStatement.executeUpdate(JdbcStatement.java:181) > at > org.gridgain.ai3tests.tests.teststeps.JdbcSteps.executeUpdateQuery(JdbcSteps.java:116) > at > org.gridgain.ai3tests.tests.UpdateTests.createAndFillTables(UpdateTests.java:247) > at io.qameta.allure.Allure.lambda$step$0(Allure.java:113) > at io.qameta.allure.Allure.lambda$step$1(Allure.java:127) > at io.qameta.allure.Allure.step(Allure.java:181) > at io.qameta.allure.Allure.step(Allure.java:125) > at io.qameta.allure.Allure.step(Allure.java:112) > at > org.gridgain.ai3tests.tests.UpdateTests.updateTableWithConditionThatHasLinkedInnerSubQueries(UpdateTests.java:147) > at java.base/java.lang.reflect.Method.invoke(Method.java:566) > at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) > at > java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) > at > java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) > at java.base/java.lang.Thread.run(Thread.java:834) {code} -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (IGNITE-21777) 'Replication is timed out' or hangs with 'aimem' storage engine
[ https://issues.apache.org/jira/browse/IGNITE-21777?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Nikita Sivkov updated IGNITE-21777: --- Description: *Steps to reproduce:* # Create a cluster with 2 nodes. # Connect via JDBC. # Repeat the following SQL statements in a loop (for example, 100 times): ** {{drop table if exists tags}} ** {{create zone if not exists "AIMEM" engine aimem}} ** {{create table tags(tagId integer not null, tag varchar(100) not null, primary key (tagId)) with PRIMARY_ZONE='AIMEM'}} ** {{insert into tags(tagId, tag) values (1,'unit'), (2,'integration'), (3,'smoke'), (4,'sanity'), (5,'regression')}} *Expected result:* No errors or hangs happen. *Actual result:* Hangs on {{Create table}} or {{Insert into}} statement. OR Get the error {{Replication is timed out}} {code:java} Replication is timed out [replicaGrpId=34_part_16] java.sql.SQLException: Replication is timed out [replicaGrpId=34_part_16] at org.apache.ignite.internal.jdbc.proto.IgniteQueryErrorCode.createJdbcSqlException(IgniteQueryErrorCode.java:57) at org.apache.ignite.internal.jdbc.JdbcStatement.execute0(JdbcStatement.java:154) at org.apache.ignite.internal.jdbc.JdbcStatement.executeUpdate(JdbcStatement.java:181) at org.gridgain.ai3tests.tests.teststeps.JdbcSteps.executeUpdateQuery(JdbcSteps.java:116) at org.gridgain.ai3tests.tests.UpdateTests.wannaCatchTheBug(UpdateTests.java:95) at java.base/java.lang.reflect.Method.invoke(Method.java:566) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) at java.base/java.lang.Thread.run(Thread.java:834) {code} OR Get the error {{Failed to get the primary replica}} {code:java} java.sql.SQLException: Failed to get the primary replica [tablePartitionId=18_part_1] at org.apache.ignite.internal.jdbc.proto.IgniteQueryErrorCode.createJdbcSqlException(IgniteQueryErrorCode.java:57) at org.apache.ignite.internal.jdbc.JdbcStatement.execute0(JdbcStatement.java:154) at org.apache.ignite.internal.jdbc.JdbcStatement.executeUpdate(JdbcStatement.java:181) at org.gridgain.ai3tests.tests.teststeps.JdbcSteps.executeUpdateQuery(JdbcSteps.java:116) at org.gridgain.ai3tests.tests.UpdateTests.createAndFillTables(UpdateTests.java:247) at io.qameta.allure.Allure.lambda$step$0(Allure.java:113) at io.qameta.allure.Allure.lambda$step$1(Allure.java:127) at io.qameta.allure.Allure.step(Allure.java:181) at io.qameta.allure.Allure.step(Allure.java:125) at io.qameta.allure.Allure.step(Allure.java:112) at org.gridgain.ai3tests.tests.UpdateTests.updateTableWithConditionThatHasLinkedInnerSubQueries(UpdateTests.java:147) at java.base/java.lang.reflect.Method.invoke(Method.java:566) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) at java.base/java.lang.Thread.run(Thread.java:834) {code} was: *Steps to reproduce:* # Create a cluster with 2 nodes. # Connect via JDBC. # Repeat the following SQL statements in a loop (for example, 100 times): ** {{drop table if exists tags}} ** {{create zone if not exists "AIMEM" engine aimem}} ** {{create table tags(tagId integer not null, tag varchar(100) not null, primary key (tagId)) with PRIMARY_ZONE='AIMEM'}} ** {{insert into tags(tagId, tag) values (1,'unit'), (2,'integration'), (3,'smoke'), (4,'sanity'), (5,'regression')}} *Expected result:* No errors or hangs happen. *Actual result:* Hangs on {{Create table}} or {{Insert into}} statement. OR Get the error {{Replication is timed out}} {code:java} Replication is timed out [replicaGrpId=34_part_16] java.sql.SQLException: Replication is timed out [replicaGrpId=34_part_16] at org.apache.ignite.internal.jdbc.proto.IgniteQueryErrorCode.createJdbcSqlException(IgniteQueryErrorCode.java:57) at org.apache.ignite.internal.jdbc.JdbcStatement.execute0(JdbcStatement.java:154) at org.apache.ignite.internal.jdbc.JdbcStatement.executeUpdate(JdbcStatement.java:181) at org.gridgain.ai3tests.tests.teststeps.JdbcSteps.executeUpdateQuery(JdbcSteps.java:116) at org.gridgain.ai3tests.tests.UpdateTests.wannaCatchTheBug(UpdateTests.java:95) at java.base/java.lang.reflect.Method.invoke(Method.java:566) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) at
[jira] [Created] (IGNITE-21777) 'Replication is timed out' or hangs with 'aimem' storage engine
Nikita Sivkov created IGNITE-21777: -- Summary: 'Replication is timed out' or hangs with 'aimem' storage engine Key: IGNITE-21777 URL: https://issues.apache.org/jira/browse/IGNITE-21777 Project: Ignite Issue Type: Bug Affects Versions: 3.0.0-beta2 Environment: Cluster of 2 nodes. Storage engine - aimem. Reporter: Nikita Sivkov *Steps to reproduce:* # Create a cluster with 2 nodes. # Connect via JDBC. # Repeat the following SQL statements in a loop (for example, 100 times): ** {{drop table if exists tags}} ** {{create zone if not exists "AIMEM" engine aimem}} ** {{create table tags(tagId integer not null, tag varchar(100) not null, primary key (tagId)) with PRIMARY_ZONE='AIMEM'}} ** {{insert into tags(tagId, tag) values (1,'unit'), (2,'integration'), (3,'smoke'), (4,'sanity'), (5,'regression')}} *Expected result:* No errors or hangs happen. *Actual result:* Hangs on {{Create table}} or {{Insert into}} statement. OR Get the error {{Replication is timed out}} {code:java} Replication is timed out [replicaGrpId=34_part_16] java.sql.SQLException: Replication is timed out [replicaGrpId=34_part_16] at org.apache.ignite.internal.jdbc.proto.IgniteQueryErrorCode.createJdbcSqlException(IgniteQueryErrorCode.java:57) at org.apache.ignite.internal.jdbc.JdbcStatement.execute0(JdbcStatement.java:154) at org.apache.ignite.internal.jdbc.JdbcStatement.executeUpdate(JdbcStatement.java:181) at org.gridgain.ai3tests.tests.teststeps.JdbcSteps.executeUpdateQuery(JdbcSteps.java:116) at org.gridgain.ai3tests.tests.UpdateTests.wannaCatchTheBug(UpdateTests.java:95) at java.base/java.lang.reflect.Method.invoke(Method.java:566) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) at java.base/java.lang.Thread.run(Thread.java:834) {code} -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Resolved] (IGNITE-21739) JDBC connection to a multi-node cluster doesn't take into account clientConnector.port from each node
[ https://issues.apache.org/jira/browse/IGNITE-21739?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Nikita Sivkov resolved IGNITE-21739. Resolution: Invalid Actually, the connection string was like {{jdbc:ignite:thin://172.24.1.2,172.24.1.3,172.24.1.4 and the actual issue is related to [IGNITE-21577|https://issues.apache.org/jira/browse/IGNITE-21577]}} > JDBC connection to a multi-node cluster doesn't take into account > clientConnector.port from each node > - > > Key: IGNITE-21739 > URL: https://issues.apache.org/jira/browse/IGNITE-21739 > Project: Ignite > Issue Type: Bug > Components: jdbc >Affects Versions: 3.0.0-beta2 > Environment: * multi-node cluster > * different `{{{}clientConnector.port{}}}` on each cluster node >Reporter: Nikita Sivkov >Priority: Major > Labels: ignite-3 > Attachments: exception.log > > > *WHEN* you create a multi-node cluster > *AND* specify different {color:#de350b}{{clientConnector.port}}{color} on > each cluster node > (for example, > node1 (172.24.1.2) - {color:#de350b}{{clientConnector.port=10800}}{color} > node2 (172.24.1.3) - {color:#de350b}{{clientConnector.port=10801}}{color} > node3 (172.24.1.4) - {color:#de350b}{{clientConnector.port=10802}}{color}) > *AND* connect to cluster like > {color:#de350b}{{{}jdbc:ignite:thin://{node1address{{color} (for example, > {{{color:#de350b}jdbc:ignite:thin://172.24.1.2{color})}} > *AND* try to insert a couple of records > *THEN* you will get an error like > {code:java} > Mar 12, 2024 7:37:21 PM org.apache.ignite.internal.logger.IgniteLogger > warnWARNING: Failed to establish connection to 172.24.1.3:10800: > org.apache.ignite.client.IgniteClientConnectionException: IGN-CLIENT-1 > TraceId:df21d718-d40c-4506-84e7-6ec141de9ab5 Client failed to connect: > Connection refused: no further information: > /172.24.1.3:10800java.util.concurrent.CompletionException: > org.apache.ignite.client.IgniteClientConnectionException: IGN-CLIENT-1 > TraceId:df21d718-d40c-4506-84e7-6ec141de9ab5 Client failed to connect: > Connection refused: no further information: /172.24.1.3:10800 at > java.base/java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:331) > at > java.base/java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:346) >at > java.base/java.util.concurrent.CompletableFuture$UniCompose.tryFire(CompletableFuture.java:1063) > at > java.base/java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506) > at > java.base/java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:2088) > at > org.apache.ignite.internal.client.io.netty.NettyClientConnectionMultiplexer.lambda$openAsync$1(NettyClientConnectionMultiplexer.java:197) > at > io.netty.util.concurrent.DefaultPromise.notifyListener0(DefaultPromise.java:590) > at > io.netty.util.concurrent.DefaultPromise.notifyListeners0(DefaultPromise.java:583) > at > io.netty.util.concurrent.DefaultPromise.notifyListenersNow(DefaultPromise.java:559) > at > io.netty.util.concurrent.DefaultPromise.notifyListeners(DefaultPromise.java:492) > at > io.netty.util.concurrent.DefaultPromise.setValue0(DefaultPromise.java:636) > at > io.netty.util.concurrent.DefaultPromise.setFailure0(DefaultPromise.java:629) > at > io.netty.util.concurrent.DefaultPromise.tryFailure(DefaultPromise.java:118) > at > io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.fulfillConnectPromise(AbstractNioChannel.java:322) > at > io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.finishConnect(AbstractNioChannel.java:338) > at > io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:776) > at > io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:724) > at > io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:650) > at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:562) at > io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) > at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) >at > io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) > at java.base/java.lang.Thread.run(Thread.java:834)Caused by: > org.apache.ignite.client.IgniteClientConnectionException: IGN-CLIENT-1 > TraceId:df21d718-d40c-4506-84e7-6ec141de9ab5 Client failed to connect: > Connection refused: no further information: /172.24.1.3:10800at > org.apache.ignite.internal.client.io.netty.NettyClientConnectionMultiplexer.lambda$openAsync$1(NettyClientConnectionMultiplexer.java:194) > ... 17
[jira] [Updated] (IGNITE-21739) JDBC connection to a multi-node cluster doesn't take into account clientConnector.port from each node
[ https://issues.apache.org/jira/browse/IGNITE-21739?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Nikita Sivkov updated IGNITE-21739: --- Affects Version/s: 3.0.0-beta2 > JDBC connection to a multi-node cluster doesn't take into account > clientConnector.port from each node > - > > Key: IGNITE-21739 > URL: https://issues.apache.org/jira/browse/IGNITE-21739 > Project: Ignite > Issue Type: Bug > Components: jdbc >Affects Versions: 3.0.0-beta2 > Environment: * multi-node cluster > * different `{{{}clientConnector.port{}}}` on each cluster node >Reporter: Nikita Sivkov >Priority: Major > Labels: ignite-3 > Attachments: exception.log > > > *WHEN* you create a multi-node cluster > *AND* specify different {color:#de350b}{{clientConnector.port}}{color} on > each cluster node > (for example, > node1 (172.24.1.2) - {color:#de350b}{{clientConnector.port=10800}}{color} > node2 (172.24.1.3) - {color:#de350b}{{clientConnector.port=10801}}{color} > node3 (172.24.1.4) - {color:#de350b}{{clientConnector.port=10802}}{color}) > *AND* connect to cluster like > {color:#de350b}{{{}jdbc:ignite:thin://{node1address{{color} (for example, > {{{color:#de350b}jdbc:ignite:thin://172.24.1.2{color})}} > *AND* try to insert a couple of records > *THEN* you will get an error like > {code:java} > Mar 12, 2024 7:37:21 PM org.apache.ignite.internal.logger.IgniteLogger > warnWARNING: Failed to establish connection to 172.24.1.3:10800: > org.apache.ignite.client.IgniteClientConnectionException: IGN-CLIENT-1 > TraceId:df21d718-d40c-4506-84e7-6ec141de9ab5 Client failed to connect: > Connection refused: no further information: > /172.24.1.3:10800java.util.concurrent.CompletionException: > org.apache.ignite.client.IgniteClientConnectionException: IGN-CLIENT-1 > TraceId:df21d718-d40c-4506-84e7-6ec141de9ab5 Client failed to connect: > Connection refused: no further information: /172.24.1.3:10800 at > java.base/java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:331) > at > java.base/java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:346) >at > java.base/java.util.concurrent.CompletableFuture$UniCompose.tryFire(CompletableFuture.java:1063) > at > java.base/java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506) > at > java.base/java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:2088) > at > org.apache.ignite.internal.client.io.netty.NettyClientConnectionMultiplexer.lambda$openAsync$1(NettyClientConnectionMultiplexer.java:197) > at > io.netty.util.concurrent.DefaultPromise.notifyListener0(DefaultPromise.java:590) > at > io.netty.util.concurrent.DefaultPromise.notifyListeners0(DefaultPromise.java:583) > at > io.netty.util.concurrent.DefaultPromise.notifyListenersNow(DefaultPromise.java:559) > at > io.netty.util.concurrent.DefaultPromise.notifyListeners(DefaultPromise.java:492) > at > io.netty.util.concurrent.DefaultPromise.setValue0(DefaultPromise.java:636) > at > io.netty.util.concurrent.DefaultPromise.setFailure0(DefaultPromise.java:629) > at > io.netty.util.concurrent.DefaultPromise.tryFailure(DefaultPromise.java:118) > at > io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.fulfillConnectPromise(AbstractNioChannel.java:322) > at > io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.finishConnect(AbstractNioChannel.java:338) > at > io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:776) > at > io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:724) > at > io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:650) > at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:562) at > io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) > at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) >at > io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) > at java.base/java.lang.Thread.run(Thread.java:834)Caused by: > org.apache.ignite.client.IgniteClientConnectionException: IGN-CLIENT-1 > TraceId:df21d718-d40c-4506-84e7-6ec141de9ab5 Client failed to connect: > Connection refused: no further information: /172.24.1.3:10800at > org.apache.ignite.internal.client.io.netty.NettyClientConnectionMultiplexer.lambda$openAsync$1(NettyClientConnectionMultiplexer.java:194) > ... 17 moreCaused by: > io.netty.channel.AbstractChannel$AnnotatedConnectException: Connection > refused: no further information: /172.24.1.3:10800Caused by: > java.net.ConnectException:
[jira] [Updated] (IGNITE-21739) JDBC connection to a multi-node cluster doesn't take into account clientConnector.port from each node
[ https://issues.apache.org/jira/browse/IGNITE-21739?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Nikita Sivkov updated IGNITE-21739: --- Labels: ignite-3 (was: ) > JDBC connection to a multi-node cluster doesn't take into account > clientConnector.port from each node > - > > Key: IGNITE-21739 > URL: https://issues.apache.org/jira/browse/IGNITE-21739 > Project: Ignite > Issue Type: Bug > Components: jdbc > Environment: * multi-node cluster > * different `{{{}clientConnector.port{}}}` on each cluster node >Reporter: Nikita Sivkov >Priority: Major > Labels: ignite-3 > Attachments: exception.log > > > *WHEN* you create a multi-node cluster > *AND* specify different {color:#de350b}{{clientConnector.port}}{color} on > each cluster node > (for example, > node1 (172.24.1.2) - {color:#de350b}{{clientConnector.port=10800}}{color} > node2 (172.24.1.3) - {color:#de350b}{{clientConnector.port=10801}}{color} > node3 (172.24.1.4) - {color:#de350b}{{clientConnector.port=10802}}{color}) > *AND* connect to cluster like > {color:#de350b}{{{}jdbc:ignite:thin://{node1address{{color} (for example, > {{{color:#de350b}jdbc:ignite:thin://172.24.1.2{color})}} > *AND* try to insert a couple of records > *THEN* you will get an error like > {code:java} > Mar 12, 2024 7:37:21 PM org.apache.ignite.internal.logger.IgniteLogger > warnWARNING: Failed to establish connection to 172.24.1.3:10800: > org.apache.ignite.client.IgniteClientConnectionException: IGN-CLIENT-1 > TraceId:df21d718-d40c-4506-84e7-6ec141de9ab5 Client failed to connect: > Connection refused: no further information: > /172.24.1.3:10800java.util.concurrent.CompletionException: > org.apache.ignite.client.IgniteClientConnectionException: IGN-CLIENT-1 > TraceId:df21d718-d40c-4506-84e7-6ec141de9ab5 Client failed to connect: > Connection refused: no further information: /172.24.1.3:10800 at > java.base/java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:331) > at > java.base/java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:346) >at > java.base/java.util.concurrent.CompletableFuture$UniCompose.tryFire(CompletableFuture.java:1063) > at > java.base/java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506) > at > java.base/java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:2088) > at > org.apache.ignite.internal.client.io.netty.NettyClientConnectionMultiplexer.lambda$openAsync$1(NettyClientConnectionMultiplexer.java:197) > at > io.netty.util.concurrent.DefaultPromise.notifyListener0(DefaultPromise.java:590) > at > io.netty.util.concurrent.DefaultPromise.notifyListeners0(DefaultPromise.java:583) > at > io.netty.util.concurrent.DefaultPromise.notifyListenersNow(DefaultPromise.java:559) > at > io.netty.util.concurrent.DefaultPromise.notifyListeners(DefaultPromise.java:492) > at > io.netty.util.concurrent.DefaultPromise.setValue0(DefaultPromise.java:636) > at > io.netty.util.concurrent.DefaultPromise.setFailure0(DefaultPromise.java:629) > at > io.netty.util.concurrent.DefaultPromise.tryFailure(DefaultPromise.java:118) > at > io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.fulfillConnectPromise(AbstractNioChannel.java:322) > at > io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.finishConnect(AbstractNioChannel.java:338) > at > io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:776) > at > io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:724) > at > io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:650) > at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:562) at > io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) > at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) >at > io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) > at java.base/java.lang.Thread.run(Thread.java:834)Caused by: > org.apache.ignite.client.IgniteClientConnectionException: IGN-CLIENT-1 > TraceId:df21d718-d40c-4506-84e7-6ec141de9ab5 Client failed to connect: > Connection refused: no further information: /172.24.1.3:10800at > org.apache.ignite.internal.client.io.netty.NettyClientConnectionMultiplexer.lambda$openAsync$1(NettyClientConnectionMultiplexer.java:194) > ... 17 moreCaused by: > io.netty.channel.AbstractChannel$AnnotatedConnectException: Connection > refused: no further information: /172.24.1.3:10800Caused by: > java.net.ConnectException: Connection refused: no further information at >
[jira] [Updated] (IGNITE-21739) JDBC connection to a multi-node cluster doesn't take into account clientConnector.port from each node
[ https://issues.apache.org/jira/browse/IGNITE-21739?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Nikita Sivkov updated IGNITE-21739: --- Attachment: exception.log > JDBC connection to a multi-node cluster doesn't take into account > clientConnector.port from each node > - > > Key: IGNITE-21739 > URL: https://issues.apache.org/jira/browse/IGNITE-21739 > Project: Ignite > Issue Type: Bug > Components: jdbc > Environment: * multi-node cluster > * different `{{{}clientConnector.port{}}}` on each cluster node >Reporter: Nikita Sivkov >Priority: Major > Attachments: exception.log > > > *WHEN* you create a multi-node cluster > *AND* specify different {color:#de350b}{{clientConnector.port}}{color} on > each cluster node > (for example, > node1 (172.24.1.2) - {color:#de350b}{{clientConnector.port=10800}}{color} > node2 (172.24.1.3) - {color:#de350b}{{clientConnector.port=10801}}{color} > node3 (172.24.1.4) - {color:#de350b}{{clientConnector.port=10802}}{color}) > *AND* connect to cluster like > {color:#de350b}{{{}jdbc:ignite:thin://{node1address{{color} (for example, > {{{color:#de350b}jdbc:ignite:thin://172.24.1.2{color})}} > *AND* try to insert a couple of records > *THEN* you will get an error like > {code:java} > Mar 12, 2024 7:37:21 PM org.apache.ignite.internal.logger.IgniteLogger > warnWARNING: Failed to establish connection to 172.24.1.3:10800: > org.apache.ignite.client.IgniteClientConnectionException: IGN-CLIENT-1 > TraceId:df21d718-d40c-4506-84e7-6ec141de9ab5 Client failed to connect: > Connection refused: no further information: > /172.24.1.3:10800java.util.concurrent.CompletionException: > org.apache.ignite.client.IgniteClientConnectionException: IGN-CLIENT-1 > TraceId:df21d718-d40c-4506-84e7-6ec141de9ab5 Client failed to connect: > Connection refused: no further information: /172.24.1.3:10800 at > java.base/java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:331) > at > java.base/java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:346) >at > java.base/java.util.concurrent.CompletableFuture$UniCompose.tryFire(CompletableFuture.java:1063) > at > java.base/java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506) > at > java.base/java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:2088) > at > org.apache.ignite.internal.client.io.netty.NettyClientConnectionMultiplexer.lambda$openAsync$1(NettyClientConnectionMultiplexer.java:197) > at > io.netty.util.concurrent.DefaultPromise.notifyListener0(DefaultPromise.java:590) > at > io.netty.util.concurrent.DefaultPromise.notifyListeners0(DefaultPromise.java:583) > at > io.netty.util.concurrent.DefaultPromise.notifyListenersNow(DefaultPromise.java:559) > at > io.netty.util.concurrent.DefaultPromise.notifyListeners(DefaultPromise.java:492) > at > io.netty.util.concurrent.DefaultPromise.setValue0(DefaultPromise.java:636) > at > io.netty.util.concurrent.DefaultPromise.setFailure0(DefaultPromise.java:629) > at > io.netty.util.concurrent.DefaultPromise.tryFailure(DefaultPromise.java:118) > at > io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.fulfillConnectPromise(AbstractNioChannel.java:322) > at > io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.finishConnect(AbstractNioChannel.java:338) > at > io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:776) > at > io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:724) > at > io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:650) > at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:562) at > io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) > at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) >at > io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) > at java.base/java.lang.Thread.run(Thread.java:834)Caused by: > org.apache.ignite.client.IgniteClientConnectionException: IGN-CLIENT-1 > TraceId:df21d718-d40c-4506-84e7-6ec141de9ab5 Client failed to connect: > Connection refused: no further information: /172.24.1.3:10800at > org.apache.ignite.internal.client.io.netty.NettyClientConnectionMultiplexer.lambda$openAsync$1(NettyClientConnectionMultiplexer.java:194) > ... 17 moreCaused by: > io.netty.channel.AbstractChannel$AnnotatedConnectException: Connection > refused: no further information: /172.24.1.3:10800Caused by: > java.net.ConnectException: Connection refused: no further information at >
[jira] [Updated] (IGNITE-21739) JDBC connection to a multi-node cluster doesn't take into account clientConnector.port from each node
[ https://issues.apache.org/jira/browse/IGNITE-21739?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Nikita Sivkov updated IGNITE-21739: --- Description: *WHEN* you create a multi-node cluster *AND* specify different {color:#de350b}{{clientConnector.port}}{color} on each cluster node (for example, node1 (172.24.1.2) - {color:#de350b}{{clientConnector.port=10800}}{color} node2 (172.24.1.3) - {color:#de350b}{{clientConnector.port=10801}}{color} node3 (172.24.1.4) - {color:#de350b}{{clientConnector.port=10802}}{color}) *AND* connect to cluster like {color:#de350b}{{{}jdbc:ignite:thin://{node1address{{color} (for example, {{{color:#de350b}jdbc:ignite:thin://172.24.1.2{color})}} *AND* try to insert a couple of records *THEN* you will get an error like {code:java} Mar 12, 2024 7:37:21 PM org.apache.ignite.internal.logger.IgniteLogger warnWARNING: Failed to establish connection to 172.24.1.3:10800: org.apache.ignite.client.IgniteClientConnectionException: IGN-CLIENT-1 TraceId:df21d718-d40c-4506-84e7-6ec141de9ab5 Client failed to connect: Connection refused: no further information: /172.24.1.3:10800java.util.concurrent.CompletionException: org.apache.ignite.client.IgniteClientConnectionException: IGN-CLIENT-1 TraceId:df21d718-d40c-4506-84e7-6ec141de9ab5 Client failed to connect: Connection refused: no further information: /172.24.1.3:10800at java.base/java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:331) at java.base/java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:346) at java.base/java.util.concurrent.CompletableFuture$UniCompose.tryFire(CompletableFuture.java:1063) at java.base/java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506) at java.base/java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:2088) at org.apache.ignite.internal.client.io.netty.NettyClientConnectionMultiplexer.lambda$openAsync$1(NettyClientConnectionMultiplexer.java:197) at io.netty.util.concurrent.DefaultPromise.notifyListener0(DefaultPromise.java:590) at io.netty.util.concurrent.DefaultPromise.notifyListeners0(DefaultPromise.java:583) at io.netty.util.concurrent.DefaultPromise.notifyListenersNow(DefaultPromise.java:559) at io.netty.util.concurrent.DefaultPromise.notifyListeners(DefaultPromise.java:492) at io.netty.util.concurrent.DefaultPromise.setValue0(DefaultPromise.java:636) at io.netty.util.concurrent.DefaultPromise.setFailure0(DefaultPromise.java:629) at io.netty.util.concurrent.DefaultPromise.tryFailure(DefaultPromise.java:118) at io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.fulfillConnectPromise(AbstractNioChannel.java:322) at io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.finishConnect(AbstractNioChannel.java:338) at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:776) at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:724) at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:650) at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:562) at io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) at io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) at java.base/java.lang.Thread.run(Thread.java:834)Caused by: org.apache.ignite.client.IgniteClientConnectionException: IGN-CLIENT-1 TraceId:df21d718-d40c-4506-84e7-6ec141de9ab5 Client failed to connect: Connection refused: no further information: /172.24.1.3:10800at org.apache.ignite.internal.client.io.netty.NettyClientConnectionMultiplexer.lambda$openAsync$1(NettyClientConnectionMultiplexer.java:194) ... 17 moreCaused by: io.netty.channel.AbstractChannel$AnnotatedConnectException: Connection refused: no further information: /172.24.1.3:10800Caused by: java.net.ConnectException: Connection refused: no further information at java.base/sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) at java.base/sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:779) at io.netty.channel.socket.nio.NioSocketChannel.doFinishConnect(NioSocketChannel.java:337) at io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.finishConnect(AbstractNioChannel.java:335) at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:776) at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:724) at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:650) at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:562) at io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) at
[jira] [Updated] (IGNITE-21739) JDBC connection to a multi-node cluster doesn't take into account clientConnector.port from each node
[ https://issues.apache.org/jira/browse/IGNITE-21739?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Nikita Sivkov updated IGNITE-21739: --- Description: Mar 12, 2024 7:37:21 PM org.apache.ignite.internal.logger.IgniteLogger warn WARNING: Failed to establish connection to 172.24.1.3:10800: org.apache.ignite.client.IgniteClientConnectionException: IGN-CLIENT-1 TraceId:df21d718-d40c-4506-84e7-6ec141de9ab5 Client failed to connect: Connection refused: no further information: /172.24.1.3:10800 java.util.concurrent.CompletionException: org.apache.ignite.client.IgniteClientConnectionException: IGN-CLIENT-1 TraceId:df21d718-d40c-4506-84e7-6ec141de9ab5 Client failed to connect: Connection refused: no further information: /172.24.1.3:10800 at java.base/java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:331) at java.base/java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:346) at java.base/java.util.concurrent.CompletableFuture$UniCompose.tryFire(CompletableFuture.java:1063) at java.base/java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506) at java.base/java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:2088) at org.apache.ignite.internal.client.io.netty.NettyClientConnectionMultiplexer.lambda$openAsync$1(NettyClientConnectionMultiplexer.java:197) at io.netty.util.concurrent.DefaultPromise.notifyListener0(DefaultPromise.java:590) at io.netty.util.concurrent.DefaultPromise.notifyListeners0(DefaultPromise.java:583) at io.netty.util.concurrent.DefaultPromise.notifyListenersNow(DefaultPromise.java:559) at io.netty.util.concurrent.DefaultPromise.notifyListeners(DefaultPromise.java:492) at io.netty.util.concurrent.DefaultPromise.setValue0(DefaultPromise.java:636) at io.netty.util.concurrent.DefaultPromise.setFailure0(DefaultPromise.java:629) at io.netty.util.concurrent.DefaultPromise.tryFailure(DefaultPromise.java:118) at io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.fulfillConnectPromise(AbstractNioChannel.java:322) at io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.finishConnect(AbstractNioChannel.java:338) at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:776) at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:724) at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:650) at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:562) at io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) at io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) at java.base/java.lang.Thread.run(Thread.java:834) Caused by: org.apache.ignite.client.IgniteClientConnectionException: IGN-CLIENT-1 TraceId:df21d718-d40c-4506-84e7-6ec141de9ab5 Client failed to connect: Connection refused: no further information: /172.24.1.3:10800 at org.apache.ignite.internal.client.io.netty.NettyClientConnectionMultiplexer.lambda$openAsync$1(NettyClientConnectionMultiplexer.java:194) ... 17 more Caused by: io.netty.channel.AbstractChannel$AnnotatedConnectException: Connection refused: no further information: /172.24.1.3:10800 Caused by: java.net.ConnectException: Connection refused: no further information at java.base/sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) at java.base/sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:779) at io.netty.channel.socket.nio.NioSocketChannel.doFinishConnect(NioSocketChannel.java:337) at io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.finishConnect(AbstractNioChannel.java:335) at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:776) at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:724) at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:650) at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:562) at io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) at io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) at java.base/java.lang.Thread.run(Thread.java:834) Mar 12, 2024 7:37:21 PM org.apache.ignite.internal.logger.IgniteLogger warn WARNING: Failed to establish connection to 172.24.1.4:10800: org.apache.ignite.client.IgniteClientConnectionException: IGN-CLIENT-1 TraceId:d7d08ee0-64f6-45ea-872f-1c2adb8d128a Client failed to connect: Connection refused: no further information: /172.24.1.4:10800 java.util.concurrent.CompletionException: org.apache.ignite.client.IgniteClientConnectionException: IGN-CLIENT-1 TraceId:d7d08ee0-64f6-45ea-872f-1c2adb8d128a Client failed to connect: Connection refused:
[jira] [Updated] (IGNITE-21739) JDBC connection to a multi-node cluster doesn't take into account clientConnector.port from each node
[ https://issues.apache.org/jira/browse/IGNITE-21739?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Nikita Sivkov updated IGNITE-21739: --- Description: *WHEN* you create a multi-node cluster *AND* specify different {{clientConnector.port}} on each cluster node (for example, node1 - 172.24.1.2 - {{clientConnector.port=10800}} node2 - 172.24.1.3 - {{clientConnector.port=10801}} node3 - 172.24.1.4 - {{{}clientConnector.port=10802{}}}) *AND* connect to cluster like {{{}jdbc:ignite:thin://{node1address{ (for example, {{jdbc:ignite:thin://172.24.1.2)}} *AND* try to insert a couple of records *THEN* you will get an error like {code:java} WARNING: Failed to establish connection to 172.24.1.3:10800: org.apache.ignite.client.IgniteClientConnectionException: IGN-CLIENT-1 TraceId:df21d718-d40c-4506-84e7-6ec141de9ab5 Client failed to connect: Connection refused: no further information: /172.24.1.3:10800 java.util.concurrent.CompletionException: org.apache.ignite.client.IgniteClientConnectionException: IGN-CLIENT-1 TraceId:df21d718-d40c-4506-84e7-6ec141de9ab5 Client failed to connect: Connection refused: no further information: /172.24.1.3:10800 at java.base/java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:331) at java.base/java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:346) at java.base/java.util.concurrent.CompletableFuture$UniCompose.tryFire(CompletableFuture.java:1063) at java.base/java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506) at java.base/java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:2088) at org.apache.ignite.internal.client.io.netty.NettyClientConnectionMultiplexer.lambda$openAsync$1(NettyClientConnectionMultiplexer.java:197) at io.netty.util.concurrent.DefaultPromise.notifyListener0(DefaultPromise.java:590) at io.netty.util.concurrent.DefaultPromise.notifyListeners0(DefaultPromise.java:583) at io.netty.util.concurrent.DefaultPromise.notifyListenersNow(DefaultPromise.java:559) at io.netty.util.concurrent.DefaultPromise.notifyListeners(DefaultPromise.java:492) at io.netty.util.concurrent.DefaultPromise.setValue0(DefaultPromise.java:636) at io.netty.util.concurrent.DefaultPromise.setFailure0(DefaultPromise.java:629) at io.netty.util.concurrent.DefaultPromise.tryFailure(DefaultPromise.java:118) at io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.fulfillConnectPromise(AbstractNioChannel.java:322) at io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.finishConnect(AbstractNioChannel.java:338) at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:776) at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:724) at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:650) at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:562) at io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) at io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) at java.base/java.lang.Thread.run(Thread.java:834) Caused by: org.apache.ignite.client.IgniteClientConnectionException: IGN-CLIENT-1 TraceId:df21d718-d40c-4506-84e7-6ec141de9ab5 Client failed to connect: Connection refused: no further information: /172.24.1.3:10800 at org.apache.ignite.internal.client.io.netty.NettyClientConnectionMultiplexer.lambda$openAsync$1(NettyClientConnectionMultiplexer.java:194) ... 17 more Caused by: io.netty.channel.AbstractChannel$AnnotatedConnectException: Connection refused: no further information: /172.24.1.3:10800 Caused by: java.net.ConnectException: Connection refused: no further information at java.base/sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) at java.base/sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:779) at io.netty.channel.socket.nio.NioSocketChannel.doFinishConnect(NioSocketChannel.java:337) at io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.finishConnect(AbstractNioChannel.java:335) at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:776) at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:724) at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:650) at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:562) at io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) at io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) at
[jira] [Updated] (IGNITE-21739) JDBC connection to a multi-node cluster doesn't take into account clientConnector.port from each node
[ https://issues.apache.org/jira/browse/IGNITE-21739?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Nikita Sivkov updated IGNITE-21739: --- Description: *WHEN* you create a multi-node cluster *AND* specify different {{clientConnector.port}} on each cluster node (for example, node1 - {{clientConnector.port=10800}} node2 - {{clientConnector.port=10801}} node3 - {{{}clientConnector.port=10802{}}}) *AND* connect to cluster like {{jdbc:ignite:thin://\{node1address}}} *AND* try to insert a couple of records *THEN* you will get an error like {code:java} WARNING: Failed to establish connection to 172.24.1.3:10800: org.apache.ignite.client.IgniteClientConnectionException: IGN-CLIENT-1 TraceId:df21d718-d40c-4506-84e7-6ec141de9ab5 Client failed to connect: Connection refused: no further information: /172.24.1.3:10800 java.util.concurrent.CompletionException: org.apache.ignite.client.IgniteClientConnectionException: IGN-CLIENT-1 TraceId:df21d718-d40c-4506-84e7-6ec141de9ab5 Client failed to connect: Connection refused: no further information: /172.24.1.3:10800 at java.base/java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:331) at java.base/java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:346) at java.base/java.util.concurrent.CompletableFuture$UniCompose.tryFire(CompletableFuture.java:1063) at java.base/java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506) at java.base/java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:2088) at org.apache.ignite.internal.client.io.netty.NettyClientConnectionMultiplexer.lambda$openAsync$1(NettyClientConnectionMultiplexer.java:197) at io.netty.util.concurrent.DefaultPromise.notifyListener0(DefaultPromise.java:590) at io.netty.util.concurrent.DefaultPromise.notifyListeners0(DefaultPromise.java:583) at io.netty.util.concurrent.DefaultPromise.notifyListenersNow(DefaultPromise.java:559) at io.netty.util.concurrent.DefaultPromise.notifyListeners(DefaultPromise.java:492) at io.netty.util.concurrent.DefaultPromise.setValue0(DefaultPromise.java:636) at io.netty.util.concurrent.DefaultPromise.setFailure0(DefaultPromise.java:629) at io.netty.util.concurrent.DefaultPromise.tryFailure(DefaultPromise.java:118) at io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.fulfillConnectPromise(AbstractNioChannel.java:322) at io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.finishConnect(AbstractNioChannel.java:338) at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:776) at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:724) at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:650) at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:562) at io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) at io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) at java.base/java.lang.Thread.run(Thread.java:834) Caused by: org.apache.ignite.client.IgniteClientConnectionException: IGN-CLIENT-1 TraceId:df21d718-d40c-4506-84e7-6ec141de9ab5 Client failed to connect: Connection refused: no further information: /172.24.1.3:10800 at org.apache.ignite.internal.client.io.netty.NettyClientConnectionMultiplexer.lambda$openAsync$1(NettyClientConnectionMultiplexer.java:194) ... 17 more Caused by: io.netty.channel.AbstractChannel$AnnotatedConnectException: Connection refused: no further information: /172.24.1.3:10800 Caused by: java.net.ConnectException: Connection refused: no further information at java.base/sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) at java.base/sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:779) at io.netty.channel.socket.nio.NioSocketChannel.doFinishConnect(NioSocketChannel.java:337) at io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.finishConnect(AbstractNioChannel.java:335) at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:776) at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:724) at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:650) at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:562) at io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) at io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) at java.base/java.lang.Thread.run(Thread.java:834) {code} was: *Issue:* WHEN you create a multi-node cluster AND
[jira] [Created] (IGNITE-21739) JDBC connection to a multi-node cluster doesn't take into account clientConnector.port from each node
Nikita Sivkov created IGNITE-21739: -- Summary: JDBC connection to a multi-node cluster doesn't take into account clientConnector.port from each node Key: IGNITE-21739 URL: https://issues.apache.org/jira/browse/IGNITE-21739 Project: Ignite Issue Type: Bug Components: jdbc Environment: * multi-node cluster * different `{{{}clientConnector.port{}}}` on each cluster node Reporter: Nikita Sivkov *Issue:* WHEN you create a multi-node cluster AND specify different `{{{}clientConnector.port{}}}` on each cluster node (for example, node1 - `{{{}clientConnector.port=10800{}}}` node2 - `{{{}clientConnector.port=10801{}}}` node3 - `{{{}clientConnector.port=10802{}}}`) AND connect to cluster like `{{{}jdbc:ignite:thin://\{node1address}{}}}` AND try to insert a couple of records THEN you will get an error like ``` WARNING: Failed to establish connection to 172.24.1.3:10800: org.apache.ignite.client.IgniteClientConnectionException: IGN-CLIENT-1 TraceId:df21d718-d40c-4506-84e7-6ec141de9ab5 Client failed to connect: Connection refused: no further information: /172.24.1.3:10800 java.util.concurrent.CompletionException: org.apache.ignite.client.IgniteClientConnectionException: IGN-CLIENT-1 TraceId:df21d718-d40c-4506-84e7-6ec141de9ab5 Client failed to connect: Connection refused: no further information: /172.24.1.3:10800 at java.base/java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:331) at java.base/java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:346) at java.base/java.util.concurrent.CompletableFuture$UniCompose.tryFire(CompletableFuture.java:1063) at java.base/java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506) at java.base/java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:2088) at org.apache.ignite.internal.client.io.netty.NettyClientConnectionMultiplexer.lambda$openAsync$1(NettyClientConnectionMultiplexer.java:197) at io.netty.util.concurrent.DefaultPromise.notifyListener0(DefaultPromise.java:590) at io.netty.util.concurrent.DefaultPromise.notifyListeners0(DefaultPromise.java:583) at io.netty.util.concurrent.DefaultPromise.notifyListenersNow(DefaultPromise.java:559) at io.netty.util.concurrent.DefaultPromise.notifyListeners(DefaultPromise.java:492) at io.netty.util.concurrent.DefaultPromise.setValue0(DefaultPromise.java:636) at io.netty.util.concurrent.DefaultPromise.setFailure0(DefaultPromise.java:629) at io.netty.util.concurrent.DefaultPromise.tryFailure(DefaultPromise.java:118) at io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.fulfillConnectPromise(AbstractNioChannel.java:322) at io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.finishConnect(AbstractNioChannel.java:338) at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:776) at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:724) at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:650) at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:562) at io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) at io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) at java.base/java.lang.Thread.run(Thread.java:834) Caused by: org.apache.ignite.client.IgniteClientConnectionException: IGN-CLIENT-1 TraceId:df21d718-d40c-4506-84e7-6ec141de9ab5 Client failed to connect: Connection refused: no further information: /172.24.1.3:10800 at org.apache.ignite.internal.client.io.netty.NettyClientConnectionMultiplexer.lambda$openAsync$1(NettyClientConnectionMultiplexer.java:194) ... 17 more Caused by: io.netty.channel.AbstractChannel$AnnotatedConnectException: Connection refused: no further information: /172.24.1.3:10800 Caused by: java.net.ConnectException: Connection refused: no further information at java.base/sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) at java.base/sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:779) at io.netty.channel.socket.nio.NioSocketChannel.doFinishConnect(NioSocketChannel.java:337) at io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.finishConnect(AbstractNioChannel.java:335) at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:776) at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:724) at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:650) at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:562) at
[jira] [Updated] (IGNITE-20154) Line separator in ODBC errors text
[ https://issues.apache.org/jira/browse/IGNITE-20154?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Nikita Sivkov updated IGNITE-20154: --- Description: *Issue:* Getting line separator symbols "\r\n" in OBDC error text. *For example:* {noformat} ALTER TABLE PUBLIC.CAR DROP COLUMN IF EXISTS NAME ('HY000', '[HY000] org.apache.ignite.sql.SqlException: Failed to parse query: Encountered "" at line 1, column 36.\r\nWas expecting one of:\r\n (262147, 3e945522-7d71-48c0-a8cf-675bbc078bb0) (0) (SQLExecDirectW)'){noformat} *Reproducer:* # Start Ignite node. # Execute command: _python3 odbc_client.py -o smoke0.sql.actual localhost:10800 smoke0.sql_ *Commit id:* 2655e406b06a2605c2d5ad9402e06d81c1a168ef was: *Issue:* Getting line separator symbols "\r\n" in OBDC error text. ** *For example:* {noformat} {noformat} *ALTER TABLE PUBLIC.CAR DROP COLUMN IF EXISTS NAME ('HY000', '[HY000] org.apache.ignite.sql.SqlException: Failed to parse query: Encountered "" at line 1, column 36.\r\nWas expecting one of:\r\n (262147, 3e945522-7d71-48c0-a8cf-675bbc078bb0) (0) (SQLExecDirectW)')* *Reproducer:* # Start Ignite node. # Execute command: `python3 odbc_client.py -o smoke0.sql.actual localhost:10800 smoke0.sql` *Commit id:* 2655e406b06a2605c2d5ad9402e06d81c1a168ef > Line separator in ODBC errors text > -- > > Key: IGNITE-20154 > URL: https://issues.apache.org/jira/browse/IGNITE-20154 > Project: Ignite > Issue Type: Bug > Components: odbc >Affects Versions: 3.0 >Reporter: Nikita Sivkov >Priority: Major > Labels: ignite-3, odbc > Attachments: odbc_client.py, smoke0.sql > > > *Issue:* > Getting line separator symbols "\r\n" in OBDC error text. > *For example:* > {noformat} > ALTER TABLE PUBLIC.CAR DROP COLUMN IF EXISTS NAME > ('HY000', '[HY000] org.apache.ignite.sql.SqlException: Failed to parse query: > Encountered "" at line 1, column 36.\r\nWas expecting one of:\r\n > (262147, 3e945522-7d71-48c0-a8cf-675bbc078bb0) (0) > (SQLExecDirectW)'){noformat} > *Reproducer:* > # Start Ignite node. > # Execute command: > _python3 odbc_client.py -o smoke0.sql.actual localhost:10800 smoke0.sql_ > *Commit id:* > 2655e406b06a2605c2d5ad9402e06d81c1a168ef > > -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Created] (IGNITE-20154) Line separator in ODBC errors text
Nikita Sivkov created IGNITE-20154: -- Summary: Line separator in ODBC errors text Key: IGNITE-20154 URL: https://issues.apache.org/jira/browse/IGNITE-20154 Project: Ignite Issue Type: Bug Components: odbc Affects Versions: 3.0 Reporter: Nikita Sivkov Attachments: odbc_client.py, smoke0.sql *Issue:* Getting line separator symbols "\r\n" in OBDC error text. ** *For example:* {noformat} {noformat} *ALTER TABLE PUBLIC.CAR DROP COLUMN IF EXISTS NAME ('HY000', '[HY000] org.apache.ignite.sql.SqlException: Failed to parse query: Encountered "" at line 1, column 36.\r\nWas expecting one of:\r\n (262147, 3e945522-7d71-48c0-a8cf-675bbc078bb0) (0) (SQLExecDirectW)')* *Reproducer:* # Start Ignite node. # Execute command: `python3 odbc_client.py -o smoke0.sql.actual localhost:10800 smoke0.sql` *Commit id:* 2655e406b06a2605c2d5ad9402e06d81c1a168ef -- This message was sent by Atlassian Jira (v8.20.10#820010)