Apache-Phoenix | 5.1 | HBase 2.1 | Build #303 SUCCESS
5.1 branch HBase 2.1 build #303 status SUCCESS Build #303 https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-mulitbranch/job/5.1/303/
Apache-Phoenix | master | HBase 2.4 | Build #622 FAILURE
master branch HBase 2.4 build #622 status FAILURE Build #622 https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-mulitbranch/job/master/622/
Apache-Phoenix | master | HBase 2.5 | Build #622 FAILURE
master branch HBase 2.5 build #622 status FAILURE Build #622 https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-mulitbranch/job/master/622/
Apache-Phoenix | 5.1 | HBase 2.2 | Build #303 FAILURE
5.1 branch HBase 2.2 build #303 status FAILURE Build #303 https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-mulitbranch/job/5.1/303/
Apache-Phoenix | 5.1 | HBase 2.3 | Build #303 SUCCESS
5.1 branch HBase 2.3 build #303 status SUCCESS Build #303 https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-mulitbranch/job/5.1/303/
Apache-Phoenix | 5.1 | HBase 2.4 | Build #303 SUCCESS
5.1 branch HBase 2.4 build #303 status SUCCESS Build #303 https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-mulitbranch/job/5.1/303/
Apache-Phoenix | 5.1 | HBase 2.5 | Build #303 FAILURE
5.1 branch HBase 2.5 build #303 status FAILURE Build #303 https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-mulitbranch/job/5.1/303/
(phoenix) branch 5.1 updated: PHOENIX-7138 Import netty-bom to make sure all netty 4.1 modules share the same version
This is an automated email from the ASF dual-hosted git repository. stoty pushed a commit to branch 5.1 in repository https://gitbox.apache.org/repos/asf/phoenix.git The following commit(s) were added to refs/heads/5.1 by this push: new db2541c129 PHOENIX-7138 Import netty-bom to make sure all netty 4.1 modules share the same version db2541c129 is described below commit db2541c12957113f6c6ac2e605a9aa7529f9ae09 Author: Istvan Toth AuthorDate: Mon Jan 8 15:37:38 2024 +0100 PHOENIX-7138 Import netty-bom to make sure all netty 4.1 modules share the same version --- pom.xml | 11 +++ 1 file changed, 11 insertions(+) diff --git a/pom.xml b/pom.xml index 9db2b381af..ddf4ac6495 100644 --- a/pom.xml +++ b/pom.xml @@ -105,6 +105,7 @@ 2.14.1 +4.1.104.Final 3.5.2 1.2.24 @@ -1383,6 +1384,16 @@ import pom + + +io.netty +netty-bom +${netty-bom.version} +import +pom + com.google.protobuf protobuf-java
(phoenix) branch master updated: PHOENIX-7138 Import netty-bom to make sure all netty 4.1 modules share the same version
This is an automated email from the ASF dual-hosted git repository. stoty pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/phoenix.git The following commit(s) were added to refs/heads/master by this push: new d331a15bfd PHOENIX-7138 Import netty-bom to make sure all netty 4.1 modules share the same version d331a15bfd is described below commit d331a15bfd90f908a2f4b16a5764ac1513c31408 Author: Istvan Toth AuthorDate: Mon Jan 8 15:37:38 2024 +0100 PHOENIX-7138 Import netty-bom to make sure all netty 4.1 modules share the same version --- pom.xml | 11 +++ 1 file changed, 11 insertions(+) diff --git a/pom.xml b/pom.xml index a2eae9..807baa4002 100644 --- a/pom.xml +++ b/pom.xml @@ -100,6 +100,7 @@ 2.14.1 +4.1.104.Final 3.5.2 1.2.19 @@ -1467,6 +1468,16 @@ import pom + + +io.netty +netty-bom +${netty-bom.version} +import +pom + com.google.protobuf protobuf-java
Apache-Phoenix | 5.1 | HBase 2.2 | Build #301 FAILURE
5.1 branch HBase 2.2 build #301 status FAILURE Build #301 https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-mulitbranch/job/5.1/301/
Apache-Phoenix | 5.1 | HBase 2.1 | Build #301 FAILURE
5.1 branch HBase 2.1 build #301 status FAILURE Build #301 https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-mulitbranch/job/5.1/301/
Apache-Phoenix | 5.1 | HBase 2.3 | Build #302 FAILURE
5.1 branch HBase 2.3 build #302 status FAILURE Build #302 https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-mulitbranch/job/5.1/302/
Apache-Phoenix | 5.1 | HBase 2.2 | Build #302 SUCCESS
5.1 branch HBase 2.2 build #302 status SUCCESS Build #302 https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-mulitbranch/job/5.1/302/
Apache-Phoenix | 5.1 | HBase 2.4 | Build #302 SUCCESS
5.1 branch HBase 2.4 build #302 status SUCCESS Build #302 https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-mulitbranch/job/5.1/302/
Apache-Phoenix | 5.1 | HBase 2.5 | Build #302 SUCCESS
5.1 branch HBase 2.5 build #302 status SUCCESS Build #302 https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-mulitbranch/job/5.1/302/
Apache-Phoenix | 5.1 | HBase 2.3 | Build #301 FAILURE
5.1 branch HBase 2.3 build #301 status FAILURE Build #301 https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-mulitbranch/job/5.1/301/
Apache-Phoenix | 5.1 | HBase 2.1 | Build #302 FAILURE
5.1 branch HBase 2.1 build #302 status FAILURE Build #302 https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-mulitbranch/job/5.1/302/
Apache-Phoenix | master | HBase 2.5 | Build #620 FAILURE
master branch HBase 2.5 build #620 status FAILURE Build #620 https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-mulitbranch/job/master/620/
Apache-Phoenix | master | HBase 2.4 | Build #620 FAILURE
master branch HBase 2.4 build #620 status FAILURE Build #620 https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-mulitbranch/job/master/620/
Apache-Phoenix | 5.1 | HBase 2.5 | Build #301 FAILURE
5.1 branch HBase 2.5 build #301 status FAILURE Build #301 https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-mulitbranch/job/5.1/301/
Apache-Phoenix | 5.1 | HBase 2.4 | Build #301 FAILURE
5.1 branch HBase 2.4 build #301 status FAILURE Build #301 https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-mulitbranch/job/5.1/301/
Apache-Phoenix | master | HBase 2.4 | Build #621 FAILURE
master branch HBase 2.4 build #621 status FAILURE Build #621 https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-mulitbranch/job/master/621/
Apache-Phoenix | master | HBase 2.5 | Build #621 FAILURE
master branch HBase 2.5 build #621 status FAILURE Build #621 https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-mulitbranch/job/master/621/
(phoenix) branch master updated: PHOENIX-7178 Use skipITs instead of skip to disable failsafe tests
This is an automated email from the ASF dual-hosted git repository. stoty pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/phoenix.git The following commit(s) were added to refs/heads/master by this push: new 1f7c3d4ee4 PHOENIX-7178 Use skipITs instead of skip to disable failsafe tests 1f7c3d4ee4 is described below commit 1f7c3d4ee453248421f012aeb7ce3151229880be Author: Istvan Toth AuthorDate: Thu Jan 11 10:01:34 2024 +0100 PHOENIX-7178 Use skipITs instead of skip to disable failsafe tests --- phoenix-tracing-webapp/pom.xml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/phoenix-tracing-webapp/pom.xml b/phoenix-tracing-webapp/pom.xml index 338cd3146d..58faa0abab 100755 --- a/phoenix-tracing-webapp/pom.xml +++ b/phoenix-tracing-webapp/pom.xml @@ -97,7 +97,7 @@ org.apache.maven.plugins maven-failsafe-plugin - true + true
(phoenix) branch 5.1 updated: PHOENIX-7178 Use skipITs instead of skip to disable failsafe tests
This is an automated email from the ASF dual-hosted git repository. stoty pushed a commit to branch 5.1 in repository https://gitbox.apache.org/repos/asf/phoenix.git The following commit(s) were added to refs/heads/5.1 by this push: new ae2da942d5 PHOENIX-7178 Use skipITs instead of skip to disable failsafe tests ae2da942d5 is described below commit ae2da942d5e772ca227b505c520dd019b771451f Author: Istvan Toth AuthorDate: Thu Jan 11 10:01:34 2024 +0100 PHOENIX-7178 Use skipITs instead of skip to disable failsafe tests --- phoenix-tracing-webapp/pom.xml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/phoenix-tracing-webapp/pom.xml b/phoenix-tracing-webapp/pom.xml index 05e9ee2f48..ecfc072046 100755 --- a/phoenix-tracing-webapp/pom.xml +++ b/phoenix-tracing-webapp/pom.xml @@ -86,7 +86,7 @@ org.apache.maven.plugins maven-failsafe-plugin - true + true
(phoenix) branch master updated: PHOENIX-7176 QueryTimeoutIT#testQueryTimeout fails with incorrect error message
This is an automated email from the ASF dual-hosted git repository. stoty pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/phoenix.git The following commit(s) were added to refs/heads/master by this push: new 2d7bfe6f1b PHOENIX-7176 QueryTimeoutIT#testQueryTimeout fails with incorrect error message 2d7bfe6f1b is described below commit 2d7bfe6f1bbe7244ffcdd04615bf36ee73d98143 Author: Aron Meszaros AuthorDate: Wed Jan 10 10:29:44 2024 +0100 PHOENIX-7176 QueryTimeoutIT#testQueryTimeout fails with incorrect error message --- .../org/apache/phoenix/end2end/QueryTimeoutIT.java | 35 -- 1 file changed, 20 insertions(+), 15 deletions(-) diff --git a/phoenix-core/src/it/java/org/apache/phoenix/end2end/QueryTimeoutIT.java b/phoenix-core/src/it/java/org/apache/phoenix/end2end/QueryTimeoutIT.java index 27bc211db5..4103655e16 100644 --- a/phoenix-core/src/it/java/org/apache/phoenix/end2end/QueryTimeoutIT.java +++ b/phoenix-core/src/it/java/org/apache/phoenix/end2end/QueryTimeoutIT.java @@ -23,15 +23,19 @@ import static org.junit.Assert.assertFalse; import static org.junit.Assert.assertTrue; import static org.junit.Assert.fail; +import java.io.IOException; import java.sql.Connection; import java.sql.DriverManager; -import java.sql.PreparedStatement; import java.sql.ResultSet; import java.sql.SQLException; import java.sql.SQLTimeoutException; import java.util.Map; import java.util.Properties; +import org.apache.hadoop.hbase.coprocessor.ObserverContext; +import org.apache.hadoop.hbase.coprocessor.RegionCoprocessorEnvironment; +import org.apache.hadoop.hbase.coprocessor.SimpleRegionObserver; +import org.apache.hadoop.hbase.regionserver.InternalScanner; import org.apache.phoenix.exception.SQLExceptionCode; import org.apache.phoenix.jdbc.PhoenixConnection; import org.apache.phoenix.jdbc.PhoenixStatement; @@ -42,6 +46,7 @@ import org.apache.phoenix.query.QueryServicesOptions; import org.apache.phoenix.util.PropertiesUtil; import org.apache.phoenix.util.QueryUtil; import org.apache.phoenix.util.ReadOnlyProps; +import org.apache.phoenix.util.TestUtil; import org.junit.After; import org.junit.Before; import org.junit.BeforeClass; @@ -107,30 +112,19 @@ public class QueryTimeoutIT extends BaseTest { @Test public void testQueryTimeout() throws Exception { -int nRows = 6; Connection conn; Properties props = PropertiesUtil.deepCopy(TEST_PROPERTIES); conn = DriverManager.getConnection(getUrl(), props); conn.createStatement().execute( "CREATE TABLE " + tableName + "(k BIGINT PRIMARY KEY, v VARCHAR)"); -PreparedStatement stmt = conn.prepareStatement("UPSERT INTO " + tableName + " VALUES(?, '')"); -for (int i = 1; i <= nRows; i++) { -stmt.setLong(1, i); -stmt.executeUpdate(); -if ((i % 2000) == 0) { -conn.commit(); -} -} -conn.commit(); -conn.createStatement().execute("UPDATE STATISTICS " + tableName); - +TestUtil.addCoprocessor(conn, tableName, QueryTimeoutIT.SleepingRegionObserver.class); + PhoenixStatement pstmt = conn.createStatement().unwrap(PhoenixStatement.class); pstmt.setQueryTimeout(1); long startTime = System.currentTimeMillis(); try { ResultSet rs = pstmt.executeQuery("SELECT count(*) FROM " + tableName); -// Force lots of chunks so query is cancelled -assertTrue(pstmt.getQueryPlan().getSplits().size() > 1000); +startTime = System.currentTimeMillis(); rs.next(); fail("Total time of query was " + (System.currentTimeMillis() - startTime) + " ms, but expected to be greater than 1000"); } catch (SQLTimeoutException e) { @@ -140,4 +134,15 @@ public class QueryTimeoutIT extends BaseTest { } conn.close(); } + +public static class SleepingRegionObserver extends SimpleRegionObserver { +@Override +public void preScannerClose(ObserverContext c, InternalScanner s) +throws IOException { +try { +Thread.sleep(1200); // Wait long enough +} catch (InterruptedException e) { +} +} +} }
(phoenix) branch 5.1 updated: PHOENIX-7176 QueryTimeoutIT#testQueryTimeout fails with incorrect error message
This is an automated email from the ASF dual-hosted git repository. stoty pushed a commit to branch 5.1 in repository https://gitbox.apache.org/repos/asf/phoenix.git The following commit(s) were added to refs/heads/5.1 by this push: new 646d63b2b0 PHOENIX-7176 QueryTimeoutIT#testQueryTimeout fails with incorrect error message 646d63b2b0 is described below commit 646d63b2b098bf9185e49e55941b79028518f09c Author: Aron Meszaros AuthorDate: Wed Jan 10 10:29:44 2024 +0100 PHOENIX-7176 QueryTimeoutIT#testQueryTimeout fails with incorrect error message --- .../org/apache/phoenix/end2end/QueryTimeoutIT.java | 35 -- 1 file changed, 20 insertions(+), 15 deletions(-) diff --git a/phoenix-core/src/it/java/org/apache/phoenix/end2end/QueryTimeoutIT.java b/phoenix-core/src/it/java/org/apache/phoenix/end2end/QueryTimeoutIT.java index 27bc211db5..4103655e16 100644 --- a/phoenix-core/src/it/java/org/apache/phoenix/end2end/QueryTimeoutIT.java +++ b/phoenix-core/src/it/java/org/apache/phoenix/end2end/QueryTimeoutIT.java @@ -23,15 +23,19 @@ import static org.junit.Assert.assertFalse; import static org.junit.Assert.assertTrue; import static org.junit.Assert.fail; +import java.io.IOException; import java.sql.Connection; import java.sql.DriverManager; -import java.sql.PreparedStatement; import java.sql.ResultSet; import java.sql.SQLException; import java.sql.SQLTimeoutException; import java.util.Map; import java.util.Properties; +import org.apache.hadoop.hbase.coprocessor.ObserverContext; +import org.apache.hadoop.hbase.coprocessor.RegionCoprocessorEnvironment; +import org.apache.hadoop.hbase.coprocessor.SimpleRegionObserver; +import org.apache.hadoop.hbase.regionserver.InternalScanner; import org.apache.phoenix.exception.SQLExceptionCode; import org.apache.phoenix.jdbc.PhoenixConnection; import org.apache.phoenix.jdbc.PhoenixStatement; @@ -42,6 +46,7 @@ import org.apache.phoenix.query.QueryServicesOptions; import org.apache.phoenix.util.PropertiesUtil; import org.apache.phoenix.util.QueryUtil; import org.apache.phoenix.util.ReadOnlyProps; +import org.apache.phoenix.util.TestUtil; import org.junit.After; import org.junit.Before; import org.junit.BeforeClass; @@ -107,30 +112,19 @@ public class QueryTimeoutIT extends BaseTest { @Test public void testQueryTimeout() throws Exception { -int nRows = 6; Connection conn; Properties props = PropertiesUtil.deepCopy(TEST_PROPERTIES); conn = DriverManager.getConnection(getUrl(), props); conn.createStatement().execute( "CREATE TABLE " + tableName + "(k BIGINT PRIMARY KEY, v VARCHAR)"); -PreparedStatement stmt = conn.prepareStatement("UPSERT INTO " + tableName + " VALUES(?, '')"); -for (int i = 1; i <= nRows; i++) { -stmt.setLong(1, i); -stmt.executeUpdate(); -if ((i % 2000) == 0) { -conn.commit(); -} -} -conn.commit(); -conn.createStatement().execute("UPDATE STATISTICS " + tableName); - +TestUtil.addCoprocessor(conn, tableName, QueryTimeoutIT.SleepingRegionObserver.class); + PhoenixStatement pstmt = conn.createStatement().unwrap(PhoenixStatement.class); pstmt.setQueryTimeout(1); long startTime = System.currentTimeMillis(); try { ResultSet rs = pstmt.executeQuery("SELECT count(*) FROM " + tableName); -// Force lots of chunks so query is cancelled -assertTrue(pstmt.getQueryPlan().getSplits().size() > 1000); +startTime = System.currentTimeMillis(); rs.next(); fail("Total time of query was " + (System.currentTimeMillis() - startTime) + " ms, but expected to be greater than 1000"); } catch (SQLTimeoutException e) { @@ -140,4 +134,15 @@ public class QueryTimeoutIT extends BaseTest { } conn.close(); } + +public static class SleepingRegionObserver extends SimpleRegionObserver { +@Override +public void preScannerClose(ObserverContext c, InternalScanner s) +throws IOException { +try { +Thread.sleep(1200); // Wait long enough +} catch (InterruptedException e) { +} +} +} }