Apache-Phoenix | master | HBase 2.2 | Build #110 SUCCESS
master branch HBase 2.2 build #110 status SUCCESS Build #110 https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-mulitbranch/job/master/110/
Apache-Phoenix | 4.x | HBase 1.4 | Build #101 SUCCESS
4.x branch HBase 1.4 build #101 status SUCCESS Build #101 https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-mulitbranch/job/4.x/101/
Apache-Phoenix | 4.x | HBase 1.6 | Build #101 SUCCESS
4.x branch HBase 1.6 build #101 status SUCCESS Build #101 https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-mulitbranch/job/4.x/101/
Apache-Phoenix | master | HBase 2.1 | Build #110 SUCCESS
master branch HBase 2.1 build #110 status SUCCESS Build #110 https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-mulitbranch/job/master/110/
Apache-Phoenix | 4.x | HBase 1.3 | Build #101 FAILURE
4.x branch HBase 1.3 build #101 status FAILURE Build #101 https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-mulitbranch/job/4.x/101/
Apache-Phoenix | master | HBase 2.3 | Build #110 FAILURE
master branch HBase 2.3 build #110 status FAILURE Build #110 https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-mulitbranch/job/master/110/
[phoenix] branch master updated: PHOENIX-5895 Leverage WALCellFilter in the SystemCatalogWALEntryFilter to replicate system catalog table
This is an automated email from the ASF dual-hosted git repository. gjacoby pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/phoenix.git The following commit(s) were added to refs/heads/master by this push: new 6cc9d50 PHOENIX-5895 Leverage WALCellFilter in the SystemCatalogWALEntryFilter to replicate system catalog table 6cc9d50 is described below commit 6cc9d5030da355abc6c8ab9eac9cc32097113e17 Author: Sandeep Pal AuthorDate: Sun Nov 1 22:38:24 2020 -0800 PHOENIX-5895 Leverage WALCellFilter in the SystemCatalogWALEntryFilter to replicate system catalog table --- .../replication/SystemCatalogWALEntryFilterIT.java | 17 +--- .../replication/SystemCatalogWALEntryFilter.java | 51 +++--- 2 files changed, 38 insertions(+), 30 deletions(-) diff --git a/phoenix-core/src/it/java/org/apache/phoenix/replication/SystemCatalogWALEntryFilterIT.java b/phoenix-core/src/it/java/org/apache/phoenix/replication/SystemCatalogWALEntryFilterIT.java index df4d97c..4a704a6 100644 --- a/phoenix-core/src/it/java/org/apache/phoenix/replication/SystemCatalogWALEntryFilterIT.java +++ b/phoenix-core/src/it/java/org/apache/phoenix/replication/SystemCatalogWALEntryFilterIT.java @@ -31,6 +31,7 @@ import org.apache.hadoop.hbase.client.Get; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.Table; import org.apache.hadoop.hbase.io.ImmutableBytesWritable; +import org.apache.hadoop.hbase.replication.ChainWALEntryFilter; import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.hbase.wal.WAL; import org.apache.hadoop.hbase.wal.WALEdit; @@ -134,11 +135,17 @@ public class SystemCatalogWALEntryFilterIT extends ParallelStatsDisabledIT { //verify that the tenant view WAL.Entry passes the filter and the non-tenant view does not SystemCatalogWALEntryFilter filter = new SystemCatalogWALEntryFilter(); -Assert.assertNull(filter.filter(nonTenantEntry)); -WAL.Entry filteredTenantEntry = filter.filter(tenantEntry); +// Chain the system catalog WAL entry filter to ChainWALEntryFilter +ChainWALEntryFilter chainWALEntryFilter = new ChainWALEntryFilter(filter); +// Asserting the WALEdit for non tenant has cells before getting filtered +Assert.assertTrue(nonTenantEntry.getEdit().size() > 0); +// All the cells will get removed by the filter since they do not belong to tenant +Assert.assertTrue("Non tenant edits for system catalog should not get filtered", +chainWALEntryFilter.filter(nonTenantEntry).getEdit().isEmpty()); +WAL.Entry filteredTenantEntry = chainWALEntryFilter.filter(tenantEntry); Assert.assertNotNull("Tenant view was filtered when it shouldn't be!", filteredTenantEntry); -Assert.assertEquals(tenantEntry.getEdit().size(), -filter.filter(tenantEntry).getEdit().size()); +Assert.assertEquals("filtered entry is not correct", +tenantEntry.getEdit().size(), filteredTenantEntry.getEdit().size()); //now check that a WAL.Entry with cells from both a tenant and a non-tenant //catalog row only allow the tenant cells through @@ -150,7 +157,7 @@ public class SystemCatalogWALEntryFilterIT extends ParallelStatsDisabledIT { Assert.assertEquals(tenantEntry.getEdit().size() + nonTenantEntry.getEdit().size() , comboEntry.getEdit().size()); Assert.assertEquals(tenantEntry.getEdit().size(), -filter.filter(comboEntry).getEdit().size()); +chainWALEntryFilter.filter(comboEntry).getEdit().size()); } public Get getGet(PTable catalogTable, byte[] tenantId, String viewName) { diff --git a/phoenix-core/src/main/java/org/apache/phoenix/replication/SystemCatalogWALEntryFilter.java b/phoenix-core/src/main/java/org/apache/phoenix/replication/SystemCatalogWALEntryFilter.java index 6e4c532..1f11807 100644 --- a/phoenix-core/src/main/java/org/apache/phoenix/replication/SystemCatalogWALEntryFilter.java +++ b/phoenix-core/src/main/java/org/apache/phoenix/replication/SystemCatalogWALEntryFilter.java @@ -17,16 +17,13 @@ */ package org.apache.phoenix.replication; -import java.util.List; - import org.apache.hadoop.hbase.Cell; -import org.apache.hadoop.hbase.io.ImmutableBytesWritable; +import org.apache.hadoop.hbase.replication.WALCellFilter; import org.apache.hadoop.hbase.replication.WALEntryFilter; import org.apache.hadoop.hbase.wal.WAL; import org.apache.phoenix.query.QueryConstants; import org.apache.phoenix.util.SchemaUtil; -import org.apache.phoenix.thirdparty.com.google.common.collect.Lists; /** * Standard replication of the SYSTEM.CATALOG table can be dangerous because schemas @@ -35,36 +32,40 @@ import org.apache.phoenix.thirdparty.com.google.common.collect.Lists; * be copied. This WALEntryFilter will only allow tenant-owned rows in SYSTEM.CATALOG to * be replicated. Data from all other tables is automatically passed. */ -public class SystemCatalogWALEntryFilter implements
[phoenix] branch master updated: PHOENIX-6212: Improve SystemCatalogIT.testSystemTableSplit() to ensure no splitting occurs when splitting is disabled
This is an automated email from the ASF dual-hosted git repository. chinmayskulkarni pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/phoenix.git The following commit(s) were added to refs/heads/master by this push: new 2ad33a1 PHOENIX-6212: Improve SystemCatalogIT.testSystemTableSplit() to ensure no splitting occurs when splitting is disabled 2ad33a1 is described below commit 2ad33a19b8d380bb1be411422de8e005047f76a1 Author: Chinmay Kulkarni AuthorDate: Tue Nov 10 17:57:59 2020 -0800 PHOENIX-6212: Improve SystemCatalogIT.testSystemTableSplit() to ensure no splitting occurs when splitting is disabled --- .../phoenix/end2end/SplitSystemCatalogIT.java | 14 ++- .../end2end/SystemCatalogRollbackEnabledIT.java| 28 +- .../java/org/apache/phoenix/query/BaseTest.java| 3 ++- 3 files changed, 27 insertions(+), 18 deletions(-) diff --git a/phoenix-core/src/it/java/org/apache/phoenix/end2end/SplitSystemCatalogIT.java b/phoenix-core/src/it/java/org/apache/phoenix/end2end/SplitSystemCatalogIT.java index 414f10c..9560a57 100644 --- a/phoenix-core/src/it/java/org/apache/phoenix/end2end/SplitSystemCatalogIT.java +++ b/phoenix-core/src/it/java/org/apache/phoenix/end2end/SplitSystemCatalogIT.java @@ -19,7 +19,6 @@ package org.apache.phoenix.end2end; import java.sql.Connection; import java.sql.DriverManager; -import java.sql.SQLException; import java.util.ArrayList; import java.util.Collections; import java.util.List; @@ -54,7 +53,8 @@ public class SplitSystemCatalogIT extends BaseTest { doSetup(null); } -public static synchronized void doSetup(Map props) throws Exception { +public static synchronized void doSetup(Map props) +throws Exception { NUM_SLAVES_BASE = 6; if (props == null) { props = Collections.emptyMap(); @@ -69,18 +69,20 @@ public class SplitSystemCatalogIT extends BaseTest { } } -protected static void splitSystemCatalog() throws SQLException, Exception { -try (Connection conn = DriverManager.getConnection(getUrl())) { +protected static void splitSystemCatalog() throws Exception { +try (Connection ignored = DriverManager.getConnection(getUrl())) { } String tableName = "TABLE"; String fullTableName1 = SchemaUtil.getTableName(SCHEMA1, tableName); String fullTableName2 = SchemaUtil.getTableName(SCHEMA2, tableName); String fullTableName3 = SchemaUtil.getTableName(SCHEMA3, tableName); String fullTableName4 = SchemaUtil.getTableName(SCHEMA4, tableName); -ArrayList tableList = Lists.newArrayList(fullTableName1, fullTableName2, fullTableName3); +ArrayList tableList = Lists.newArrayList(fullTableName1, +fullTableName2, fullTableName3); Map> tenantToTableMap = Maps.newHashMap(); tenantToTableMap.put(null, tableList); -tenantToTableMap.put(TENANT1, Lists.newArrayList(fullTableName2, fullTableName3)); +tenantToTableMap.put(TENANT1, Lists.newArrayList(fullTableName2, +fullTableName3)); tenantToTableMap.put(TENANT2, Lists.newArrayList(fullTableName4)); splitSystemCatalog(tenantToTableMap); } diff --git a/phoenix-core/src/it/java/org/apache/phoenix/end2end/SystemCatalogRollbackEnabledIT.java b/phoenix-core/src/it/java/org/apache/phoenix/end2end/SystemCatalogRollbackEnabledIT.java index 5a34a4f..fb2e379 100644 --- a/phoenix-core/src/it/java/org/apache/phoenix/end2end/SystemCatalogRollbackEnabledIT.java +++ b/phoenix-core/src/it/java/org/apache/phoenix/end2end/SystemCatalogRollbackEnabledIT.java @@ -22,6 +22,7 @@ import static org.junit.Assert.assertFalse; import static org.junit.Assert.assertTrue; import static org.junit.Assert.fail; +import java.io.IOException; import java.sql.Connection; import java.sql.DriverManager; import java.sql.ResultSet; @@ -51,8 +52,7 @@ import org.junit.experimental.categories.Category; * Tests various scenarios when * {@link QueryServices#ALLOW_SPLITTABLE_SYSTEM_CATALOG_ROLLBACK} * is set to true and SYSTEM.CATALOG should not be allowed to split. - * Note that this config must - * be set on both the client and server + * Note that this config must be set on both the client and server */ @Category(NeedsOwnMiniClusterTest.class) public class SystemCatalogRollbackEnabledIT extends BaseTest { @@ -99,37 +99,43 @@ public class SystemCatalogRollbackEnabledIT extends BaseTest { return DriverManager.getConnection(getUrl(), tenantProps); } +private void assertNumRegions(HBaseTestingUtility testUtil, +TableName tableName, int expectedNumRegions) throws IOException { +RegionLocator rl = testUtil.getConnection().getRegionLocator(tableName); +assertEquals(expectedNumRegions, rl.getAllRegionLocations().size()); +} /** * Make sure that SYSTEM.CATALOG
[phoenix] branch 4.x updated: PHOENIX-6212: Improve SystemCatalogIT.testSystemTableSplit() to ensure no splitting occurs when splitting is disabled
This is an automated email from the ASF dual-hosted git repository. chinmayskulkarni pushed a commit to branch 4.x in repository https://gitbox.apache.org/repos/asf/phoenix.git The following commit(s) were added to refs/heads/4.x by this push: new 565b0ea PHOENIX-6212: Improve SystemCatalogIT.testSystemTableSplit() to ensure no splitting occurs when splitting is disabled 565b0ea is described below commit 565b0eaa17de82e4bf6ea9634a5a30473e4167bf Author: Chinmay Kulkarni AuthorDate: Tue Nov 10 17:57:59 2020 -0800 PHOENIX-6212: Improve SystemCatalogIT.testSystemTableSplit() to ensure no splitting occurs when splitting is disabled --- .../phoenix/end2end/SplitSystemCatalogIT.java | 14 + .../end2end/SystemCatalogRollbackEnabledIT.java| 36 ++ .../java/org/apache/phoenix/query/BaseTest.java| 3 +- 3 files changed, 26 insertions(+), 27 deletions(-) diff --git a/phoenix-core/src/it/java/org/apache/phoenix/end2end/SplitSystemCatalogIT.java b/phoenix-core/src/it/java/org/apache/phoenix/end2end/SplitSystemCatalogIT.java index dce530f..b2075a7 100644 --- a/phoenix-core/src/it/java/org/apache/phoenix/end2end/SplitSystemCatalogIT.java +++ b/phoenix-core/src/it/java/org/apache/phoenix/end2end/SplitSystemCatalogIT.java @@ -19,7 +19,6 @@ package org.apache.phoenix.end2end; import java.sql.Connection; import java.sql.DriverManager; -import java.sql.SQLException; import java.util.ArrayList; import java.util.Collections; import java.util.List; @@ -54,7 +53,8 @@ public class SplitSystemCatalogIT extends BaseTest { doSetup(null); } -public static synchronized void doSetup(Map props) throws Exception { +public static synchronized void doSetup(Map props) +throws Exception { NUM_SLAVES_BASE = 6; if (props == null) { props = Collections.emptyMap(); @@ -69,18 +69,20 @@ public class SplitSystemCatalogIT extends BaseTest { } } -protected static void splitSystemCatalog() throws SQLException, Exception { -try (Connection conn = DriverManager.getConnection(getUrl())) { +protected static void splitSystemCatalog() throws Exception { +try (Connection ignored = DriverManager.getConnection(getUrl())) { } String tableName = "TABLE"; String fullTableName1 = SchemaUtil.getTableName(SCHEMA1, tableName); String fullTableName2 = SchemaUtil.getTableName(SCHEMA2, tableName); String fullTableName3 = SchemaUtil.getTableName(SCHEMA3, tableName); String fullTableName4 = SchemaUtil.getTableName(SCHEMA4, tableName); -ArrayList tableList = Lists.newArrayList(fullTableName1, fullTableName2, fullTableName3); +ArrayList tableList = Lists.newArrayList(fullTableName1, +fullTableName2, fullTableName3); Map> tenantToTableMap = Maps.newHashMap(); tenantToTableMap.put(null, tableList); -tenantToTableMap.put(TENANT1, Lists.newArrayList(fullTableName2, fullTableName3)); +tenantToTableMap.put(TENANT1, Lists.newArrayList(fullTableName2, +fullTableName3)); tenantToTableMap.put(TENANT2, Lists.newArrayList(fullTableName4)); splitSystemCatalog(tenantToTableMap); } diff --git a/phoenix-core/src/it/java/org/apache/phoenix/end2end/SystemCatalogRollbackEnabledIT.java b/phoenix-core/src/it/java/org/apache/phoenix/end2end/SystemCatalogRollbackEnabledIT.java index 2b69596..0241eab 100644 --- a/phoenix-core/src/it/java/org/apache/phoenix/end2end/SystemCatalogRollbackEnabledIT.java +++ b/phoenix-core/src/it/java/org/apache/phoenix/end2end/SystemCatalogRollbackEnabledIT.java @@ -22,6 +22,7 @@ import static org.junit.Assert.assertFalse; import static org.junit.Assert.assertTrue; import static org.junit.Assert.fail; +import java.io.IOException; import java.sql.Connection; import java.sql.DriverManager; import java.sql.ResultSet; @@ -31,7 +32,6 @@ import java.util.HashMap; import java.util.Map; import java.util.Properties; -import org.apache.hadoop.hbase.DoNotRetryIOException; import org.apache.hadoop.hbase.HBaseTestingUtility; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.client.RegionLocator; @@ -51,8 +51,7 @@ import org.junit.experimental.categories.Category; * Tests various scenarios when * {@link QueryServices#ALLOW_SPLITTABLE_SYSTEM_CATALOG_ROLLBACK} * is set to true and SYSTEM.CATALOG should not be allowed to split. - * Note that this config must - * be set on both the client and server + * Note that this config must be set on both the client and server */ @Category(NeedsOwnMiniClusterTest.class) public class SystemCatalogRollbackEnabledIT extends BaseTest { @@ -100,37 +99,34 @@ public class SystemCatalogRollbackEnabledIT extends BaseTest { return DriverManager.getConnection(getUrl(), tenantProps); } +private void assertNumRegions(HBaseTestingUtility testUtil, +
[phoenix] branch 4.x-PHOENIX-5923 created (now c9c80b2)
This is an automated email from the ASF dual-hosted git repository. gokcen pushed a change to branch 4.x-PHOENIX-5923 in repository https://gitbox.apache.org/repos/asf/phoenix.git. at c9c80b2 PHOENIX-5955 OrphanViewToolIT is flapping No new revisions were added by this update.
[phoenix-omid] branch master updated: OMID-120 Utilize protobuf-maven-plugin for build
This is an automated email from the ASF dual-hosted git repository. stoty pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/phoenix-omid.git The following commit(s) were added to refs/heads/master by this push: new ebe39cd OMID-120 Utilize protobuf-maven-plugin for build ebe39cd is described below commit ebe39cd3ae755ad7ee65842e6f634feeaf51fda5 Author: Chia-Ping Tsai AuthorDate: Wed Nov 11 19:46:34 2020 +0800 OMID-120 Utilize protobuf-maven-plugin for build --- .travis.yml| 5 - common/pom.xml | 36 +++- pom.xml| 3 ++- 3 files changed, 17 insertions(+), 27 deletions(-) diff --git a/.travis.yml b/.travis.yml index dbfc9e3..535f249 100644 --- a/.travis.yml +++ b/.travis.yml @@ -28,11 +28,6 @@ before_script: - git config --global user.email "d...@omid.incubator.apache.org" - git config --global user.name "Omid CI" - REPO_DIR=$(pwd) - # Install protobuf to genearte TSO client-server protocol in each compilation - - cd .. - - wget https://github.com/google/protobuf/releases/download/v2.5.0/protobuf-2.5.0.tar.gz - - tar -xzvf protobuf-2.5.0.tar.gz - - cd protobuf-2.5.0 && ./configure --prefix=/usr && make && sudo make install - cd $REPO_DIR script: diff --git a/common/pom.xml b/common/pom.xml index bfef674..e39a276 100644 --- a/common/pom.xml +++ b/common/pom.xml @@ -88,37 +88,31 @@ + + +kr.motd.maven +os-maven-plugin +${os.plugin.version} + + -maven-antrun-plugin -${maven-antrun-plugin.version} +org.xolstice.maven.plugins +protobuf-maven-plugin +${protobuf.plugin.version} +compile-protoc generate-sources -default-cli - - - - - - - - - - - - - - - - -run +compile + + com.google.protobuf:protoc:${protobuf.version}:exe:${os.detected.classifier} + - org.codehaus.mojo build-helper-maven-plugin diff --git a/pom.xml b/pom.xml index 788f3a0..fa25a57 100644 --- a/pom.xml +++ b/pom.xml @@ -182,7 +182,8 @@ 1.2.17 3.2.6.Final 2.5.0 - +0.6.1 +1.6.2 4.13 1.9.5 3.2.0