Apache-Phoenix | 4.x | HBase 1.6 | Build #168 SUCCESS
4.x branch HBase 1.6 build #168 status SUCCESS Build #168 https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-mulitbranch/job/4.x/168/
Apache-Phoenix | 4.x | HBase 1.3 | Build #168 FAILURE
4.x branch HBase 1.3 build #168 status FAILURE Build #168 https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-mulitbranch/job/4.x/168/
Apache-Phoenix | 4.x | HBase 1.4 | Build #168 FAILURE
4.x branch HBase 1.4 build #168 status FAILURE Build #168 https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-mulitbranch/job/4.x/168/
Apache-Phoenix | master | HBase 2.2 | Build #173 SUCCESS
master branch HBase 2.2 build #173 status SUCCESS Build #173 https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-mulitbranch/job/master/173/
Apache-Phoenix | master | HBase 2.3 | Build #173 FAILURE
master branch HBase 2.3 build #173 status FAILURE Build #173 https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-mulitbranch/job/master/173/
Apache-Phoenix | master | HBase 2.1 | Build #173 SUCCESS
master branch HBase 2.1 build #173 status SUCCESS Build #173 https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-mulitbranch/job/master/173/
Apache-Phoenix | master | HBase 2.2 | Build #172 SUCCESS
master branch HBase 2.2 build #172 status SUCCESS Build #172 https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-mulitbranch/job/master/172/
Apache-Phoenix | 4.x | HBase 1.6 | Build #167 SUCCESS
4.x branch HBase 1.6 build #167 status SUCCESS Build #167 https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-mulitbranch/job/4.x/167/
Apache-Phoenix | 4.x | HBase 1.4 | Build #167 SUCCESS
4.x branch HBase 1.4 build #167 status SUCCESS Build #167 https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-mulitbranch/job/4.x/167/
Apache-Phoenix | 4.x | HBase 1.3 | Build #167 FAILURE
4.x branch HBase 1.3 build #167 status FAILURE Build #167 https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-mulitbranch/job/4.x/167/
Apache-Phoenix | master | HBase 2.3 | Build #172 FAILURE
master branch HBase 2.3 build #172 status FAILURE Build #172 https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-mulitbranch/job/master/172/
Apache-Phoenix | master | HBase 2.1 | Build #172 FAILURE
master branch HBase 2.1 build #172 status FAILURE Build #172 https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-mulitbranch/job/master/172/
[phoenix] branch 4.x updated: PHOENIX-6288 Minicluster startup problems on Jenkins
This is an automated email from the ASF dual-hosted git repository. stoty pushed a commit to branch 4.x in repository https://gitbox.apache.org/repos/asf/phoenix.git The following commit(s) were added to refs/heads/4.x by this push: new 237b724 PHOENIX-6288 Minicluster startup problems on Jenkins 237b724 is described below commit 237b7248dafe51087a8a95aa5bc67e5ce141e89d Author: Istvan Toth AuthorDate: Wed Dec 30 16:16:47 2020 +0100 PHOENIX-6288 Minicluster startup problems on Jenkins increase minicluster master startup timeout to 200 seconds log minicluster startup time --- phoenix-core/src/test/java/org/apache/phoenix/query/BaseTest.java | 5 + 1 file changed, 5 insertions(+) diff --git a/phoenix-core/src/test/java/org/apache/phoenix/query/BaseTest.java b/phoenix-core/src/test/java/org/apache/phoenix/query/BaseTest.java index bbe0371..05773fb 100644 --- a/phoenix-core/src/test/java/org/apache/phoenix/query/BaseTest.java +++ b/phoenix-core/src/test/java/org/apache/phoenix/query/BaseTest.java @@ -547,7 +547,10 @@ public abstract class BaseTest { setUpConfigForMiniCluster(conf, overrideProps); utility = new HBaseTestingUtility(conf); try { +long startTime = System.currentTimeMillis(); utility.startMiniCluster(NUM_SLAVES_BASE); +long startupTime = System.currentTimeMillis()-startTime; +LOGGER.info("HBase minicluster startup complete in {} ms", startupTime); return getLocalClusterUrl(utility); } catch (Throwable t) { throw new RuntimeException(t); @@ -626,6 +629,8 @@ public abstract class BaseTest { conf.setInt("hbase.assignment.zkevent.workers", 5); conf.setInt("hbase.assignment.threads.max", 5); conf.setInt("hbase.catalogjanitor.interval", 5000); +//Allow for an extra long miniCluster startup time in case of an overloaded test machine +conf.setInt("hbase.master.start.timeout.localHBaseCluster", 20); conf.setInt(QueryServices.TASK_HANDLING_INTERVAL_MS_ATTRIB, 1); conf.setInt(HConstants.HBASE_CLIENT_RETRIES_NUMBER, 2); conf.setInt(NUM_CONCURRENT_INDEX_WRITER_THREADS_CONF_KEY, 1);
[phoenix] branch master updated: PHOENIX-6288 Minicluster startup problems on Jenkins
This is an automated email from the ASF dual-hosted git repository. stoty pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/phoenix.git The following commit(s) were added to refs/heads/master by this push: new d097857 PHOENIX-6288 Minicluster startup problems on Jenkins d097857 is described below commit d0978570a65b3191cff046abdfbe95d0d04f1121 Author: Istvan Toth AuthorDate: Wed Dec 30 16:16:47 2020 +0100 PHOENIX-6288 Minicluster startup problems on Jenkins increase minicluster master startup timeout to 200 seconds log minicluster startup time --- phoenix-core/src/test/java/org/apache/phoenix/query/BaseTest.java | 5 + 1 file changed, 5 insertions(+) diff --git a/phoenix-core/src/test/java/org/apache/phoenix/query/BaseTest.java b/phoenix-core/src/test/java/org/apache/phoenix/query/BaseTest.java index 2960081..4894727 100644 --- a/phoenix-core/src/test/java/org/apache/phoenix/query/BaseTest.java +++ b/phoenix-core/src/test/java/org/apache/phoenix/query/BaseTest.java @@ -548,7 +548,10 @@ public abstract class BaseTest { setUpConfigForMiniCluster(conf, overrideProps); utility = new HBaseTestingUtility(conf); try { +long startTime = System.currentTimeMillis(); utility.startMiniCluster(NUM_SLAVES_BASE); +long startupTime = System.currentTimeMillis()-startTime; +LOGGER.info("HBase minicluster startup complete in {} ms", startupTime); return getLocalClusterUrl(utility); } catch (Throwable t) { throw new RuntimeException(t); @@ -628,6 +631,8 @@ public abstract class BaseTest { conf.setInt("hbase.assignment.zkevent.workers", 5); conf.setInt("hbase.assignment.threads.max", 5); conf.setInt("hbase.catalogjanitor.interval", 5000); +//Allow for an extra long miniCluster startup time in case of an overloaded test machine +conf.setInt("hbase.master.start.timeout.localHBaseCluster", 20); conf.setInt(QueryServices.TASK_HANDLING_INTERVAL_MS_ATTRIB, 1); conf.setInt(HConstants.HBASE_CLIENT_RETRIES_NUMBER, 2); conf.setInt(NUM_CONCURRENT_INDEX_WRITER_THREADS_CONF_KEY, 1);
[phoenix] branch 4.x updated: PHOENIX-6286 Extend syscat RegionSplitPolicies from the default Policy for the HBase branch
This is an automated email from the ASF dual-hosted git repository. stoty pushed a commit to branch 4.x in repository https://gitbox.apache.org/repos/asf/phoenix.git The following commit(s) were added to refs/heads/4.x by this push: new 7bdc3b4 PHOENIX-6286 Extend syscat RegionSplitPolicies from the default Policy for the HBase branch 7bdc3b4 is described below commit 7bdc3b4dce7e61fa73bf6d38b4a34c809fe54540 Author: Istvan Toth AuthorDate: Wed Dec 30 08:42:50 2020 +0100 PHOENIX-6286 Extend syscat RegionSplitPolicies from the default Policy for the HBase branch --- .../org/apache/phoenix/schema/SplitOnLeadingVarCharColumnsPolicy.java | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/phoenix-core/src/main/java/org/apache/phoenix/schema/SplitOnLeadingVarCharColumnsPolicy.java b/phoenix-core/src/main/java/org/apache/phoenix/schema/SplitOnLeadingVarCharColumnsPolicy.java index d481998..249038f 100644 --- a/phoenix-core/src/main/java/org/apache/phoenix/schema/SplitOnLeadingVarCharColumnsPolicy.java +++ b/phoenix-core/src/main/java/org/apache/phoenix/schema/SplitOnLeadingVarCharColumnsPolicy.java @@ -17,10 +17,10 @@ */ package org.apache.phoenix.schema; -import org.apache.hadoop.hbase.regionserver.ConstantSizeRegionSplitPolicy; +import org.apache.hadoop.hbase.regionserver.IncreasingToUpperBoundRegionSplitPolicy; import org.apache.phoenix.util.SchemaUtil; -public abstract class SplitOnLeadingVarCharColumnsPolicy extends ConstantSizeRegionSplitPolicy { +public abstract class SplitOnLeadingVarCharColumnsPolicy extends IncreasingToUpperBoundRegionSplitPolicy { abstract protected int getColumnToSplitAt(); protected final byte[] getSplitPoint(byte[] splitPoint) {
[phoenix] branch master updated: PHOENIX-6286 Extend syscat RegionSplitPolicies from the default Policy for the HBase branch
This is an automated email from the ASF dual-hosted git repository. stoty pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/phoenix.git The following commit(s) were added to refs/heads/master by this push: new 1b5c873 PHOENIX-6286 Extend syscat RegionSplitPolicies from the default Policy for the HBase branch 1b5c873 is described below commit 1b5c873cc8487aee5ef9a447d2d645e0980ca44d Author: Istvan Toth AuthorDate: Wed Dec 30 08:39:51 2020 +0100 PHOENIX-6286 Extend syscat RegionSplitPolicies from the default Policy for the HBase branch --- .../org/apache/phoenix/schema/SplitOnLeadingVarCharColumnsPolicy.java | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/phoenix-core/src/main/java/org/apache/phoenix/schema/SplitOnLeadingVarCharColumnsPolicy.java b/phoenix-core/src/main/java/org/apache/phoenix/schema/SplitOnLeadingVarCharColumnsPolicy.java index 6e0e856..80551f5 100644 --- a/phoenix-core/src/main/java/org/apache/phoenix/schema/SplitOnLeadingVarCharColumnsPolicy.java +++ b/phoenix-core/src/main/java/org/apache/phoenix/schema/SplitOnLeadingVarCharColumnsPolicy.java @@ -17,10 +17,10 @@ */ package org.apache.phoenix.schema; -import org.apache.hadoop.hbase.regionserver.ConstantSizeRegionSplitPolicy; +import org.apache.hadoop.hbase.regionserver.SteppingSplitPolicy; import org.apache.phoenix.util.SchemaUtil; -public abstract class SplitOnLeadingVarCharColumnsPolicy extends ConstantSizeRegionSplitPolicy { +public abstract class SplitOnLeadingVarCharColumnsPolicy extends SteppingSplitPolicy { abstract protected int getColumnToSplitAt(); protected final byte[] getSplitPoint(byte[] splitPoint) {
Apache-Phoenix | 4.x | HBase 1.4 | Build #166 SUCCESS
4.x branch HBase 1.4 build #166 status SUCCESS Build #166 https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-mulitbranch/job/4.x/166/
Apache-Phoenix | 4.x | HBase 1.6 | Build #165 SUCCESS
4.x branch HBase 1.6 build #165 status SUCCESS Build #165 https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-mulitbranch/job/4.x/165/
Apache-Phoenix | master | HBase 2.2 | Build #171 SUCCESS
master branch HBase 2.2 build #171 status SUCCESS Build #171 https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-mulitbranch/job/master/171/
Apache-Phoenix | master | HBase 2.3 | Build #171 SUCCESS
master branch HBase 2.3 build #171 status SUCCESS Build #171 https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-mulitbranch/job/master/171/
Apache-Phoenix | master | HBase 2.1 | Build #171 FAILURE
master branch HBase 2.1 build #171 status FAILURE Build #171 https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-mulitbranch/job/master/171/
Apache-Phoenix | 4.x | HBase 1.3 | Build #166 FAILURE
4.x branch HBase 1.3 build #166 status FAILURE Build #166 https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-mulitbranch/job/4.x/166/
Apache-Phoenix | 4.x | HBase 1.4 | Build #165 SUCCESS
4.x branch HBase 1.4 build #165 status SUCCESS Build #165 https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-mulitbranch/job/4.x/165/
Apache-Phoenix | master | HBase 2.2 | Build #170 FAILURE
master branch HBase 2.2 build #170 status FAILURE Build #170 https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-mulitbranch/job/master/170/
Apache-Phoenix | 4.x | HBase 1.3 | Build #165 FAILURE
4.x branch HBase 1.3 build #165 status FAILURE Build #165 https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-mulitbranch/job/4.x/165/
Apache-Phoenix | master | HBase 2.3 | Build #170 FAILURE
master branch HBase 2.3 build #170 status FAILURE Build #170 https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-mulitbranch/job/master/170/
Apache-Phoenix | master | HBase 2.1 | Build #170 FAILURE
master branch HBase 2.1 build #170 status FAILURE Build #170 https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-mulitbranch/job/master/170/
[phoenix] branch master updated: PHOENIX-6287 Fix incorrect log in ParallelIterators.submitWork
This is an automated email from the ASF dual-hosted git repository. chenglei pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/phoenix.git The following commit(s) were added to refs/heads/master by this push: new 3891050 PHOENIX-6287 Fix incorrect log in ParallelIterators.submitWork 3891050 is described below commit 38910507659976c8a12e5f0d355beef55add382d Author: chenglei AuthorDate: Wed Dec 30 19:09:45 2020 +0800 PHOENIX-6287 Fix incorrect log in ParallelIterators.submitWork --- .../org/apache/phoenix/iterate/ParallelIterators.java | 19 +-- 1 file changed, 13 insertions(+), 6 deletions(-) diff --git a/phoenix-core/src/main/java/org/apache/phoenix/iterate/ParallelIterators.java b/phoenix-core/src/main/java/org/apache/phoenix/iterate/ParallelIterators.java index 8947668..828de39 100644 --- a/phoenix-core/src/main/java/org/apache/phoenix/iterate/ParallelIterators.java +++ b/phoenix-core/src/main/java/org/apache/phoenix/iterate/ParallelIterators.java @@ -123,12 +123,12 @@ public class ParallelIterators extends BaseResultIterators { @Override public PeekingResultIterator call() throws Exception { long startTime = EnvironmentEdgeManager.currentTimeMillis(); -if (LOGGER.isDebugEnabled()) { -LOGGER.debug(LogUtil.addCustomAnnotations("Id: " + scanId + ", Time: " + -(EnvironmentEdgeManager.currentTimeMillis() - startTime) + -"ms, Scan: " + scan, ScanUtil.getCustomAnnotations(scan))); -} -PeekingResultIterator iterator = iteratorFactory.newIterator(context, tableResultItr, scan, physicalTableName, ParallelIterators.this.plan); +PeekingResultIterator iterator = iteratorFactory.newIterator( +context, +tableResultItr, +scan, +physicalTableName, +ParallelIterators.this.plan); if (initFirstScanOnly) { if ((!isReverse && scanLocation.isFirstScan()) || (isReverse && scanLocation.isLastScan())) { // Fill the scanner's cache. This helps reduce latency since we are parallelizing the I/O needed. @@ -137,6 +137,13 @@ public class ParallelIterators extends BaseResultIterators { } else { iterator.peek(); } + +if (LOGGER.isDebugEnabled()) { +LOGGER.debug(LogUtil.addCustomAnnotations("Id: " + scanId + ", Time: " + +(EnvironmentEdgeManager.currentTimeMillis() - startTime) + +"ms, Scan: " + scan, ScanUtil.getCustomAnnotations(scan))); +} + allIterators.add(iterator); return iterator; }
Apache-Phoenix | master | HBase 2.2 | Build #169 SUCCESS
master branch HBase 2.2 build #169 status SUCCESS Build #169 https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-mulitbranch/job/master/169/
[phoenix] branch 4.x updated: PHOENIX-6283 MutableIndexExtendedIT#testCompactDisabledIndex consistently fails with HBase 2.3
This is an automated email from the ASF dual-hosted git repository. stoty pushed a commit to branch 4.x in repository https://gitbox.apache.org/repos/asf/phoenix.git The following commit(s) were added to refs/heads/4.x by this push: new 0e8bef1 PHOENIX-6283 MutableIndexExtendedIT#testCompactDisabledIndex consistently fails with HBase 2.3 0e8bef1 is described below commit 0e8bef1f94bb5b0fb16dbe8da9198ef8158d02f1 Author: Istvan Toth AuthorDate: Wed Dec 30 08:15:26 2020 +0100 PHOENIX-6283 MutableIndexExtendedIT#testCompactDisabledIndex consistently fails with HBase 2.3 add missing test to 4.x --- .../end2end/NoLookbackMutableIndexExtendedIT.java | 158 + 1 file changed, 158 insertions(+) diff --git a/phoenix-core/src/it/java/org/apache/phoenix/end2end/NoLookbackMutableIndexExtendedIT.java b/phoenix-core/src/it/java/org/apache/phoenix/end2end/NoLookbackMutableIndexExtendedIT.java new file mode 100644 index 000..376a7b9 --- /dev/null +++ b/phoenix-core/src/it/java/org/apache/phoenix/end2end/NoLookbackMutableIndexExtendedIT.java @@ -0,0 +1,158 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ +package org.apache.phoenix.end2end; + +import static org.apache.phoenix.util.TestUtil.TEST_PROPERTIES; +import static org.junit.Assert.assertEquals; + +import java.sql.Connection; +import java.sql.DriverManager; +import java.sql.SQLException; +import java.util.Arrays; +import java.util.Collection; +import java.util.List; +import java.util.Map; +import java.util.Properties; + +import org.apache.hadoop.hbase.TableName; +import org.apache.hadoop.hbase.client.Table; +import org.apache.hadoop.hbase.regionserver.HRegion; +import org.apache.hadoop.hbase.util.Bytes; +import org.apache.phoenix.jdbc.PhoenixConnection; +import org.apache.phoenix.jdbc.PhoenixDatabaseMetaData; +import org.apache.phoenix.query.QueryServices; +import org.apache.phoenix.schema.PIndexState; +import org.apache.phoenix.thirdparty.com.google.common.collect.Maps; +import org.apache.phoenix.util.EnvironmentEdgeManager; +import org.apache.phoenix.util.IndexUtil; +import org.apache.phoenix.util.PropertiesUtil; +import org.apache.phoenix.util.ReadOnlyProps; +import org.apache.phoenix.util.SchemaUtil; +import org.apache.phoenix.util.TestUtil; +import org.junit.BeforeClass; +import org.junit.Test; +import org.junit.experimental.categories.Category; +import org.junit.runner.RunWith; +import org.junit.runners.Parameterized; + +@RunWith(Parameterized.class) +@Category(NeedsOwnMiniClusterTest.class) +public class NoLookbackMutableIndexExtendedIT extends BaseUniqueNamesOwnClusterIT { +protected final boolean localIndex; +protected final String tableDDLOptions; + +@BeforeClass +public static synchronized void doSetup() throws Exception { +Map props = Maps.newHashMapWithExpectedSize(1); +setUpTestDriver(new ReadOnlyProps(props.entrySet().iterator())); +} + +public NoLookbackMutableIndexExtendedIT(Boolean localIndex, String txProvider, Boolean columnEncoded) { +this.localIndex = localIndex; +StringBuilder optionBuilder = new StringBuilder(); +if (txProvider != null) { +optionBuilder +.append("TRANSACTIONAL=true," + PhoenixDatabaseMetaData.TRANSACTION_PROVIDER ++ "='" + txProvider + "'"); +} +if (!columnEncoded) { +if (optionBuilder.length() != 0) optionBuilder.append(","); +optionBuilder.append("COLUMN_ENCODED_BYTES=0"); +} +this.tableDDLOptions = optionBuilder.toString(); +} + +private static Connection getConnection(Properties props) throws SQLException { + props.setProperty(QueryServices.INDEX_MUTATE_BATCH_SIZE_THRESHOLD_ATTRIB, +Integer.toString(1)); +Connection conn = DriverManager.getConnection(getUrl(), props); +return conn; +} + +protected static Connection getConnection() throws SQLException { +Properties props = PropertiesUtil.deepCopy(TEST_PROPERTIES); +return getConnection(props); +} + +@Parameterized.Parameters(name = "NoLookbackMutableIndexExtendedIT_localIndex={0},transactionProvider={
Apache-Phoenix | master | HBase 2.1 | Build #169 SUCCESS
master branch HBase 2.1 build #169 status SUCCESS Build #169 https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-mulitbranch/job/master/169/
[phoenix] branch master updated: PHOENIX-6283 MutableIndexExtendedIT#testCompactDisabledIndex consistently fails with HBase 2.3
This is an automated email from the ASF dual-hosted git repository. stoty pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/phoenix.git The following commit(s) were added to refs/heads/master by this push: new c04a07b PHOENIX-6283 MutableIndexExtendedIT#testCompactDisabledIndex consistently fails with HBase 2.3 c04a07b is described below commit c04a07b25172e5bb195d71a5d5615a2c5295773c Author: Istvan Toth AuthorDate: Wed Dec 30 08:15:26 2020 +0100 PHOENIX-6283 MutableIndexExtendedIT#testCompactDisabledIndex consistently fails with HBase 2.3 --- .../end2end/NoLookbackMutableIndexExtendedIT.java | 158 + .../end2end/index/MutableIndexExtendedIT.java | 59 2 files changed, 158 insertions(+), 59 deletions(-) diff --git a/phoenix-core/src/it/java/org/apache/phoenix/end2end/NoLookbackMutableIndexExtendedIT.java b/phoenix-core/src/it/java/org/apache/phoenix/end2end/NoLookbackMutableIndexExtendedIT.java new file mode 100644 index 000..376a7b9 --- /dev/null +++ b/phoenix-core/src/it/java/org/apache/phoenix/end2end/NoLookbackMutableIndexExtendedIT.java @@ -0,0 +1,158 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ +package org.apache.phoenix.end2end; + +import static org.apache.phoenix.util.TestUtil.TEST_PROPERTIES; +import static org.junit.Assert.assertEquals; + +import java.sql.Connection; +import java.sql.DriverManager; +import java.sql.SQLException; +import java.util.Arrays; +import java.util.Collection; +import java.util.List; +import java.util.Map; +import java.util.Properties; + +import org.apache.hadoop.hbase.TableName; +import org.apache.hadoop.hbase.client.Table; +import org.apache.hadoop.hbase.regionserver.HRegion; +import org.apache.hadoop.hbase.util.Bytes; +import org.apache.phoenix.jdbc.PhoenixConnection; +import org.apache.phoenix.jdbc.PhoenixDatabaseMetaData; +import org.apache.phoenix.query.QueryServices; +import org.apache.phoenix.schema.PIndexState; +import org.apache.phoenix.thirdparty.com.google.common.collect.Maps; +import org.apache.phoenix.util.EnvironmentEdgeManager; +import org.apache.phoenix.util.IndexUtil; +import org.apache.phoenix.util.PropertiesUtil; +import org.apache.phoenix.util.ReadOnlyProps; +import org.apache.phoenix.util.SchemaUtil; +import org.apache.phoenix.util.TestUtil; +import org.junit.BeforeClass; +import org.junit.Test; +import org.junit.experimental.categories.Category; +import org.junit.runner.RunWith; +import org.junit.runners.Parameterized; + +@RunWith(Parameterized.class) +@Category(NeedsOwnMiniClusterTest.class) +public class NoLookbackMutableIndexExtendedIT extends BaseUniqueNamesOwnClusterIT { +protected final boolean localIndex; +protected final String tableDDLOptions; + +@BeforeClass +public static synchronized void doSetup() throws Exception { +Map props = Maps.newHashMapWithExpectedSize(1); +setUpTestDriver(new ReadOnlyProps(props.entrySet().iterator())); +} + +public NoLookbackMutableIndexExtendedIT(Boolean localIndex, String txProvider, Boolean columnEncoded) { +this.localIndex = localIndex; +StringBuilder optionBuilder = new StringBuilder(); +if (txProvider != null) { +optionBuilder +.append("TRANSACTIONAL=true," + PhoenixDatabaseMetaData.TRANSACTION_PROVIDER ++ "='" + txProvider + "'"); +} +if (!columnEncoded) { +if (optionBuilder.length() != 0) optionBuilder.append(","); +optionBuilder.append("COLUMN_ENCODED_BYTES=0"); +} +this.tableDDLOptions = optionBuilder.toString(); +} + +private static Connection getConnection(Properties props) throws SQLException { + props.setProperty(QueryServices.INDEX_MUTATE_BATCH_SIZE_THRESHOLD_ATTRIB, +Integer.toString(1)); +Connection conn = DriverManager.getConnection(getUrl(), props); +return conn; +} + +protected static Connection getConnection() throws SQLException { +Properties props = PropertiesUtil.deepCopy(TEST_PROPERTIES); +return getConnection(props); +} + +@Parameterized.Parameters(name = "NoLookbackM
Apache-Phoenix | master | HBase 2.3 | Build #169 FAILURE
master branch HBase 2.3 build #169 status FAILURE Build #169 https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-mulitbranch/job/master/169/
[phoenix] branch 4.x updated: PHOENIX-6287 Fix incorrect log in ParallelIterators.submitWork
This is an automated email from the ASF dual-hosted git repository. chenglei pushed a commit to branch 4.x in repository https://gitbox.apache.org/repos/asf/phoenix.git The following commit(s) were added to refs/heads/4.x by this push: new 4ed30f2 PHOENIX-6287 Fix incorrect log in ParallelIterators.submitWork 4ed30f2 is described below commit 4ed30f23c74b0409c9bd4ebe0997de95ac4fc898 Author: chenglei AuthorDate: Wed Dec 30 17:43:38 2020 +0800 PHOENIX-6287 Fix incorrect log in ParallelIterators.submitWork --- .../org/apache/phoenix/iterate/ParallelIterators.java | 19 +-- 1 file changed, 13 insertions(+), 6 deletions(-) diff --git a/phoenix-core/src/main/java/org/apache/phoenix/iterate/ParallelIterators.java b/phoenix-core/src/main/java/org/apache/phoenix/iterate/ParallelIterators.java index 7b1229b..88b48c0 100644 --- a/phoenix-core/src/main/java/org/apache/phoenix/iterate/ParallelIterators.java +++ b/phoenix-core/src/main/java/org/apache/phoenix/iterate/ParallelIterators.java @@ -123,12 +123,12 @@ public class ParallelIterators extends BaseResultIterators { @Override public PeekingResultIterator call() throws Exception { long startTime = EnvironmentEdgeManager.currentTimeMillis(); -if (LOGGER.isDebugEnabled()) { -LOGGER.debug(LogUtil.addCustomAnnotations("Id: " + scanId + ", Time: " + -(EnvironmentEdgeManager.currentTimeMillis() - startTime) + -"ms, Scan: " + scan, ScanUtil.getCustomAnnotations(scan))); -} -PeekingResultIterator iterator = iteratorFactory.newIterator(context, tableResultItr, scan, physicalTableName, ParallelIterators.this.plan); +PeekingResultIterator iterator = iteratorFactory.newIterator( +context, +tableResultItr, +scan, +physicalTableName, +ParallelIterators.this.plan); if (initFirstScanOnly) { if ((!isReverse && scanLocation.isFirstScan()) || (isReverse && scanLocation.isLastScan())) { // Fill the scanner's cache. This helps reduce latency since we are parallelizing the I/O needed. @@ -137,6 +137,13 @@ public class ParallelIterators extends BaseResultIterators { } else { iterator.peek(); } + +if (LOGGER.isDebugEnabled()) { +LOGGER.debug(LogUtil.addCustomAnnotations("Id: " + scanId + ", Time: " + +(EnvironmentEdgeManager.currentTimeMillis() - startTime) + +"ms, Scan: " + scan, ScanUtil.getCustomAnnotations(scan))); +} + allIterators.add(iterator); return iterator; }
Apache-Phoenix | master | HBase 2.2 | Build #168 SUCCESS
master branch HBase 2.2 build #168 status SUCCESS Build #168 https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-mulitbranch/job/master/168/
Apache-Phoenix | master | HBase 2.1 | Build #168 SUCCESS
master branch HBase 2.1 build #168 status SUCCESS Build #168 https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-mulitbranch/job/master/168/
Apache-Phoenix | master | HBase 2.3 | Build #168 FAILURE
master branch HBase 2.3 build #168 status FAILURE Build #168 https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-mulitbranch/job/master/168/