[jira] [Commented] (HBASE-13876) Improving performance of HeapMemoryManager
[ https://issues.apache.org/jira/browse/HBASE-13876?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14581710#comment-14581710 ] Hadoop QA commented on HBASE-13876: --- {color:red}-1 overall{color}. Here are the results of testing the latest attachment http://issues.apache.org/jira/secure/attachment/12738995/HBASE-13876-v5.patch against master branch at commit 349cbe102a130b50852201e84dc7ac3bea4fc1f5. ATTACHMENT ID: 12738995 {color:green}+1 @author{color}. The patch does not contain any @author tags. {color:green}+1 tests included{color}. The patch appears to include 3 new or modified tests. {color:green}+1 hadoop versions{color}. The patch compiles with all supported hadoop versions (2.4.1 2.5.2 2.6.0) {color:green}+1 javac{color}. The applied patch does not increase the total number of javac compiler warnings. {color:green}+1 protoc{color}. The applied patch does not increase the total number of protoc compiler warnings. {color:green}+1 javadoc{color}. The javadoc tool did not generate any warning messages. {color:green}+1 checkstyle{color}. The applied patch does not increase the total number of checkstyle errors {color:green}+1 findbugs{color}. The patch does not introduce any new Findbugs (version 2.0.3) warnings. {color:green}+1 release audit{color}. The applied patch does not increase the total number of release audit warnings. {color:green}+1 lineLengths{color}. The patch does not introduce lines longer than 100 {color:green}+1 site{color}. The mvn site goal succeeds with this patch. {color:red}-1 core tests{color}. The patch failed these unit tests: org.apache.hadoop.hbase.TestRegionRebalancing Test results: https://builds.apache.org/job/PreCommit-HBASE-Build/14378//testReport/ Release Findbugs (version 2.0.3)warnings: https://builds.apache.org/job/PreCommit-HBASE-Build/14378//artifact/patchprocess/newFindbugsWarnings.html Checkstyle Errors: https://builds.apache.org/job/PreCommit-HBASE-Build/14378//artifact/patchprocess/checkstyle-aggregate.html Console output: https://builds.apache.org/job/PreCommit-HBASE-Build/14378//console This message is automatically generated. Improving performance of HeapMemoryManager -- Key: HBASE-13876 URL: https://issues.apache.org/jira/browse/HBASE-13876 Project: HBase Issue Type: Improvement Components: hbase, regionserver Affects Versions: 2.0.0, 1.0.1, 1.1.0, 1.1.1 Reporter: Abhilash Assignee: Abhilash Priority: Minor Attachments: HBASE-13876-v2.patch, HBASE-13876-v3.patch, HBASE-13876-v4.patch, HBASE-13876-v5.patch, HBASE-13876-v5.patch, HBASE-13876.patch I am trying to improve the performance of DefaultHeapMemoryTuner by introducing some more checks. The current checks under which the DefaultHeapMemoryTuner works are very rare so I am trying to weaken these checks to improve its performance. Check current memstore size and current block cache size. For say if we are using less than 50% of currently available block cache size we say block cache is sufficient and same for memstore. This check will be very effective when server is either load heavy or write heavy. Earlier version just waited for number of evictions / number of flushes to be zero which are very rare. Otherwise based on percent change in number of cache misses and number of flushes we increase / decrease memory provided for caching / memstore. After doing so, on next call of HeapMemoryTuner we verify that last change has indeed decreased number of evictions / flush either of which it was expected to do. We also check that it does not make the other (evictions / flush) increase much. I am doing this analysis by comparing percent change (which is basically nothing but normalized derivative) of number of evictions and number of flushes during last two periods. The main motive for doing this was that if we have random reads then we will be having a lot of cache misses. But even after increasing block cache we wont be able to decrease number of cache misses and we will revert back and eventually we will not waste memory on block caches. This will also help us ignore random short term spikes in reads / writes. I have also tried to take care not to tune memory if do do not have enough hints as unnecessary tuning my slow down the system. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Resolved] (HBASE-13855) Race in multi threaded PartitionedMobCompactor causes NPE
[ https://issues.apache.org/jira/browse/HBASE-13855?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Anoop Sam John resolved HBASE-13855. Resolution: Fixed Hadoop Flags: Reviewed Thanks Jingcheng. Race in multi threaded PartitionedMobCompactor causes NPE - Key: HBASE-13855 URL: https://issues.apache.org/jira/browse/HBASE-13855 Project: HBase Issue Type: Sub-task Components: mob Affects Versions: hbase-11339 Reporter: Jingcheng Du Assignee: Jingcheng Du Priority: Critical Fix For: hbase-11339 Attachments: HBASE-13855.diff In PartitionedMobCompactor, mob files are split into partitions, the compactions of partitions run in parallel. The partitions share the same set of del files. There might be race conditions when open readers of del store files in each partition which can cause NPE. In this patch, we will pre-create the reader for each del store file to avoid this issue. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HBASE-13828) Add group permissions testing coverage to AC.
[ https://issues.apache.org/jira/browse/HBASE-13828?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14581703#comment-14581703 ] Hudson commented on HBASE-13828: SUCCESS: Integrated in HBase-0.98 #1025 (See [https://builds.apache.org/job/HBase-0.98/1025/]) HBASE-13828 Add group permissions testing coverage to AC (apurtell: rev 95cc075a8acbe6313852c37836b8ed145cfdbb33) * hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestNamespaceCommands.java * hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestAccessController2.java * hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestAccessController.java Add group permissions testing coverage to AC. - Key: HBASE-13828 URL: https://issues.apache.org/jira/browse/HBASE-13828 Project: HBase Issue Type: Improvement Reporter: Srikanth Srungarapu Assignee: Ashish Singhi Fix For: 2.0.0, 0.98.13, 1.0.2, 1.2.0, 1.1.1 Attachments: HBASE-13828-0.98-v2.patch, HBASE-13828-branch-1-v2.patch, HBASE-13828-branch-1.0-v2.patch, HBASE-13828-branch-1.1-v2.patch, HBASE-13828-v1.patch, HBASE-13828-v2.patch, HBASE-13828-v3.patch, HBASE-13828.patch We suffered a regression HBASE-13826 recently due to lack of testing coverage for group permissions for AC. With the recent perf boost provided by HBASE-13658, it wouldn't be a bad idea to add checks for group level users to applicable unit tests in TestAccessController. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HBASE-13855) Race in multi threaded PartitionedMobCompactor causes NPE
[ https://issues.apache.org/jira/browse/HBASE-13855?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Anoop Sam John updated HBASE-13855: --- Summary: Race in multi threaded PartitionedMobCompactor causes NPE (was: NPE in PartitionedMobCompactor) Race in multi threaded PartitionedMobCompactor causes NPE - Key: HBASE-13855 URL: https://issues.apache.org/jira/browse/HBASE-13855 Project: HBase Issue Type: Sub-task Components: mob Affects Versions: hbase-11339 Reporter: Jingcheng Du Assignee: Jingcheng Du Priority: Critical Fix For: hbase-11339 Attachments: HBASE-13855.diff In PartitionedMobCompactor, mob files are split into partitions, the compactions of partitions run in parallel. The partitions share the same set of del files. There might be race conditions when open readers of del store files in each partition which can cause NPE. In this patch, we will pre-create the reader for each del store file to avoid this issue. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Resolved] (HBASE-12713) IncreasingToUpperBoundRegionSplitPolicy invalid when region count greater than 100
[ https://issues.apache.org/jira/browse/HBASE-12713?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Liu Junhong resolved HBASE-12713. - Resolution: Won't Fix IncreasingToUpperBoundRegionSplitPolicy invalid when region count greater than 100 -- Key: HBASE-12713 URL: https://issues.apache.org/jira/browse/HBASE-12713 Project: HBase Issue Type: Improvement Components: regionserver Affects Versions: 0.98.6 Reporter: Liu Junhong Original Estimate: 4m Remaining Estimate: 4m I find that the policy of region split which in IncreasingToUpperBoundRegionSplitPolicy will be use the value of maxfilesize when the count of region is greater than 100. But sometimes 100 regions is not too much for a cluster that has 50 or more regionservers. So i think this policy should consider the density of the regions but not the total count of the regions. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HBASE-13828) Add group permissions testing coverage to AC.
[ https://issues.apache.org/jira/browse/HBASE-13828?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14581782#comment-14581782 ] Hudson commented on HBASE-13828: SUCCESS: Integrated in HBase-0.98-on-Hadoop-1.1 #978 (See [https://builds.apache.org/job/HBase-0.98-on-Hadoop-1.1/978/]) HBASE-13828 Add group permissions testing coverage to AC (apurtell: rev 95cc075a8acbe6313852c37836b8ed145cfdbb33) * hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestAccessController.java * hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestNamespaceCommands.java * hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestAccessController2.java Add group permissions testing coverage to AC. - Key: HBASE-13828 URL: https://issues.apache.org/jira/browse/HBASE-13828 Project: HBase Issue Type: Improvement Reporter: Srikanth Srungarapu Assignee: Ashish Singhi Fix For: 2.0.0, 0.98.13, 1.0.2, 1.2.0, 1.1.1 Attachments: HBASE-13828-0.98-v2.patch, HBASE-13828-branch-1-v2.patch, HBASE-13828-branch-1.0-v2.patch, HBASE-13828-branch-1.1-v2.patch, HBASE-13828-v1.patch, HBASE-13828-v2.patch, HBASE-13828-v3.patch, HBASE-13828.patch We suffered a regression HBASE-13826 recently due to lack of testing coverage for group permissions for AC. With the recent perf boost provided by HBASE-13658, it wouldn't be a bad idea to add checks for group level users to applicable unit tests in TestAccessController. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HBASE-13888) refill bug from HBASE-13686
[ https://issues.apache.org/jira/browse/HBASE-13888?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14581954#comment-14581954 ] Ted Yu commented on HBASE-13888: Any chance of a unit test that shows the problem ? Thanks refill bug from HBASE-13686 --- Key: HBASE-13888 URL: https://issues.apache.org/jira/browse/HBASE-13888 Project: HBase Issue Type: Bug Affects Versions: 2.0.0 Reporter: Guanghao Zhang Assignee: Guanghao Zhang Attachments: HBASE-13888-v1.patch As I report the RateLimiter fail to limit in HBASE-13686, then [~ashish singhi] fix that problem by support two kinds of RateLimiter: AverageIntervalRateLimiter and FixedIntervalRateLimiter. But in my use of the code, I found a new bug about refill() in AverageIntervalRateLimiter. {code} long delta = (limit * (now - nextRefillTime)) / super.getTimeUnitInMillis(); if (delta 0) { this.nextRefillTime = now; return Math.min(limit, available + delta); } {code} When delta 0, refill maybe return available + delta. Then in the canExecute(), avail will add refillAmount again. So the new avail maybe 2 * avail + delta. {code} long refillAmount = refill(limit, avail); if (refillAmount == 0 avail amount) { return false; } // check for positive overflow if (avail = Long.MAX_VALUE - refillAmount) { avail = Math.max(0, Math.min(avail + refillAmount, limit)); } else { avail = Math.max(0, limit); } {code} I will add more unit tests for RateLimiter in the next days. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HBASE-13881) Bugs in HTable incrementColumnValue implementation
[ https://issues.apache.org/jira/browse/HBASE-13881?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14582038#comment-14582038 ] Jerry Lam commented on HBASE-13881: --- No I don't but I can create one. I believe the semantic of writeToWAL = true is equal to Durability.SYNC_WAL? Please confirm. Thank you. Bugs in HTable incrementColumnValue implementation -- Key: HBASE-13881 URL: https://issues.apache.org/jira/browse/HBASE-13881 Project: HBase Issue Type: Bug Components: Client Affects Versions: 0.98.6.1, 1.0.1 Reporter: Jerry Lam The exact method I'm talking about is: {code} @Deprecated @Override public long incrementColumnValue(final byte [] row, final byte [] family, final byte [] qualifier, final long amount, final boolean writeToWAL) throws IOException { return incrementColumnValue(row, family, qualifier, amount, writeToWAL? Durability.SKIP_WAL: Durability.USE_DEFAULT); } {code} Setting writeToWAL to true, Durability will be set to SKIP_WAL which does not make much sense unless the meaning of SKIP_WAL is negated. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HBASE-13560) Large compaction queue should steal from small compaction queue when idle
[ https://issues.apache.org/jira/browse/HBASE-13560?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14582193#comment-14582193 ] Elliott Clark commented on HBASE-13560: --- bq.Should there be a config which disables this feature ? Please no more configs. Large compaction queue should steal from small compaction queue when idle - Key: HBASE-13560 URL: https://issues.apache.org/jira/browse/HBASE-13560 Project: HBase Issue Type: Bug Components: Compaction Affects Versions: 2.0.0, 1.2.0 Reporter: Elliott Clark Assignee: Changgeng Li Attachments: HBASE-13560.patch, queuestealwork-v1.patch, queuestealwork-v4.patch, queuestealwork-v5.patch, queuestealwork-v6.patch, queuestealwork-v7.patch If you tune compaction threads so that a server is never over commited when large and small compaction threads are busy then it should be possible to have the large compactions steal work. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HBASE-13887) Document 0.98 release build differences
[ https://issues.apache.org/jira/browse/HBASE-13887?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=1458#comment-1458 ] Nick Dimiduk commented on HBASE-13887: -- bq. This JDK suffers from JDK-6521495 Ouch. Document 0.98 release build differences --- Key: HBASE-13887 URL: https://issues.apache.org/jira/browse/HBASE-13887 Project: HBase Issue Type: Bug Components: documentation, site Reporter: Andrew Purtell Fix For: 2.0.0 The build instructions in the online manual do not describe the extra steps we need to take to build 0.98. Add a section on this. A quick enumeration of the differences: 1. Source assemblies will be missing the hbase-hadoop1-compat module. This should be fixed in the POM somehow. What I do now is untar the src tarball, cp -a the module over, then tar up the result. (It's a hack in a release script.) 2. We must munge POMs for building hadoop1 and hadoop2 variants and then execute two builds pointing Maven at each munged POM. The generate-hadoop-X-poms script requires bash {noformat} $ bash dev-support/generate-hadoopX-poms.sh $version $version-hadoop1 $ bash dev-support/generate-hadoopX-poms.sh $version $version-hadoop2 {noformat} Build Hadoop 1 {noformat} $ mvn -f pom.xml.hadoop1 clean install -DskipTests -Prelease \ mvn -f pom.xml.hadoop1 install -DskipTests site assembly:single \ -Prelease \ mvn -f pom.xml.hadoop1 deploy -DskipTests -Papache-release $ cp hbase-assembly/target/hbase*-bin.tar.gz $release_dir {noformat} Build Hadoop 2 {noformat} $ mvn -f pom.xml.hadoop2 clean install -DskipTests -Prelease \ mvn -f pom.xml.hadoop2 install -DskipTests site assembly:single \ -Prelease \ mvn -f pom.xml.hadoop2 deploy -DskipTests -Papache-release $ cp hbase-assembly/target/hbase*-bin.tar.gz $release_dir {noformat} 3. Current HEAD of 0.98 branch enforces a requirement that the release be built with a JDK no more recent than the compile language level. For 0.98, that is 1.6, therefore the ancient 6u45 JDK. This JDK suffers from [JDK-6521495|http://bugs.java.com/bugdatabase/view_bug.do?bug_id=6521495] so the following workaround is required in order to deploy artifacts to Apache's Nexus: 3.a. Download https://www.bouncycastle.org/download/bcprov-jdk15on-152.jar and https://www.bouncycastle.org/download/bcprov-ext-jdk15on-152.jar into $JAVA_HOME/lib/ext. 3.b. Edit $JAVA_HOME/lib/security/java.security and add the BouncyCastle provider as the first provider: {noformat} security.provider.1=org.bouncycastle.jce.provider.BouncyCastleProvider {noformat} -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Created] (HBASE-13889) hbase-shaded-client artifact is missing dependency (therefore, does not work)
Dmitry Minkovsky created HBASE-13889: Summary: hbase-shaded-client artifact is missing dependency (therefore, does not work) Key: HBASE-13889 URL: https://issues.apache.org/jira/browse/HBASE-13889 Project: HBase Issue Type: Bug Components: Client Affects Versions: 1.1.0, 1.1.0.1 Environment: N/A? Reporter: Dmitry Minkovsky Priority: Blocker The {{hbase-shaded-client}} artifact was introduced in [HBASE-13517|https://issues.apache.org/jira/browse/HBASE-13517]. Thank you very much for this, as I am new to Java building and was having a very slow-moving time resolving conflicts. However, the shaded client artifact seems to be missing {{javax.xml.transform.TransformerException}}. I examined the JAR, which does not have this package/class. Steps to reproduce: Java: {code} package com.mycompany.app; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.client.Connection; import org.apache.hadoop.hbase.client.ConnectionFactory; public class App { public static void main( String[] args ) throws java.io.IOException { Configuration config = HBaseConfiguration.create(); Connection connection = ConnectionFactory.createConnection(config); } } {code} POM: {code} project xmlns=http://maven.apache.org/POM/4.0.0; xmlns:xsi=http://www.w3.org/2001/XMLSchema-instance; xsi:schemaLocation=http://maven.apache.org/POM/4.0.0 http://maven.apache.org/xsd/maven-4.0.0.xsd; modelVersion4.0.0/modelVersion groupIdcom.mycompany.app/groupId artifactIdmy-app/artifactId version1.0-SNAPSHOT/version packagingjar/packaging namemy-app/name
[jira] [Updated] (HBASE-13889) hbase-shaded-client artifact is missing dependency (therefore, does not work)
[ https://issues.apache.org/jira/browse/HBASE-13889?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Nick Dimiduk updated HBASE-13889: - Fix Version/s: 1.1.1 1.2.0 2.0.0 hbase-shaded-client artifact is missing dependency (therefore, does not work) - Key: HBASE-13889 URL: https://issues.apache.org/jira/browse/HBASE-13889 Project: HBase Issue Type: Bug Components: Client Affects Versions: 1.1.0, 1.1.0.1 Environment: N/A? Reporter: Dmitry Minkovsky Priority: Blocker Fix For: 2.0.0, 1.2.0, 1.1.1 The {{hbase-shaded-client}} artifact was introduced in [HBASE-13517|https://issues.apache.org/jira/browse/HBASE-13517]. Thank you very much for this, as I am new to Java building and was having a very slow-moving time resolving conflicts. However, the shaded client artifact seems to be missing {{javax.xml.transform.TransformerException}}. I examined the JAR, which does not have this package/class. Steps to reproduce: Java: {code} package com.mycompany.app; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.client.Connection; import org.apache.hadoop.hbase.client.ConnectionFactory; public class App { public static void main( String[] args ) throws java.io.IOException { Configuration config = HBaseConfiguration.create(); Connection connection = ConnectionFactory.createConnection(config); } } {code} POM: {code} project xmlns=http://maven.apache.org/POM/4.0.0; xmlns:xsi=http://www.w3.org/2001/XMLSchema-instance; xsi:schemaLocation=http://maven.apache.org/POM/4.0.0 http://maven.apache.org/xsd/maven-4.0.0.xsd; modelVersion4.0.0/modelVersion groupIdcom.mycompany.app/groupId artifactIdmy-app/artifactId version1.0-SNAPSHOT/version packagingjar/packaging
[jira] [Commented] (HBASE-13889) hbase-shaded-client artifact is missing dependency (therefore, does not work)
[ https://issues.apache.org/jira/browse/HBASE-13889?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14582271#comment-14582271 ] Nick Dimiduk commented on HBASE-13889: -- This is a JDK class, so should be excluded from the shade. http://docs.oracle.com/javase/7/docs/api/javax/xml/transform/TransformerException.html hbase-shaded-client artifact is missing dependency (therefore, does not work) - Key: HBASE-13889 URL: https://issues.apache.org/jira/browse/HBASE-13889 Project: HBase Issue Type: Bug Components: Client Affects Versions: 1.1.0, 1.1.0.1 Environment: N/A? Reporter: Dmitry Minkovsky Priority: Blocker Fix For: 2.0.0, 1.2.0, 1.1.1 The {{hbase-shaded-client}} artifact was introduced in [HBASE-13517|https://issues.apache.org/jira/browse/HBASE-13517]. Thank you very much for this, as I am new to Java building and was having a very slow-moving time resolving conflicts. However, the shaded client artifact seems to be missing {{javax.xml.transform.TransformerException}}. I examined the JAR, which does not have this package/class. Steps to reproduce: Java: {code} package com.mycompany.app; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.client.Connection; import org.apache.hadoop.hbase.client.ConnectionFactory; public class App { public static void main( String[] args ) throws java.io.IOException { Configuration config = HBaseConfiguration.create(); Connection connection = ConnectionFactory.createConnection(config); } } {code} POM: {code} project xmlns=http://maven.apache.org/POM/4.0.0; xmlns:xsi=http://www.w3.org/2001/XMLSchema-instance; xsi:schemaLocation=http://maven.apache.org/POM/4.0.0 http://maven.apache.org/xsd/maven-4.0.0.xsd; modelVersion4.0.0/modelVersion groupIdcom.mycompany.app/groupId artifactIdmy-app/artifactId version1.0-SNAPSHOT/version packagingjar/packaging
[jira] [Commented] (HBASE-13889) hbase-shaded-client artifact is missing dependency (therefore, does not work)
[ https://issues.apache.org/jira/browse/HBASE-13889?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14582272#comment-14582272 ] Dmitry Minkovsky commented on HBASE-13889: -- Oh yes obviously, good call. Excellent opportunity for me to try to fix this myself while maybe [~eclark] or someone else who knows how to fix this quickly can step in. hbase-shaded-client artifact is missing dependency (therefore, does not work) - Key: HBASE-13889 URL: https://issues.apache.org/jira/browse/HBASE-13889 Project: HBase Issue Type: Bug Components: Client Affects Versions: 1.1.0, 1.1.0.1 Environment: N/A? Reporter: Dmitry Minkovsky Priority: Blocker Fix For: 2.0.0, 1.2.0, 1.1.1 The {{hbase-shaded-client}} artifact was introduced in [HBASE-13517|https://issues.apache.org/jira/browse/HBASE-13517]. Thank you very much for this, as I am new to Java building and was having a very slow-moving time resolving conflicts. However, the shaded client artifact seems to be missing {{javax.xml.transform.TransformerException}}. I examined the JAR, which does not have this package/class. Steps to reproduce: Java: {code} package com.mycompany.app; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.client.Connection; import org.apache.hadoop.hbase.client.ConnectionFactory; public class App { public static void main( String[] args ) throws java.io.IOException { Configuration config = HBaseConfiguration.create(); Connection connection = ConnectionFactory.createConnection(config); } } {code} POM: {code} project xmlns=http://maven.apache.org/POM/4.0.0; xmlns:xsi=http://www.w3.org/2001/XMLSchema-instance; xsi:schemaLocation=http://maven.apache.org/POM/4.0.0 http://maven.apache.org/xsd/maven-4.0.0.xsd; modelVersion4.0.0/modelVersion groupIdcom.mycompany.app/groupId artifactIdmy-app/artifactId version1.0-SNAPSHOT/version
[jira] [Commented] (HBASE-13889) hbase-shaded-client artifact is missing dependency (therefore, does not work)
[ https://issues.apache.org/jira/browse/HBASE-13889?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14582286#comment-14582286 ] Elliott Clark commented on HBASE-13889: --- H it looks like the shaded jar doesn't include any shaded classes. hbase-shaded-client artifact is missing dependency (therefore, does not work) - Key: HBASE-13889 URL: https://issues.apache.org/jira/browse/HBASE-13889 Project: HBase Issue Type: Bug Components: Client Affects Versions: 1.1.0, 1.1.0.1 Environment: N/A? Reporter: Dmitry Minkovsky Priority: Blocker Fix For: 2.0.0, 1.2.0, 1.1.1 The {{hbase-shaded-client}} artifact was introduced in [HBASE-13517|https://issues.apache.org/jira/browse/HBASE-13517]. Thank you very much for this, as I am new to Java building and was having a very slow-moving time resolving conflicts. However, the shaded client artifact seems to be missing {{javax.xml.transform.TransformerException}}. I examined the JAR, which does not have this package/class. Steps to reproduce: Java: {code} package com.mycompany.app; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.client.Connection; import org.apache.hadoop.hbase.client.ConnectionFactory; public class App { public static void main( String[] args ) throws java.io.IOException { Configuration config = HBaseConfiguration.create(); Connection connection = ConnectionFactory.createConnection(config); } } {code} POM: {code} project xmlns=http://maven.apache.org/POM/4.0.0; xmlns:xsi=http://www.w3.org/2001/XMLSchema-instance; xsi:schemaLocation=http://maven.apache.org/POM/4.0.0 http://maven.apache.org/xsd/maven-4.0.0.xsd; modelVersion4.0.0/modelVersion groupIdcom.mycompany.app/groupId artifactIdmy-app/artifactId version1.0-SNAPSHOT/version packagingjar/packaging
[jira] [Updated] (HBASE-13889) hbase-shaded-client artifact is missing dependency (therefore, does not work)
[ https://issues.apache.org/jira/browse/HBASE-13889?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Elliott Clark updated HBASE-13889: -- Attachment: Screen Shot 2015-06-11 at 10.59.55 AM.png hbase-shaded-client artifact is missing dependency (therefore, does not work) - Key: HBASE-13889 URL: https://issues.apache.org/jira/browse/HBASE-13889 Project: HBase Issue Type: Bug Components: Client Affects Versions: 1.1.0, 1.1.0.1 Environment: N/A? Reporter: Dmitry Minkovsky Priority: Blocker Fix For: 2.0.0, 1.2.0, 1.1.1 Attachments: Screen Shot 2015-06-11 at 10.59.55 AM.png The {{hbase-shaded-client}} artifact was introduced in [HBASE-13517|https://issues.apache.org/jira/browse/HBASE-13517]. Thank you very much for this, as I am new to Java building and was having a very slow-moving time resolving conflicts. However, the shaded client artifact seems to be missing {{javax.xml.transform.TransformerException}}. I examined the JAR, which does not have this package/class. Steps to reproduce: Java: {code} package com.mycompany.app; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.client.Connection; import org.apache.hadoop.hbase.client.ConnectionFactory; public class App { public static void main( String[] args ) throws java.io.IOException { Configuration config = HBaseConfiguration.create(); Connection connection = ConnectionFactory.createConnection(config); } } {code} POM: {code} project xmlns=http://maven.apache.org/POM/4.0.0; xmlns:xsi=http://www.w3.org/2001/XMLSchema-instance; xsi:schemaLocation=http://maven.apache.org/POM/4.0.0 http://maven.apache.org/xsd/maven-4.0.0.xsd; modelVersion4.0.0/modelVersion groupIdcom.mycompany.app/groupId artifactIdmy-app/artifactId version1.0-SNAPSHOT/version packagingjar/packaging
[jira] [Commented] (HBASE-13877) Interrupt to flush from TableFlushProcedure causes dataloss in ITBLL
[ https://issues.apache.org/jira/browse/HBASE-13877?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14582298#comment-14582298 ] Enis Soztutar commented on HBASE-13877: --- bq. Enis Soztutar any comment on Duo Zhang remark? Let me see how I can simplify the contract. bq. Are dealing w/ above, I'm +1 on commit. I did not find incidence of the original issue in my run after looking in all logs. In my case, I am seeing double-assignment over a master restart. Agreed, lets commit this and continue on yet another issue. Interrupt to flush from TableFlushProcedure causes dataloss in ITBLL Key: HBASE-13877 URL: https://issues.apache.org/jira/browse/HBASE-13877 Project: HBase Issue Type: Bug Reporter: Enis Soztutar Assignee: Enis Soztutar Priority: Blocker Fix For: 2.0.0, 1.2.0, 1.1.1 Attachments: hbase-13877_v1.patch, hbase-13877_v2-branch-1.1.patch ITBLL with 1.25B rows failed for me (and Stack as reported in https://issues.apache.org/jira/browse/HBASE-13811?focusedCommentId=14577834page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-14577834) HBASE-13811 and HBASE-13853 fixed an issue with WAL edit filtering. The root cause this time seems to be different. It is due to procedure based flush interrupting the flush request in case the procedure is cancelled from an exception elsewhere. This leaves the memstore snapshot intact without aborting the server. The next flush, then flushes the previous memstore with the current seqId (as opposed to seqId from the memstore snapshot). This creates an hfile with larger seqId than what its contents are. Previous behavior in 0.98 and 1.0 (I believe) is that after flush prepare and interruption / exception will cause RS abort. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HBASE-13889) hbase-shaded-client artifact is missing dependency (therefore, does not work)
[ https://issues.apache.org/jira/browse/HBASE-13889?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14582305#comment-14582305 ] Elliott Clark commented on HBASE-13889: --- Blah nevermind. Yeah it includes some but not all of the javax stuff. hbase-shaded-client artifact is missing dependency (therefore, does not work) - Key: HBASE-13889 URL: https://issues.apache.org/jira/browse/HBASE-13889 Project: HBase Issue Type: Bug Components: Client Affects Versions: 1.1.0, 1.1.0.1 Environment: N/A? Reporter: Dmitry Minkovsky Priority: Blocker Fix For: 2.0.0, 1.2.0, 1.1.1 Attachments: Screen Shot 2015-06-11 at 10.59.55 AM.png The {{hbase-shaded-client}} artifact was introduced in [HBASE-13517|https://issues.apache.org/jira/browse/HBASE-13517]. Thank you very much for this, as I am new to Java building and was having a very slow-moving time resolving conflicts. However, the shaded client artifact seems to be missing {{javax.xml.transform.TransformerException}}. I examined the JAR, which does not have this package/class. Steps to reproduce: Java: {code} package com.mycompany.app; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.client.Connection; import org.apache.hadoop.hbase.client.ConnectionFactory; public class App { public static void main( String[] args ) throws java.io.IOException { Configuration config = HBaseConfiguration.create(); Connection connection = ConnectionFactory.createConnection(config); } } {code} POM: {code} project xmlns=http://maven.apache.org/POM/4.0.0; xmlns:xsi=http://www.w3.org/2001/XMLSchema-instance; xsi:schemaLocation=http://maven.apache.org/POM/4.0.0 http://maven.apache.org/xsd/maven-4.0.0.xsd; modelVersion4.0.0/modelVersion groupIdcom.mycompany.app/groupId artifactIdmy-app/artifactId version1.0-SNAPSHOT/version packagingjar/packaging
[jira] [Updated] (HBASE-13833) LoadIncrementalHFile.doBulkLoad(Path,HTable) doesn't handle unmanaged connections when using SecureBulkLoad
[ https://issues.apache.org/jira/browse/HBASE-13833?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Nick Dimiduk updated HBASE-13833: - Attachment: HBASE-13833.02.branch-1.patch LoadIncrementalHFile.doBulkLoad(Path,HTable) doesn't handle unmanaged connections when using SecureBulkLoad --- Key: HBASE-13833 URL: https://issues.apache.org/jira/browse/HBASE-13833 Project: HBase Issue Type: Bug Affects Versions: 1.1.0.1 Reporter: Nick Dimiduk Assignee: Nick Dimiduk Fix For: 1.1.1 Attachments: HBASE-13833.00.branch-1.1.patch, HBASE-13833.01.branch-1.1.patch, HBASE-13833.02.branch-1.1.patch, HBASE-13833.02.branch-1.patch Seems HBASE-13328 wasn't quite sufficient. {noformat} 015-06-02 05:49:23,578|beaver.machine|INFO|2828|7140|MainThread|15/06/02 05:49:23 WARN mapreduce.LoadIncrementalHFiles: Skipping non-directory hdfs://dal-pqc1:8020/tmp/192f21dd-cc89-4354-8ba1-78d1f228e7c7/LARGE_TABLE/_SUCCESS 2015-06-02 05:49:23,720|beaver.machine|INFO|2828|7140|MainThread|15/06/02 05:49:23 INFO hfile.CacheConfig: CacheConfig:disabled 2015-06-02 05:49:23,859|beaver.machine|INFO|2828|7140|MainThread|15/06/02 05:49:23 INFO mapreduce.LoadIncrementalHFiles: Trying to load hfile=hdfs://dal-pqc1:8020/tmp/192f21dd-cc89-4354-8ba1-78d1f228e7c7/LARGE_TABLE/0/00870fd0a7544373b32b6f1e976bf47f first=\x80\x00\x00\x00 last=\x80LK? 2015-06-02 05:50:32,028|beaver.machine|INFO|2828|7140|MainThread|15/06/02 05:50:32 INFO client.RpcRetryingCaller: Call exception, tries=10, retries=35, started=68154 ms ago, cancelled=false, msg=row '' on table 'LARGE_TABLE' at region=LARGE_TABLE,,1433222865285.e01e02483f30a060d3f7abb1846ea029., hostname=dal-pqc5,16020,1433222547221, seqNum=2 2015-06-02 05:50:52,128|beaver.machine|INFO|2828|7140|MainThread|15/06/02 05:50:52 INFO client.RpcRetryingCaller: Call exception, tries=11, retries=35, started=88255 ms ago, cancelled=false, msg=row '' on table 'LARGE_TABLE' at region=LARGE_TABLE,,1433222865285.e01e02483f30a060d3f7abb1846ea029., hostname=dal-pqc5,16020,1433222547221, seqNum=2 ... ... 2015-06-02 05:01:56,121|beaver.machine|INFO|7800|2276|MainThread|15/06/02 05:01:56 ERROR mapreduce.CsvBulkLoadTool: Import job on table=LARGE_TABLE failed due to exception. 2015-06-02 05:01:56,121|beaver.machine|INFO|7800|2276|MainThread|java.io.IOException: BulkLoad encountered an unrecoverable problem 2015-06-02 05:01:56,121|beaver.machine|INFO|7800|2276|MainThread|at org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.bulkLoadPhase(LoadIncrementalHFiles.java:474) 2015-06-02 05:01:56,122|beaver.machine|INFO|7800|2276|MainThread|at org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.doBulkLoad(LoadIncrementalHFiles.java:405) 2015-06-02 05:01:56,122|beaver.machine|INFO|7800|2276|MainThread|at org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.doBulkLoad(LoadIncrementalHFiles.java:300) 2015-06-02 05:01:56,122|beaver.machine|INFO|7800|2276|MainThread|at org.apache.phoenix.mapreduce.CsvBulkLoadTool$TableLoader.call(CsvBulkLoadTool.java:517) 2015-06-02 05:01:56,122|beaver.machine|INFO|7800|2276|MainThread|at org.apache.phoenix.mapreduce.CsvBulkLoadTool$TableLoader.call(CsvBulkLoadTool.java:466) 2015-06-02 05:01:56,122|beaver.machine|INFO|7800|2276|MainThread|at java.util.concurrent.FutureTask.run(FutureTask.java:266) 2015-06-02 05:01:56,122|beaver.machine|INFO|7800|2276|MainThread|at org.apache.phoenix.job.JobManager$InstrumentedJobFutureTask.run(JobManager.java:172) 2015-06-02 05:01:56,122|beaver.machine|INFO|7800|2276|MainThread|at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) 2015-06-02 05:01:56,124|beaver.machine|INFO|7800|2276|MainThread|at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) 2015-06-02 05:01:56,124|beaver.machine|INFO|7800|2276|MainThread|at java.lang.Thread.run(Thread.java:745) ... ... ... 2015-06-02 05:58:34,993|beaver.machine|INFO|2828|7140|MainThread|Caused by: org.apache.hadoop.hbase.client.NeedUnmanagedConnectionException: The connection has to be unmanaged. 2015-06-02 05:58:34,993|beaver.machine|INFO|2828|7140|MainThread|at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getTable(ConnectionManager.java:724) 2015-06-02 05:58:34,994|beaver.machine|INFO|2828|7140|MainThread|at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getTable(ConnectionManager.java:708) 2015-06-02 05:58:34,994|beaver.machine|INFO|2828|7140|MainThread|at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getTable(ConnectionManager.java:542) 2015-06-02 05:58:34,994|beaver.machine|INFO|2828|7140|MainThread|at
[jira] [Updated] (HBASE-13833) LoadIncrementalHFile.doBulkLoad(Path,HTable) doesn't handle unmanaged connections when using SecureBulkLoad
[ https://issues.apache.org/jira/browse/HBASE-13833?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Nick Dimiduk updated HBASE-13833: - Attachment: HBASE-13833.02.branch-1.0.patch Here's a patch for branch-1.0. The original does not apply cleanly, so maybe [~enis] wants to take a look? Test\{Secure,\}LoadIncrementalHFiles passes locally for me. LoadIncrementalHFile.doBulkLoad(Path,HTable) doesn't handle unmanaged connections when using SecureBulkLoad --- Key: HBASE-13833 URL: https://issues.apache.org/jira/browse/HBASE-13833 Project: HBase Issue Type: Bug Affects Versions: 1.1.0.1 Reporter: Nick Dimiduk Assignee: Nick Dimiduk Fix For: 1.1.1 Attachments: HBASE-13833.00.branch-1.1.patch, HBASE-13833.01.branch-1.1.patch, HBASE-13833.02.branch-1.0.patch, HBASE-13833.02.branch-1.1.patch, HBASE-13833.02.branch-1.patch Seems HBASE-13328 wasn't quite sufficient. {noformat} 015-06-02 05:49:23,578|beaver.machine|INFO|2828|7140|MainThread|15/06/02 05:49:23 WARN mapreduce.LoadIncrementalHFiles: Skipping non-directory hdfs://dal-pqc1:8020/tmp/192f21dd-cc89-4354-8ba1-78d1f228e7c7/LARGE_TABLE/_SUCCESS 2015-06-02 05:49:23,720|beaver.machine|INFO|2828|7140|MainThread|15/06/02 05:49:23 INFO hfile.CacheConfig: CacheConfig:disabled 2015-06-02 05:49:23,859|beaver.machine|INFO|2828|7140|MainThread|15/06/02 05:49:23 INFO mapreduce.LoadIncrementalHFiles: Trying to load hfile=hdfs://dal-pqc1:8020/tmp/192f21dd-cc89-4354-8ba1-78d1f228e7c7/LARGE_TABLE/0/00870fd0a7544373b32b6f1e976bf47f first=\x80\x00\x00\x00 last=\x80LK? 2015-06-02 05:50:32,028|beaver.machine|INFO|2828|7140|MainThread|15/06/02 05:50:32 INFO client.RpcRetryingCaller: Call exception, tries=10, retries=35, started=68154 ms ago, cancelled=false, msg=row '' on table 'LARGE_TABLE' at region=LARGE_TABLE,,1433222865285.e01e02483f30a060d3f7abb1846ea029., hostname=dal-pqc5,16020,1433222547221, seqNum=2 2015-06-02 05:50:52,128|beaver.machine|INFO|2828|7140|MainThread|15/06/02 05:50:52 INFO client.RpcRetryingCaller: Call exception, tries=11, retries=35, started=88255 ms ago, cancelled=false, msg=row '' on table 'LARGE_TABLE' at region=LARGE_TABLE,,1433222865285.e01e02483f30a060d3f7abb1846ea029., hostname=dal-pqc5,16020,1433222547221, seqNum=2 ... ... 2015-06-02 05:01:56,121|beaver.machine|INFO|7800|2276|MainThread|15/06/02 05:01:56 ERROR mapreduce.CsvBulkLoadTool: Import job on table=LARGE_TABLE failed due to exception. 2015-06-02 05:01:56,121|beaver.machine|INFO|7800|2276|MainThread|java.io.IOException: BulkLoad encountered an unrecoverable problem 2015-06-02 05:01:56,121|beaver.machine|INFO|7800|2276|MainThread|at org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.bulkLoadPhase(LoadIncrementalHFiles.java:474) 2015-06-02 05:01:56,122|beaver.machine|INFO|7800|2276|MainThread|at org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.doBulkLoad(LoadIncrementalHFiles.java:405) 2015-06-02 05:01:56,122|beaver.machine|INFO|7800|2276|MainThread|at org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.doBulkLoad(LoadIncrementalHFiles.java:300) 2015-06-02 05:01:56,122|beaver.machine|INFO|7800|2276|MainThread|at org.apache.phoenix.mapreduce.CsvBulkLoadTool$TableLoader.call(CsvBulkLoadTool.java:517) 2015-06-02 05:01:56,122|beaver.machine|INFO|7800|2276|MainThread|at org.apache.phoenix.mapreduce.CsvBulkLoadTool$TableLoader.call(CsvBulkLoadTool.java:466) 2015-06-02 05:01:56,122|beaver.machine|INFO|7800|2276|MainThread|at java.util.concurrent.FutureTask.run(FutureTask.java:266) 2015-06-02 05:01:56,122|beaver.machine|INFO|7800|2276|MainThread|at org.apache.phoenix.job.JobManager$InstrumentedJobFutureTask.run(JobManager.java:172) 2015-06-02 05:01:56,122|beaver.machine|INFO|7800|2276|MainThread|at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) 2015-06-02 05:01:56,124|beaver.machine|INFO|7800|2276|MainThread|at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) 2015-06-02 05:01:56,124|beaver.machine|INFO|7800|2276|MainThread|at java.lang.Thread.run(Thread.java:745) ... ... ... 2015-06-02 05:58:34,993|beaver.machine|INFO|2828|7140|MainThread|Caused by: org.apache.hadoop.hbase.client.NeedUnmanagedConnectionException: The connection has to be unmanaged. 2015-06-02 05:58:34,993|beaver.machine|INFO|2828|7140|MainThread|at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getTable(ConnectionManager.java:724) 2015-06-02 05:58:34,994|beaver.machine|INFO|2828|7140|MainThread|at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getTable(ConnectionManager.java:708) 2015-06-02
[jira] [Created] (HBASE-13888) refill bug from HBASE-13686
Guanghao Zhang created HBASE-13888: -- Summary: refill bug from HBASE-13686 Key: HBASE-13888 URL: https://issues.apache.org/jira/browse/HBASE-13888 Project: HBase Issue Type: Bug Affects Versions: 2.0.0 Reporter: Guanghao Zhang Assignee: Guanghao Zhang As I report the RateLimiter fail to limit in HBASE-13686, then [~ashish singhi] fix that problem by support two kinds of RateLimiter: AverageIntervalRateLimiter and FixedIntervalRateLimiter. But in my use of the code, I found a new bug about refill() in AverageIntervalRateLimiter. {code} long delta = (limit * (now - nextRefillTime)) / super.getTimeUnitInMillis(); if (delta 0) { this.nextRefillTime = now; return Math.min(limit, available + delta); } {code} When delta 0, refill maybe return available + delta. Then in the canExecute(), avail will add refillAmount again. So the new avail maybe 2 * avail + delta. {code} long refillAmount = refill(limit, avail); if (refillAmount == 0 avail amount) { return false; } // check for positive overflow if (avail = Long.MAX_VALUE - refillAmount) { avail = Math.max(0, Math.min(avail + refillAmount, limit)); } else { avail = Math.max(0, limit); } {code} I will add more unit tests for RateLimiter in the next days. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HBASE-13888) refill bug from HBASE-13686
[ https://issues.apache.org/jira/browse/HBASE-13888?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Guanghao Zhang updated HBASE-13888: --- Attachment: HBASE-13888-v1.patch refill bug from HBASE-13686 --- Key: HBASE-13888 URL: https://issues.apache.org/jira/browse/HBASE-13888 Project: HBase Issue Type: Bug Affects Versions: 2.0.0 Reporter: Guanghao Zhang Assignee: Guanghao Zhang Attachments: HBASE-13888-v1.patch As I report the RateLimiter fail to limit in HBASE-13686, then [~ashish singhi] fix that problem by support two kinds of RateLimiter: AverageIntervalRateLimiter and FixedIntervalRateLimiter. But in my use of the code, I found a new bug about refill() in AverageIntervalRateLimiter. {code} long delta = (limit * (now - nextRefillTime)) / super.getTimeUnitInMillis(); if (delta 0) { this.nextRefillTime = now; return Math.min(limit, available + delta); } {code} When delta 0, refill maybe return available + delta. Then in the canExecute(), avail will add refillAmount again. So the new avail maybe 2 * avail + delta. {code} long refillAmount = refill(limit, avail); if (refillAmount == 0 avail amount) { return false; } // check for positive overflow if (avail = Long.MAX_VALUE - refillAmount) { avail = Math.max(0, Math.min(avail + refillAmount, limit)); } else { avail = Math.max(0, limit); } {code} I will add more unit tests for RateLimiter in the next days. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HBASE-13864) HColumnDescriptor should parse the output from master and from describe for ttl
[ https://issues.apache.org/jira/browse/HBASE-13864?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Ashu Pachauri updated HBASE-13864: -- Attachment: HBASE-13864.patch HColumnDescriptor should parse the output from master and from describe for ttl --- Key: HBASE-13864 URL: https://issues.apache.org/jira/browse/HBASE-13864 Project: HBase Issue Type: Bug Components: shell Reporter: Elliott Clark Assignee: Ashu Pachauri Attachments: HBASE-13864.patch The TTL printing on HColumnDescriptor adds a human readable time. When using that string for the create command it throws an error. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HBASE-13864) HColumnDescriptor should parse the output from master and from describe for ttl
[ https://issues.apache.org/jira/browse/HBASE-13864?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Ashu Pachauri updated HBASE-13864: -- Attachment: (was: HBASE-13684.patch) HColumnDescriptor should parse the output from master and from describe for ttl --- Key: HBASE-13864 URL: https://issues.apache.org/jira/browse/HBASE-13864 Project: HBase Issue Type: Bug Components: shell Reporter: Elliott Clark Assignee: Ashu Pachauri Attachments: HBASE-13864.patch The TTL printing on HColumnDescriptor adds a human readable time. When using that string for the create command it throws an error. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HBASE-13864) HColumnDescriptor should parse the output from master and from describe for ttl
[ https://issues.apache.org/jira/browse/HBASE-13864?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Ashu Pachauri updated HBASE-13864: -- Status: Patch Available (was: Open) HColumnDescriptor should parse the output from master and from describe for ttl --- Key: HBASE-13864 URL: https://issues.apache.org/jira/browse/HBASE-13864 Project: HBase Issue Type: Bug Components: shell Reporter: Elliott Clark Assignee: Ashu Pachauri Attachments: HBASE-13864.patch The TTL printing on HColumnDescriptor adds a human readable time. When using that string for the create command it throws an error. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HBASE-13378) RegionScannerImpl synchronized for READ_UNCOMMITTED Isolation Levels
[ https://issues.apache.org/jira/browse/HBASE-13378?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14582545#comment-14582545 ] Lars Hofhansl commented on HBASE-13378: --- Fair enough. Also... Now that I think about it again. Maybe this isn't so benign after all. Is that even what you want [~jleach]? It means that as you use this scanner HBase can at any point flush or compact the data away that the scanner should see based in MVCC. RegionScannerImpl synchronized for READ_UNCOMMITTED Isolation Levels Key: HBASE-13378 URL: https://issues.apache.org/jira/browse/HBASE-13378 Project: HBase Issue Type: New Feature Reporter: John Leach Assignee: John Leach Priority: Minor Attachments: HBASE-13378.patch, HBASE-13378.txt Original Estimate: 2h Time Spent: 2h Remaining Estimate: 0h This block of code below coupled with the close method could be changed so that READ_UNCOMMITTED does not synchronize. {CODE:JAVA} // synchronize on scannerReadPoints so that nobody calculates // getSmallestReadPoint, before scannerReadPoints is updated. IsolationLevel isolationLevel = scan.getIsolationLevel(); synchronized(scannerReadPoints) { this.readPt = getReadpoint(isolationLevel); scannerReadPoints.put(this, this.readPt); } {CODE} This hotspots for me under heavy get requests. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HBASE-13833) LoadIncrementalHFile.doBulkLoad(Path,HTable) doesn't handle unmanaged connections when using SecureBulkLoad
[ https://issues.apache.org/jira/browse/HBASE-13833?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14582512#comment-14582512 ] Hadoop QA commented on HBASE-13833: --- {color:red}-1 overall{color}. Here are the results of testing the latest attachment http://issues.apache.org/jira/secure/attachment/12739101/HBASE-13833.02.branch-1.0.patch against branch-1.0 branch at commit 47bd7de6d87f35b2c07ca83be209f73ae2b41c27. ATTACHMENT ID: 12739101 {color:green}+1 @author{color}. The patch does not contain any @author tags. {color:red}-1 tests included{color}. The patch doesn't appear to include any new or modified tests. Please justify why no new tests are needed for this patch. Also please list what manual steps were performed to verify this patch. {color:green}+1 hadoop versions{color}. The patch compiles with all supported hadoop versions (2.4.1 2.5.2 2.6.0) {color:green}+1 javac{color}. The applied patch does not increase the total number of javac compiler warnings. {color:green}+1 protoc{color}. The applied patch does not increase the total number of protoc compiler warnings. {color:green}+1 javadoc{color}. The javadoc tool did not generate any warning messages. {color:green}+1 checkstyle{color}. The applied patch does not increase the total number of checkstyle errors {color:green}+1 findbugs{color}. The patch does not introduce any new Findbugs (version 2.0.3) warnings. {color:green}+1 release audit{color}. The applied patch does not increase the total number of release audit warnings. {color:green}+1 lineLengths{color}. The patch does not introduce lines longer than 100 {color:green}+1 site{color}. The mvn site goal succeeds with this patch. {color:green}+1 core tests{color}. The patch passed unit tests in . Test results: https://builds.apache.org/job/PreCommit-HBASE-Build/14380//testReport/ Release Findbugs (version 2.0.3)warnings: https://builds.apache.org/job/PreCommit-HBASE-Build/14380//artifact/patchprocess/newFindbugsWarnings.html Checkstyle Errors: https://builds.apache.org/job/PreCommit-HBASE-Build/14380//artifact/patchprocess/checkstyle-aggregate.html Console output: https://builds.apache.org/job/PreCommit-HBASE-Build/14380//console This message is automatically generated. LoadIncrementalHFile.doBulkLoad(Path,HTable) doesn't handle unmanaged connections when using SecureBulkLoad --- Key: HBASE-13833 URL: https://issues.apache.org/jira/browse/HBASE-13833 Project: HBase Issue Type: Bug Affects Versions: 1.1.0.1 Reporter: Nick Dimiduk Assignee: Nick Dimiduk Fix For: 1.1.1 Attachments: HBASE-13833.00.branch-1.1.patch, HBASE-13833.01.branch-1.1.patch, HBASE-13833.02.branch-1.0.patch, HBASE-13833.02.branch-1.1.patch, HBASE-13833.02.branch-1.patch Seems HBASE-13328 wasn't quite sufficient. {noformat} 015-06-02 05:49:23,578|beaver.machine|INFO|2828|7140|MainThread|15/06/02 05:49:23 WARN mapreduce.LoadIncrementalHFiles: Skipping non-directory hdfs://dal-pqc1:8020/tmp/192f21dd-cc89-4354-8ba1-78d1f228e7c7/LARGE_TABLE/_SUCCESS 2015-06-02 05:49:23,720|beaver.machine|INFO|2828|7140|MainThread|15/06/02 05:49:23 INFO hfile.CacheConfig: CacheConfig:disabled 2015-06-02 05:49:23,859|beaver.machine|INFO|2828|7140|MainThread|15/06/02 05:49:23 INFO mapreduce.LoadIncrementalHFiles: Trying to load hfile=hdfs://dal-pqc1:8020/tmp/192f21dd-cc89-4354-8ba1-78d1f228e7c7/LARGE_TABLE/0/00870fd0a7544373b32b6f1e976bf47f first=\x80\x00\x00\x00 last=\x80LK? 2015-06-02 05:50:32,028|beaver.machine|INFO|2828|7140|MainThread|15/06/02 05:50:32 INFO client.RpcRetryingCaller: Call exception, tries=10, retries=35, started=68154 ms ago, cancelled=false, msg=row '' on table 'LARGE_TABLE' at region=LARGE_TABLE,,1433222865285.e01e02483f30a060d3f7abb1846ea029., hostname=dal-pqc5,16020,1433222547221, seqNum=2 2015-06-02 05:50:52,128|beaver.machine|INFO|2828|7140|MainThread|15/06/02 05:50:52 INFO client.RpcRetryingCaller: Call exception, tries=11, retries=35, started=88255 ms ago, cancelled=false, msg=row '' on table 'LARGE_TABLE' at region=LARGE_TABLE,,1433222865285.e01e02483f30a060d3f7abb1846ea029., hostname=dal-pqc5,16020,1433222547221, seqNum=2 ... ... 2015-06-02 05:01:56,121|beaver.machine|INFO|7800|2276|MainThread|15/06/02 05:01:56 ERROR mapreduce.CsvBulkLoadTool: Import job on table=LARGE_TABLE failed due to exception. 2015-06-02 05:01:56,121|beaver.machine|INFO|7800|2276|MainThread|java.io.IOException: BulkLoad encountered an unrecoverable problem 2015-06-02 05:01:56,121|beaver.machine|INFO|7800|2276|MainThread|at
[jira] [Updated] (HBASE-13560) Large compaction queue should steal from small compaction queue when idle
[ https://issues.apache.org/jira/browse/HBASE-13560?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Elliott Clark updated HBASE-13560: -- Resolution: Fixed Fix Version/s: 1.2.0 2.0.0 Hadoop Flags: Reviewed Status: Resolved (was: Patch Available) Thanks for the patch. Large compaction queue should steal from small compaction queue when idle - Key: HBASE-13560 URL: https://issues.apache.org/jira/browse/HBASE-13560 Project: HBase Issue Type: Bug Components: Compaction Affects Versions: 2.0.0, 1.2.0 Reporter: Elliott Clark Assignee: Changgeng Li Fix For: 2.0.0, 1.2.0 Attachments: HBASE-13560.patch, queuestealwork-v1.patch, queuestealwork-v4.patch, queuestealwork-v5.patch, queuestealwork-v6.patch, queuestealwork-v7.patch If you tune compaction threads so that a server is never over commited when large and small compaction threads are busy then it should be possible to have the large compactions steal work. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HBASE-13833) LoadIncrementalHFile.doBulkLoad(Path,HTable) doesn't handle unmanaged connections when using SecureBulkLoad
[ https://issues.apache.org/jira/browse/HBASE-13833?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14582495#comment-14582495 ] Enis Soztutar commented on HBASE-13833: --- Should we also close this? {code} +conn.getRegionLocator(t.getName()).getStartEndKeys(); {code} LoadIncrementalHFile.doBulkLoad(Path,HTable) doesn't handle unmanaged connections when using SecureBulkLoad --- Key: HBASE-13833 URL: https://issues.apache.org/jira/browse/HBASE-13833 Project: HBase Issue Type: Bug Affects Versions: 1.1.0.1 Reporter: Nick Dimiduk Assignee: Nick Dimiduk Fix For: 1.1.1 Attachments: HBASE-13833.00.branch-1.1.patch, HBASE-13833.01.branch-1.1.patch, HBASE-13833.02.branch-1.0.patch, HBASE-13833.02.branch-1.1.patch, HBASE-13833.02.branch-1.patch Seems HBASE-13328 wasn't quite sufficient. {noformat} 015-06-02 05:49:23,578|beaver.machine|INFO|2828|7140|MainThread|15/06/02 05:49:23 WARN mapreduce.LoadIncrementalHFiles: Skipping non-directory hdfs://dal-pqc1:8020/tmp/192f21dd-cc89-4354-8ba1-78d1f228e7c7/LARGE_TABLE/_SUCCESS 2015-06-02 05:49:23,720|beaver.machine|INFO|2828|7140|MainThread|15/06/02 05:49:23 INFO hfile.CacheConfig: CacheConfig:disabled 2015-06-02 05:49:23,859|beaver.machine|INFO|2828|7140|MainThread|15/06/02 05:49:23 INFO mapreduce.LoadIncrementalHFiles: Trying to load hfile=hdfs://dal-pqc1:8020/tmp/192f21dd-cc89-4354-8ba1-78d1f228e7c7/LARGE_TABLE/0/00870fd0a7544373b32b6f1e976bf47f first=\x80\x00\x00\x00 last=\x80LK? 2015-06-02 05:50:32,028|beaver.machine|INFO|2828|7140|MainThread|15/06/02 05:50:32 INFO client.RpcRetryingCaller: Call exception, tries=10, retries=35, started=68154 ms ago, cancelled=false, msg=row '' on table 'LARGE_TABLE' at region=LARGE_TABLE,,1433222865285.e01e02483f30a060d3f7abb1846ea029., hostname=dal-pqc5,16020,1433222547221, seqNum=2 2015-06-02 05:50:52,128|beaver.machine|INFO|2828|7140|MainThread|15/06/02 05:50:52 INFO client.RpcRetryingCaller: Call exception, tries=11, retries=35, started=88255 ms ago, cancelled=false, msg=row '' on table 'LARGE_TABLE' at region=LARGE_TABLE,,1433222865285.e01e02483f30a060d3f7abb1846ea029., hostname=dal-pqc5,16020,1433222547221, seqNum=2 ... ... 2015-06-02 05:01:56,121|beaver.machine|INFO|7800|2276|MainThread|15/06/02 05:01:56 ERROR mapreduce.CsvBulkLoadTool: Import job on table=LARGE_TABLE failed due to exception. 2015-06-02 05:01:56,121|beaver.machine|INFO|7800|2276|MainThread|java.io.IOException: BulkLoad encountered an unrecoverable problem 2015-06-02 05:01:56,121|beaver.machine|INFO|7800|2276|MainThread|at org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.bulkLoadPhase(LoadIncrementalHFiles.java:474) 2015-06-02 05:01:56,122|beaver.machine|INFO|7800|2276|MainThread|at org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.doBulkLoad(LoadIncrementalHFiles.java:405) 2015-06-02 05:01:56,122|beaver.machine|INFO|7800|2276|MainThread|at org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.doBulkLoad(LoadIncrementalHFiles.java:300) 2015-06-02 05:01:56,122|beaver.machine|INFO|7800|2276|MainThread|at org.apache.phoenix.mapreduce.CsvBulkLoadTool$TableLoader.call(CsvBulkLoadTool.java:517) 2015-06-02 05:01:56,122|beaver.machine|INFO|7800|2276|MainThread|at org.apache.phoenix.mapreduce.CsvBulkLoadTool$TableLoader.call(CsvBulkLoadTool.java:466) 2015-06-02 05:01:56,122|beaver.machine|INFO|7800|2276|MainThread|at java.util.concurrent.FutureTask.run(FutureTask.java:266) 2015-06-02 05:01:56,122|beaver.machine|INFO|7800|2276|MainThread|at org.apache.phoenix.job.JobManager$InstrumentedJobFutureTask.run(JobManager.java:172) 2015-06-02 05:01:56,122|beaver.machine|INFO|7800|2276|MainThread|at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) 2015-06-02 05:01:56,124|beaver.machine|INFO|7800|2276|MainThread|at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) 2015-06-02 05:01:56,124|beaver.machine|INFO|7800|2276|MainThread|at java.lang.Thread.run(Thread.java:745) ... ... ... 2015-06-02 05:58:34,993|beaver.machine|INFO|2828|7140|MainThread|Caused by: org.apache.hadoop.hbase.client.NeedUnmanagedConnectionException: The connection has to be unmanaged. 2015-06-02 05:58:34,993|beaver.machine|INFO|2828|7140|MainThread|at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getTable(ConnectionManager.java:724) 2015-06-02 05:58:34,994|beaver.machine|INFO|2828|7140|MainThread|at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getTable(ConnectionManager.java:708) 2015-06-02 05:58:34,994|beaver.machine|INFO|2828|7140|MainThread|at
[jira] [Commented] (HBASE-13378) RegionScannerImpl synchronized for READ_UNCOMMITTED Isolation Levels
[ https://issues.apache.org/jira/browse/HBASE-13378?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14582547#comment-14582547 ] John Leach commented on HBASE-13378: Can you point me to the line of code that uses the scannerReadPoints for determining whether to flush or compact the data? RegionScannerImpl synchronized for READ_UNCOMMITTED Isolation Levels Key: HBASE-13378 URL: https://issues.apache.org/jira/browse/HBASE-13378 Project: HBase Issue Type: New Feature Reporter: John Leach Assignee: John Leach Priority: Minor Attachments: HBASE-13378.patch, HBASE-13378.txt Original Estimate: 2h Time Spent: 2h Remaining Estimate: 0h This block of code below coupled with the close method could be changed so that READ_UNCOMMITTED does not synchronize. {CODE:JAVA} // synchronize on scannerReadPoints so that nobody calculates // getSmallestReadPoint, before scannerReadPoints is updated. IsolationLevel isolationLevel = scan.getIsolationLevel(); synchronized(scannerReadPoints) { this.readPt = getReadpoint(isolationLevel); scannerReadPoints.put(this, this.readPt); } {CODE} This hotspots for me under heavy get requests. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HBASE-13470) High level Integration test for master DDL operations
[ https://issues.apache.org/jira/browse/HBASE-13470?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14582478#comment-14582478 ] Stephen Yuan Jiang commented on HBASE-13470: The patch can be reviewed in https://reviews.apache.org/r/33603/ High level Integration test for master DDL operations - Key: HBASE-13470 URL: https://issues.apache.org/jira/browse/HBASE-13470 Project: HBase Issue Type: Sub-task Components: master Reporter: Enis Soztutar Assignee: Sophia Feng Fix For: 2.0.0, 1.2.0, 1.1.1 Attachments: HBASE-13470-v0.patch, HBASE-13470-v1.patch, HBASE-13470-v2.patch, HBASE-13470-v3.patch, HBASE-13470-v4.patch Our [~fengs] has an integration test which executes DDL operations with a new monkey to kill the active master as a high level test for the proc v2 changes. The test does random DDL operations from 20 client threads. The DDL statements are create / delete / modify / enable / disable table and CF operations. It runs HBCK to verify the end state. The test can be run on a single master, or multi master setup. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HBASE-13833) LoadIncrementalHFile.doBulkLoad(Path,HTable) doesn't handle unmanaged connections when using SecureBulkLoad
[ https://issues.apache.org/jira/browse/HBASE-13833?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14582484#comment-14582484 ] Hadoop QA commented on HBASE-13833: --- {color:red}-1 overall{color}. Here are the results of testing the latest attachment http://issues.apache.org/jira/secure/attachment/12739097/HBASE-13833.02.branch-1.patch against branch-1 branch at commit 47bd7de6d87f35b2c07ca83be209f73ae2b41c27. ATTACHMENT ID: 12739097 {color:green}+1 @author{color}. The patch does not contain any @author tags. {color:red}-1 tests included{color}. The patch doesn't appear to include any new or modified tests. Please justify why no new tests are needed for this patch. Also please list what manual steps were performed to verify this patch. {color:green}+1 hadoop versions{color}. The patch compiles with all supported hadoop versions (2.4.1 2.5.2 2.6.0) {color:green}+1 javac{color}. The applied patch does not increase the total number of javac compiler warnings. {color:green}+1 protoc{color}. The applied patch does not increase the total number of protoc compiler warnings. {color:green}+1 javadoc{color}. The javadoc tool did not generate any warning messages. {color:red}-1 checkstyle{color}. The applied patch generated 3801 checkstyle errors (more than the master's current 3800 errors). {color:green}+1 findbugs{color}. The patch does not introduce any new Findbugs (version 2.0.3) warnings. {color:green}+1 release audit{color}. The applied patch does not increase the total number of release audit warnings. {color:green}+1 lineLengths{color}. The patch does not introduce lines longer than 100 {color:green}+1 site{color}. The mvn site goal succeeds with this patch. {color:red}-1 core tests{color}. The patch failed these unit tests: org.apache.hadoop.hbase.TestRegionRebalancing {color:red}-1 core zombie tests{color}. There are 4 zombie test(s): at org.apache.hadoop.hbase.regionserver.TestSplitTransactionOnCluster.testSplitWithRegionReplicas(TestSplitTransactionOnCluster.java:996) at org.apache.camel.component.jetty.jettyproducer.HttpJettyProducerRecipientListCustomThreadPoolTest.testRecipientList(HttpJettyProducerRecipientListCustomThreadPoolTest.java:40) Test results: https://builds.apache.org/job/PreCommit-HBASE-Build/14379//testReport/ Release Findbugs (version 2.0.3)warnings: https://builds.apache.org/job/PreCommit-HBASE-Build/14379//artifact/patchprocess/newFindbugsWarnings.html Checkstyle Errors: https://builds.apache.org/job/PreCommit-HBASE-Build/14379//artifact/patchprocess/checkstyle-aggregate.html Console output: https://builds.apache.org/job/PreCommit-HBASE-Build/14379//console This message is automatically generated. LoadIncrementalHFile.doBulkLoad(Path,HTable) doesn't handle unmanaged connections when using SecureBulkLoad --- Key: HBASE-13833 URL: https://issues.apache.org/jira/browse/HBASE-13833 Project: HBase Issue Type: Bug Affects Versions: 1.1.0.1 Reporter: Nick Dimiduk Assignee: Nick Dimiduk Fix For: 1.1.1 Attachments: HBASE-13833.00.branch-1.1.patch, HBASE-13833.01.branch-1.1.patch, HBASE-13833.02.branch-1.0.patch, HBASE-13833.02.branch-1.1.patch, HBASE-13833.02.branch-1.patch Seems HBASE-13328 wasn't quite sufficient. {noformat} 015-06-02 05:49:23,578|beaver.machine|INFO|2828|7140|MainThread|15/06/02 05:49:23 WARN mapreduce.LoadIncrementalHFiles: Skipping non-directory hdfs://dal-pqc1:8020/tmp/192f21dd-cc89-4354-8ba1-78d1f228e7c7/LARGE_TABLE/_SUCCESS 2015-06-02 05:49:23,720|beaver.machine|INFO|2828|7140|MainThread|15/06/02 05:49:23 INFO hfile.CacheConfig: CacheConfig:disabled 2015-06-02 05:49:23,859|beaver.machine|INFO|2828|7140|MainThread|15/06/02 05:49:23 INFO mapreduce.LoadIncrementalHFiles: Trying to load hfile=hdfs://dal-pqc1:8020/tmp/192f21dd-cc89-4354-8ba1-78d1f228e7c7/LARGE_TABLE/0/00870fd0a7544373b32b6f1e976bf47f first=\x80\x00\x00\x00 last=\x80LK? 2015-06-02 05:50:32,028|beaver.machine|INFO|2828|7140|MainThread|15/06/02 05:50:32 INFO client.RpcRetryingCaller: Call exception, tries=10, retries=35, started=68154 ms ago, cancelled=false, msg=row '' on table 'LARGE_TABLE' at region=LARGE_TABLE,,1433222865285.e01e02483f30a060d3f7abb1846ea029., hostname=dal-pqc5,16020,1433222547221, seqNum=2 2015-06-02 05:50:52,128|beaver.machine|INFO|2828|7140|MainThread|15/06/02 05:50:52 INFO client.RpcRetryingCaller: Call exception, tries=11, retries=35, started=88255 ms ago, cancelled=false, msg=row '' on table 'LARGE_TABLE' at
[jira] [Commented] (HBASE-13889) hbase-shaded-client artifact is missing dependency (therefore, does not work)
[ https://issues.apache.org/jira/browse/HBASE-13889?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14582562#comment-14582562 ] Dmitry Minkovsky commented on HBASE-13889: -- So can [these lines|https://github.com/apache/hbase/blob/47bd7de6d87f35b2c07ca83be209f73ae2b41c27/hbase-shaded/pom.xml#L98-L101] simply be removed? Or somehow further qualified beyond {{javax}}. I built locally without these. I didn't get the same error. I did however get many other errors. Not sure if they are related though. hbase-shaded-client artifact is missing dependency (therefore, does not work) - Key: HBASE-13889 URL: https://issues.apache.org/jira/browse/HBASE-13889 Project: HBase Issue Type: Bug Components: Client Affects Versions: 1.1.0, 1.1.0.1 Environment: N/A? Reporter: Dmitry Minkovsky Priority: Blocker Fix For: 2.0.0, 1.2.0, 1.1.1 Attachments: Screen Shot 2015-06-11 at 10.59.55 AM.png The {{hbase-shaded-client}} artifact was introduced in [HBASE-13517|https://issues.apache.org/jira/browse/HBASE-13517]. Thank you very much for this, as I am new to Java building and was having a very slow-moving time resolving conflicts. However, the shaded client artifact seems to be missing {{javax.xml.transform.TransformerException}}. I examined the JAR, which does not have this package/class. Steps to reproduce: Java: {code} package com.mycompany.app; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.client.Connection; import org.apache.hadoop.hbase.client.ConnectionFactory; public class App { public static void main( String[] args ) throws java.io.IOException { Configuration config = HBaseConfiguration.create(); Connection connection = ConnectionFactory.createConnection(config); } } {code} POM: {code} project xmlns=http://maven.apache.org/POM/4.0.0; xmlns:xsi=http://www.w3.org/2001/XMLSchema-instance; xsi:schemaLocation=http://maven.apache.org/POM/4.0.0 http://maven.apache.org/xsd/maven-4.0.0.xsd; modelVersion4.0.0/modelVersion groupIdcom.mycompany.app/groupId artifactIdmy-app/artifactId
[jira] [Created] (HBASE-13890) Get/Scan from MemStore only (Client API)
Vladimir Rodionov created HBASE-13890: - Summary: Get/Scan from MemStore only (Client API) Key: HBASE-13890 URL: https://issues.apache.org/jira/browse/HBASE-13890 Project: HBase Issue Type: New Feature Components: API, Client, Scanners Reporter: Vladimir Rodionov Assignee: Vladimir Rodionov This is short-circuit read for get/scan when recent data (version) of a cell can be found only in MemStore (with very high probability). Good examples are: Atomic counters and appends. This feature will allow to bypass completely store file scanners and improve performance and latency. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HBASE-13864) HColumnDescriptor should parse the output from master and from describe for ttl
[ https://issues.apache.org/jira/browse/HBASE-13864?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14582657#comment-14582657 ] Hadoop QA commented on HBASE-13864: --- {color:red}-1 overall{color}. Here are the results of testing the latest attachment http://issues.apache.org/jira/secure/attachment/12739122/HBASE-13864.patch against master branch at commit 47bd7de6d87f35b2c07ca83be209f73ae2b41c27. ATTACHMENT ID: 12739122 {color:green}+1 @author{color}. The patch does not contain any @author tags. {color:green}+1 tests included{color}. The patch appears to include 2 new or modified tests. {color:green}+1 hadoop versions{color}. The patch compiles with all supported hadoop versions (2.4.1 2.5.2 2.6.0) {color:green}+1 javac{color}. The applied patch does not increase the total number of javac compiler warnings. {color:green}+1 protoc{color}. The applied patch does not increase the total number of protoc compiler warnings. {color:green}+1 javadoc{color}. The javadoc tool did not generate any warning messages. {color:red}-1 checkstyle{color}. The applied patch generated 1916 checkstyle errors (more than the master's current 1912 errors). {color:green}+1 findbugs{color}. The patch does not introduce any new Findbugs (version 2.0.3) warnings. {color:green}+1 release audit{color}. The applied patch does not increase the total number of release audit warnings. {color:green}+1 lineLengths{color}. The patch does not introduce lines longer than 100 {color:green}+1 site{color}. The mvn site goal succeeds with this patch. {color:red}-1 core tests{color}. The patch failed these unit tests: Test results: https://builds.apache.org/job/PreCommit-HBASE-Build/14381//testReport/ Release Findbugs (version 2.0.3)warnings: https://builds.apache.org/job/PreCommit-HBASE-Build/14381//artifact/patchprocess/newFindbugsWarnings.html Checkstyle Errors: https://builds.apache.org/job/PreCommit-HBASE-Build/14381//artifact/patchprocess/checkstyle-aggregate.html Console output: https://builds.apache.org/job/PreCommit-HBASE-Build/14381//console This message is automatically generated. HColumnDescriptor should parse the output from master and from describe for ttl --- Key: HBASE-13864 URL: https://issues.apache.org/jira/browse/HBASE-13864 Project: HBase Issue Type: Bug Components: shell Reporter: Elliott Clark Assignee: Ashu Pachauri Attachments: HBASE-13864.patch The TTL printing on HColumnDescriptor adds a human readable time. When using that string for the create command it throws an error. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HBASE-13881) Bugs in HTable incrementColumnValue implementation
[ https://issues.apache.org/jira/browse/HBASE-13881?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14582695#comment-14582695 ] Ted Yu commented on HBASE-13881: bq. writeToWAL = true is equal to Durability.SYNC_WAL? I think so. Bugs in HTable incrementColumnValue implementation -- Key: HBASE-13881 URL: https://issues.apache.org/jira/browse/HBASE-13881 Project: HBase Issue Type: Bug Components: Client Affects Versions: 0.98.6.1, 1.0.1 Reporter: Jerry Lam The exact method I'm talking about is: {code} @Deprecated @Override public long incrementColumnValue(final byte [] row, final byte [] family, final byte [] qualifier, final long amount, final boolean writeToWAL) throws IOException { return incrementColumnValue(row, family, qualifier, amount, writeToWAL? Durability.SKIP_WAL: Durability.USE_DEFAULT); } {code} Setting writeToWAL to true, Durability will be set to SKIP_WAL which does not make much sense unless the meaning of SKIP_WAL is negated. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Created] (HBASE-13891) AM should handle RegionServerStoppedException during assignment
Nick Dimiduk created HBASE-13891: Summary: AM should handle RegionServerStoppedException during assignment Key: HBASE-13891 URL: https://issues.apache.org/jira/browse/HBASE-13891 Project: HBase Issue Type: Bug Components: master, Region Assignment Affects Versions: 1.1.0.1 Reporter: Nick Dimiduk I noticed the following in the master logs {noformat} 2015-06-11 11:04:55,278 WARN [AM.ZK.Worker-pool2-t337] master.AssignmentManager: Failed assignment of SYSTEM.SEQUENCE,\x8E\x00\x00\x00,1434010321127.d2be67cf43d6bd600c7f461701ca908f. to ip-172-31-32-232.ec2.internal,16020,1434020633773, trying to assign elsewhere instead; try=1 of 10 org.apache.hadoop.hbase.regionserver.RegionServerStoppedException: org.apache.hadoop.hbase.regionserver.RegionServerStoppedException: Server ip-172-31-32-232.ec2.internal,16020,1434020633773 not running, aborting at org.apache.hadoop.hbase.regionserver.RSRpcServices.checkOpen(RSRpcServices.java:980) at org.apache.hadoop.hbase.regionserver.RSRpcServices.openRegion(RSRpcServices.java:1382) at org.apache.hadoop.hbase.protobuf.generated.AdminProtos$AdminService$2.callBlockingMethod(AdminProtos.java:22117) at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2112) at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101) at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130) at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107) at java.lang.Thread.run(Thread.java:745) at sun.reflect.GeneratedConstructorAccessor26.newInstance(Unknown Source) at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) at java.lang.reflect.Constructor.newInstance(Constructor.java:526) at org.apache.hadoop.ipc.RemoteException.instantiateException(RemoteException.java:106) at org.apache.hadoop.ipc.RemoteException.unwrapRemoteException(RemoteException.java:95) at org.apache.hadoop.hbase.protobuf.ProtobufUtil.getRemoteException(ProtobufUtil.java:322) at org.apache.hadoop.hbase.master.ServerManager.sendRegionOpen(ServerManager.java:752) at org.apache.hadoop.hbase.master.AssignmentManager.assign(AssignmentManager.java:2136) at org.apache.hadoop.hbase.master.AssignmentManager.assign(AssignmentManager.java:1590) at org.apache.hadoop.hbase.master.AssignmentManager.assign(AssignmentManager.java:1568) at org.apache.hadoop.hbase.master.handler.ClosedRegionHandler.process(ClosedRegionHandler.java:106) at org.apache.hadoop.hbase.master.AssignmentManager.handleRegion(AssignmentManager.java:1063) at org.apache.hadoop.hbase.master.AssignmentManager$6.run(AssignmentManager.java:1511) at org.apache.hadoop.hbase.master.AssignmentManager$3.run(AssignmentManager.java:1295) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471) at java.util.concurrent.FutureTask.run(FutureTask.java:262) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) at java.lang.Thread.run(Thread.java:745) Caused by: org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.regionserver.RegionServerStoppedException): org.apache.hadoop.hbase.regionserver.RegionServerStoppedException: Server ip-172-31-32-232.ec2.internal,16020,1434020633773 not running, aborting at org.apache.hadoop.hbase.regionserver.RSRpcServices.checkOpen(RSRpcServices.java:980) at org.apache.hadoop.hbase.regionserver.RSRpcServices.openRegion(RSRpcServices.java:1382) at org.apache.hadoop.hbase.protobuf.generated.AdminProtos$AdminService$2.callBlockingMethod(AdminProtos.java:22117) at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2112) at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101) at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130) at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107) at java.lang.Thread.run(Thread.java:745) at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1206) at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213) at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287) at org.apache.hadoop.hbase.protobuf.generated.AdminProtos$AdminService$BlockingStub.openRegion(AdminProtos.java:23003) at org.apache.hadoop.hbase.master.ServerManager.sendRegionOpen(ServerManager.java:749) ... 12 more ... 2015-06-11 11:04:55,289 INFO
[jira] [Commented] (HBASE-13889) hbase-shaded-client artifact is missing dependency (therefore, does not work)
[ https://issues.apache.org/jira/browse/HBASE-13889?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14582572#comment-14582572 ] Nick Dimiduk commented on HBASE-13889: -- I agree, that bit looks suspicious. I'm not sure what other classes are provided in {{javax}}; I'll have a look as well. What other errors did you get? For a start, think you can write a unit or integration test that exposes the error? Maybe it'll have to go into {{hbase-it}}. hbase-shaded-client artifact is missing dependency (therefore, does not work) - Key: HBASE-13889 URL: https://issues.apache.org/jira/browse/HBASE-13889 Project: HBase Issue Type: Bug Components: Client Affects Versions: 1.1.0, 1.1.0.1 Environment: N/A? Reporter: Dmitry Minkovsky Priority: Blocker Fix For: 2.0.0, 1.2.0, 1.1.1 Attachments: Screen Shot 2015-06-11 at 10.59.55 AM.png The {{hbase-shaded-client}} artifact was introduced in [HBASE-13517|https://issues.apache.org/jira/browse/HBASE-13517]. Thank you very much for this, as I am new to Java building and was having a very slow-moving time resolving conflicts. However, the shaded client artifact seems to be missing {{javax.xml.transform.TransformerException}}. I examined the JAR, which does not have this package/class. Steps to reproduce: Java: {code} package com.mycompany.app; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.client.Connection; import org.apache.hadoop.hbase.client.ConnectionFactory; public class App { public static void main( String[] args ) throws java.io.IOException { Configuration config = HBaseConfiguration.create(); Connection connection = ConnectionFactory.createConnection(config); } } {code} POM: {code} project xmlns=http://maven.apache.org/POM/4.0.0; xmlns:xsi=http://www.w3.org/2001/XMLSchema-instance; xsi:schemaLocation=http://maven.apache.org/POM/4.0.0 http://maven.apache.org/xsd/maven-4.0.0.xsd; modelVersion4.0.0/modelVersion groupIdcom.mycompany.app/groupId artifactIdmy-app/artifactId
[jira] [Commented] (HBASE-13885) ZK watches leaks during snapshots
[ https://issues.apache.org/jira/browse/HBASE-13885?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14582973#comment-14582973 ] Lars Hofhansl commented on HBASE-13885: --- So apparently before ZOOKEEPER-442 we must trigger a watch in order to remove it. Otherwise the watches will linger and accumulate. [~jesse_yates], [~mbertozzi], any ideas? ZK watches leaks during snapshots - Key: HBASE-13885 URL: https://issues.apache.org/jira/browse/HBASE-13885 Project: HBase Issue Type: Bug Components: snapshots Affects Versions: 0.98.12 Reporter: Abhishek Singh Chouhan Priority: Critical When taking snapshot of a table a watcher over /hbase/online-snapshot/abort/snapshot-name is created which is never cleared when the snapshot is successful. If we use snapshots to take backups daily we accumulate a lot of watches. Steps to reproduce - 1) Take snapshot of a table - snapshot 'table_1', 'abc' 2) Run the following on zk node or alternatively observe zk watches metric echo wchc | nc localhost 2181 /hbase/online-snapshot/abort/abc can be found. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HBASE-13877) Interrupt to flush from TableFlushProcedure causes dataloss in ITBLL
[ https://issues.apache.org/jira/browse/HBASE-13877?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14582909#comment-14582909 ] stack commented on HBASE-13877: --- On patch, in this section: 699 } 700 if (rows.addAndGet(1) MISSING_ROWS_TO_LOG) { 701 context.getCounter(FOUND_GROUP_KEY, keyStr + _in_ 702 + context.getInputSplit().toString()).increment(1); When YARN dumps result, this nice addition gets cut off. How you looking at it [~enis]? Patch is great. +1 after you get hadoopqa to pass. Interrupt to flush from TableFlushProcedure causes dataloss in ITBLL Key: HBASE-13877 URL: https://issues.apache.org/jira/browse/HBASE-13877 Project: HBase Issue Type: Bug Reporter: Enis Soztutar Assignee: Enis Soztutar Priority: Blocker Fix For: 2.0.0, 1.2.0, 1.1.1 Attachments: hbase-13877_v1.patch, hbase-13877_v2-branch-1.1.patch, hbase-13877_v3-branch-1.1.patch ITBLL with 1.25B rows failed for me (and Stack as reported in https://issues.apache.org/jira/browse/HBASE-13811?focusedCommentId=14577834page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-14577834) HBASE-13811 and HBASE-13853 fixed an issue with WAL edit filtering. The root cause this time seems to be different. It is due to procedure based flush interrupting the flush request in case the procedure is cancelled from an exception elsewhere. This leaves the memstore snapshot intact without aborting the server. The next flush, then flushes the previous memstore with the current seqId (as opposed to seqId from the memstore snapshot). This creates an hfile with larger seqId than what its contents are. Previous behavior in 0.98 and 1.0 (I believe) is that after flush prepare and interruption / exception will cause RS abort. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HBASE-13605) RegionStates should not keep its list of dead servers
[ https://issues.apache.org/jira/browse/HBASE-13605?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14582910#comment-14582910 ] Hadoop QA commented on HBASE-13605: --- {color:red}-1 overall{color}. Here are the results of testing the latest attachment http://issues.apache.org/jira/secure/attachment/12739154/hbase-13605_v3-branch-1.1.patch against branch-1.1 branch at commit 9d3422ed16004da1b0f9a874a98bd140b46b7a6f. ATTACHMENT ID: 12739154 {color:green}+1 @author{color}. The patch does not contain any @author tags. {color:green}+1 tests included{color}. The patch appears to include 3 new or modified tests. {color:green}+1 hadoop versions{color}. The patch compiles with all supported hadoop versions (2.4.1 2.5.2 2.6.0) {color:green}+1 javac{color}. The applied patch does not increase the total number of javac compiler warnings. {color:green}+1 protoc{color}. The applied patch does not increase the total number of protoc compiler warnings. {color:green}+1 javadoc{color}. The javadoc tool did not generate any warning messages. {color:red}-1 checkstyle{color}. The applied patch generated 3815 checkstyle errors (more than the master's current 3813 errors). {color:green}+1 findbugs{color}. The patch does not introduce any new Findbugs (version 2.0.3) warnings. {color:green}+1 release audit{color}. The applied patch does not increase the total number of release audit warnings. {color:green}+1 lineLengths{color}. The patch does not introduce lines longer than 100 {color:green}+1 site{color}. The mvn site goal succeeds with this patch. {color:red}-1 core tests{color}. The patch failed these unit tests: Test results: https://builds.apache.org/job/PreCommit-HBASE-Build/14386//testReport/ Release Findbugs (version 2.0.3)warnings: https://builds.apache.org/job/PreCommit-HBASE-Build/14386//artifact/patchprocess/newFindbugsWarnings.html Checkstyle Errors: https://builds.apache.org/job/PreCommit-HBASE-Build/14386//artifact/patchprocess/checkstyle-aggregate.html Console output: https://builds.apache.org/job/PreCommit-HBASE-Build/14386//console This message is automatically generated. RegionStates should not keep its list of dead servers - Key: HBASE-13605 URL: https://issues.apache.org/jira/browse/HBASE-13605 Project: HBase Issue Type: Bug Reporter: Enis Soztutar Assignee: Enis Soztutar Priority: Critical Fix For: 2.0.0, 1.0.2, 1.1.1 Attachments: hbase-13605_v1.patch, hbase-13605_v3-branch-1.1.patch As mentioned in https://issues.apache.org/jira/browse/HBASE-9514?focusedCommentId=13769761page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-13769761 and HBASE-12844 we should have only 1 source of cluster membership. The list of dead server and RegionStates doing it's own liveliness check (ServerManager.isServerReachable()) has caused an assignment problem again in a test cluster where the region states thinks that the server is dead and SSH will handle the region assignment. However the RS is not dead at all, living happily, and never gets zk expiry or YouAreDeadException or anything. This leaves the list of regions unassigned in OFFLINE state. master assigning the region: {code} 15-04-20 09:02:25,780 DEBUG [AM.ZK.Worker-pool3-t330] master.RegionStates: Onlined 77dddcd50c22e56bfff133c0e1f9165b on os-amb-r6-us-1429512014-hbase4-6.novalocal,16020,1429520535268 {ENCODED = 77dddcd50c {code} Master then disabled the table, and unassigned the region: {code} 2015-04-20 09:02:27,158 WARN [ProcedureExecutorThread-1] zookeeper.ZKTableStateManager: Moving table loadtest_d1 state from DISABLING to DISABLING Starting unassign of loadtest_d1,,1429520544378.77dddcd50c22e56bfff133c0e1f9165b. (offlining), current state: {77dddcd50c22e56bfff133c0e1f9165b state=OPEN, ts=1429520545780, server=os-amb-r6-us-1429512014-hbase4-6.novalocal,16020,1429520535268} bleProcedure$BulkDisabler-0] master.AssignmentManager: Sent CLOSE to os-amb-r6-us-1429512014-hbase4-6.novalocal,16020,1429520535268 for region loadtest_d1,,1429520544378.77dddcd50c22e56bfff133c0e1f9165b. 2015-04-20 09:02:27,414 INFO [AM.ZK.Worker-pool3-t316] master.RegionStates: Offlined 77dddcd50c22e56bfff133c0e1f9165b from os-amb-r6-us-1429512014-hbase4-6.novalocal,16020,1429520535268 {code} On table re-enable, AM does not assign the region: {code} 2015-04-20 09:02:30,415 INFO [ProcedureExecutorThread-3] balancer.BaseLoadBalancer: Reassigned 25 regions. 25 retained the pre-restart assignment.· 2015-04-20 09:02:30,415 INFO [ProcedureExecutorThread-3] procedure.EnableTableProcedure: Bulk assigning 25 region(s)
[jira] [Commented] (HBASE-13892) Scanner with all results filtered out results in NPE
[ https://issues.apache.org/jira/browse/HBASE-13892?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14582911#comment-14582911 ] Hudson commented on HBASE-13892: SUCCESS: Integrated in HBase-1.1 #538 (See [https://builds.apache.org/job/HBase-1.1/538/]) HBASE-13892 NPE in ClientScanner on null results array (apurtell: rev 05cef0bbdd87822b0862726c2d946588fe6e1698) * hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java * hbase-client/src/main/java/org/apache/hadoop/hbase/client/ClientScanner.java Scanner with all results filtered out results in NPE Key: HBASE-13892 URL: https://issues.apache.org/jira/browse/HBASE-13892 Project: HBase Issue Type: Bug Components: Client Affects Versions: 1.1.0 Reporter: Josh Elser Assignee: Josh Elser Priority: Critical Fix For: 2.0.0, 0.98.14, 1.0.2, 1.2.0, 1.1.1 Attachments: HBASE-13892.patch Saw a failure during some testing with region_mover.rb {code} NativeException: java.lang.NullPointerException: null __ensure__ at /usr/hdp/current/hbase-master/bin/region_mover.rb:110 isSuccessfulScan at /usr/hdp/current/hbase-master/bin/region_mover.rb:109 isSuccessfulScan at /usr/hdp/current/hbase-master/bin/region_mover.rb:104 unloadRegions at /usr/hdp/current/hbase-master/bin/region_mover.rb:328 {code} To try to get a real stacktrace, I wrote a simple test. Turns out, it was really simple to just produce the NPE within ClientScanner. {code} java.lang.NullPointerException: null at org.apache.hadoop.hbase.client.ClientScanner.getResultsToAddToCache(ClientScanner.java:576) at org.apache.hadoop.hbase.client.ClientScanner.loadCache(ClientScanner.java:492) at org.apache.hadoop.hbase.client.ClientScanner.next(ClientScanner.java:364) {code} Patch with fix and test incoming. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HBASE-13885) ZK watches leaks during snapshots
[ https://issues.apache.org/jira/browse/HBASE-13885?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14582916#comment-14582916 ] Lars Hofhansl commented on HBASE-13885: --- Making critical as it will render the cluster unusable over time. ZK watches leaks during snapshots - Key: HBASE-13885 URL: https://issues.apache.org/jira/browse/HBASE-13885 Project: HBase Issue Type: Bug Components: snapshots Affects Versions: 0.98.12 Reporter: Abhishek Singh Chouhan Priority: Critical When taking snapshot of a table a watcher over /hbase/online-snapshot/abort/snapshot-name is created which is never cleared when the snapshot is successful. If we use snapshots to take backups daily we accumulate a lot of watches. Steps to reproduce - 1) Take snapshot of a table - snapshot 'table_1', 'abc' 2) Run the following on zk node or alternatively observe zk watches metric echo wchc | nc localhost 2181 /hbase/online-snapshot/abort/abc can be found. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HBASE-13885) ZK watches leaks during snapshots
[ https://issues.apache.org/jira/browse/HBASE-13885?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Lars Hofhansl updated HBASE-13885: -- Priority: Critical (was: Major) ZK watches leaks during snapshots - Key: HBASE-13885 URL: https://issues.apache.org/jira/browse/HBASE-13885 Project: HBase Issue Type: Bug Components: snapshots Affects Versions: 0.98.12 Reporter: Abhishek Singh Chouhan Priority: Critical When taking snapshot of a table a watcher over /hbase/online-snapshot/abort/snapshot-name is created which is never cleared when the snapshot is successful. If we use snapshots to take backups daily we accumulate a lot of watches. Steps to reproduce - 1) Take snapshot of a table - snapshot 'table_1', 'abc' 2) Run the following on zk node or alternatively observe zk watches metric echo wchc | nc localhost 2181 /hbase/online-snapshot/abort/abc can be found. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HBASE-13892) Scanner with all results filtered out results in NPE
[ https://issues.apache.org/jira/browse/HBASE-13892?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14582925#comment-14582925 ] Hudson commented on HBASE-13892: SUCCESS: Integrated in HBase-1.2 #147 (See [https://builds.apache.org/job/HBase-1.2/147/]) HBASE-13892 NPE in ClientScanner on null results array (apurtell: rev e78572a985fcf76b049b22bf7bbc7734c6601077) * hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java * hbase-client/src/main/java/org/apache/hadoop/hbase/client/ClientScanner.java Scanner with all results filtered out results in NPE Key: HBASE-13892 URL: https://issues.apache.org/jira/browse/HBASE-13892 Project: HBase Issue Type: Bug Components: Client Affects Versions: 1.1.0 Reporter: Josh Elser Assignee: Josh Elser Priority: Critical Fix For: 2.0.0, 0.98.14, 1.0.2, 1.2.0, 1.1.1 Attachments: HBASE-13892.patch Saw a failure during some testing with region_mover.rb {code} NativeException: java.lang.NullPointerException: null __ensure__ at /usr/hdp/current/hbase-master/bin/region_mover.rb:110 isSuccessfulScan at /usr/hdp/current/hbase-master/bin/region_mover.rb:109 isSuccessfulScan at /usr/hdp/current/hbase-master/bin/region_mover.rb:104 unloadRegions at /usr/hdp/current/hbase-master/bin/region_mover.rb:328 {code} To try to get a real stacktrace, I wrote a simple test. Turns out, it was really simple to just produce the NPE within ClientScanner. {code} java.lang.NullPointerException: null at org.apache.hadoop.hbase.client.ClientScanner.getResultsToAddToCache(ClientScanner.java:576) at org.apache.hadoop.hbase.client.ClientScanner.loadCache(ClientScanner.java:492) at org.apache.hadoop.hbase.client.ClientScanner.next(ClientScanner.java:364) {code} Patch with fix and test incoming. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HBASE-13892) Scanner with all results filtered out results in NPE
[ https://issues.apache.org/jira/browse/HBASE-13892?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14582945#comment-14582945 ] Hudson commented on HBASE-13892: FAILURE: Integrated in HBase-TRUNK #6565 (See [https://builds.apache.org/job/HBase-TRUNK/6565/]) HBASE-13892 NPE in ClientScanner on null results array (apurtell: rev 8cef99e5062d889a748c8442595a0e0644e11458) * hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java * hbase-client/src/main/java/org/apache/hadoop/hbase/client/ClientScanner.java Scanner with all results filtered out results in NPE Key: HBASE-13892 URL: https://issues.apache.org/jira/browse/HBASE-13892 Project: HBase Issue Type: Bug Components: Client Affects Versions: 1.1.0 Reporter: Josh Elser Assignee: Josh Elser Priority: Critical Fix For: 2.0.0, 0.98.14, 1.0.2, 1.2.0, 1.1.1 Attachments: HBASE-13892.patch Saw a failure during some testing with region_mover.rb {code} NativeException: java.lang.NullPointerException: null __ensure__ at /usr/hdp/current/hbase-master/bin/region_mover.rb:110 isSuccessfulScan at /usr/hdp/current/hbase-master/bin/region_mover.rb:109 isSuccessfulScan at /usr/hdp/current/hbase-master/bin/region_mover.rb:104 unloadRegions at /usr/hdp/current/hbase-master/bin/region_mover.rb:328 {code} To try to get a real stacktrace, I wrote a simple test. Turns out, it was really simple to just produce the NPE within ClientScanner. {code} java.lang.NullPointerException: null at org.apache.hadoop.hbase.client.ClientScanner.getResultsToAddToCache(ClientScanner.java:576) at org.apache.hadoop.hbase.client.ClientScanner.loadCache(ClientScanner.java:492) at org.apache.hadoop.hbase.client.ClientScanner.next(ClientScanner.java:364) {code} Patch with fix and test incoming. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HBASE-13876) Improving performance of HeapMemoryManager
[ https://issues.apache.org/jira/browse/HBASE-13876?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Abhilash updated HBASE-13876: - Attachment: HBASE-13876-v5.patch Improving performance of HeapMemoryManager -- Key: HBASE-13876 URL: https://issues.apache.org/jira/browse/HBASE-13876 Project: HBase Issue Type: Improvement Components: hbase, regionserver Affects Versions: 2.0.0, 1.0.1, 1.1.0, 1.1.1 Reporter: Abhilash Assignee: Abhilash Priority: Minor Attachments: HBASE-13876-v2.patch, HBASE-13876-v3.patch, HBASE-13876-v4.patch, HBASE-13876-v5.patch, HBASE-13876-v5.patch, HBASE-13876.patch I am trying to improve the performance of DefaultHeapMemoryTuner by introducing some more checks. The current checks under which the DefaultHeapMemoryTuner works are very rare so I am trying to weaken these checks to improve its performance. Check current memstore size and current block cache size. If we are using less than 50% of currently available block cache size we say block cache is sufficient and same for memstore. This check will be very effective when server is either load heavy or write heavy. Earlier version just waited for number of evictions / number of flushes to be zero which are very rare. Otherwise based on percent change in number of cache misses and number of flushes we increase / decrease memory provided for caching / memstore. After doing so, on next call of HeapMemoryTuner we verify that last change has indeed decreased number of evictions / flush ( combined). I am doing this analysis by comparing percent change (which is basically nothing but normalized derivative) of number of evictions and number of flushes during last two periods. The main motive for doing this was that if we have random reads then we will be having a lot of cache misses. But even after increasing block cache we wont be able to decrease number of cache misses and we will revert back and eventually we will not waste memory on block caches. This will also help us ignore random short term spikes in reads / writes. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HBASE-13876) Improving performance of HeapMemoryManager
[ https://issues.apache.org/jira/browse/HBASE-13876?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14581503#comment-14581503 ] Abhilash commented on HBASE-13876: -- Reattached the patch because it passes that failed test on my machine. Improving performance of HeapMemoryManager -- Key: HBASE-13876 URL: https://issues.apache.org/jira/browse/HBASE-13876 Project: HBase Issue Type: Improvement Components: hbase, regionserver Affects Versions: 2.0.0, 1.0.1, 1.1.0, 1.1.1 Reporter: Abhilash Assignee: Abhilash Priority: Minor Attachments: HBASE-13876-v2.patch, HBASE-13876-v3.patch, HBASE-13876-v4.patch, HBASE-13876-v5.patch, HBASE-13876-v5.patch, HBASE-13876.patch I am trying to improve the performance of DefaultHeapMemoryTuner by introducing some more checks. The current checks under which the DefaultHeapMemoryTuner works are very rare so I am trying to weaken these checks to improve its performance. Check current memstore size and current block cache size. If we are using less than 50% of currently available block cache size we say block cache is sufficient and same for memstore. This check will be very effective when server is either load heavy or write heavy. Earlier version just waited for number of evictions / number of flushes to be zero which are very rare. Otherwise based on percent change in number of cache misses and number of flushes we increase / decrease memory provided for caching / memstore. After doing so, on next call of HeapMemoryTuner we verify that last change has indeed decreased number of evictions / flush ( combined). I am doing this analysis by comparing percent change (which is basically nothing but normalized derivative) of number of evictions and number of flushes during last two periods. The main motive for doing this was that if we have random reads then we will be having a lot of cache misses. But even after increasing block cache we wont be able to decrease number of cache misses and we will revert back and eventually we will not waste memory on block caches. This will also help us ignore random short term spikes in reads / writes. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HBASE-13828) Add group permissions testing coverage to AC.
[ https://issues.apache.org/jira/browse/HBASE-13828?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14581572#comment-14581572 ] Hudson commented on HBASE-13828: SUCCESS: Integrated in HBase-1.1 #537 (See [https://builds.apache.org/job/HBase-1.1/537/]) HBASE-13828 Add group permissions testing coverage to AC (apurtell: rev 7125dd4f97cb691462a9b79c2a818534c5c2de17) * hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestNamespaceCommands.java * hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestAccessController.java * hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestAccessController2.java Add group permissions testing coverage to AC. - Key: HBASE-13828 URL: https://issues.apache.org/jira/browse/HBASE-13828 Project: HBase Issue Type: Improvement Reporter: Srikanth Srungarapu Assignee: Ashish Singhi Fix For: 2.0.0, 0.98.13, 1.0.2, 1.2.0, 1.1.1 Attachments: HBASE-13828-0.98-v2.patch, HBASE-13828-branch-1-v2.patch, HBASE-13828-branch-1.0-v2.patch, HBASE-13828-branch-1.1-v2.patch, HBASE-13828-v1.patch, HBASE-13828-v2.patch, HBASE-13828-v3.patch, HBASE-13828.patch We suffered a regression HBASE-13826 recently due to lack of testing coverage for group permissions for AC. With the recent perf boost provided by HBASE-13658, it wouldn't be a bad idea to add checks for group level users to applicable unit tests in TestAccessController. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HBASE-13855) NPE in PartitionedMobCompactor
[ https://issues.apache.org/jira/browse/HBASE-13855?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14581585#comment-14581585 ] Jingcheng Du commented on HBASE-13855: -- Hi Jon [~jmhsieh], do you want to take a look? NPE in PartitionedMobCompactor -- Key: HBASE-13855 URL: https://issues.apache.org/jira/browse/HBASE-13855 Project: HBase Issue Type: Sub-task Components: mob Affects Versions: hbase-11339 Reporter: Jingcheng Du Assignee: Jingcheng Du Fix For: hbase-11339 Attachments: HBASE-13855.diff In PartitionedMobCompactor, mob files are split into partitions, the compactions of partitions run in parallel. The partitions share the same set of del files. There might be race conditions when open readers of del store files in each partition which can cause NPE. In this patch, we will pre-create the reader for each del store file to avoid this issue. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HBASE-13828) Add group permissions testing coverage to AC.
[ https://issues.apache.org/jira/browse/HBASE-13828?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14581591#comment-14581591 ] Hudson commented on HBASE-13828: FAILURE: Integrated in HBase-TRUNK #6562 (See [https://builds.apache.org/job/HBase-TRUNK/6562/]) HBASE-13828 Add group permissions testing coverage to AC (apurtell: rev 349cbe102a130b50852201e84dc7ac3bea4fc1f5) * hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestAccessController2.java * hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestNamespaceCommands.java * hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestAccessController.java Add group permissions testing coverage to AC. - Key: HBASE-13828 URL: https://issues.apache.org/jira/browse/HBASE-13828 Project: HBase Issue Type: Improvement Reporter: Srikanth Srungarapu Assignee: Ashish Singhi Fix For: 2.0.0, 0.98.13, 1.0.2, 1.2.0, 1.1.1 Attachments: HBASE-13828-0.98-v2.patch, HBASE-13828-branch-1-v2.patch, HBASE-13828-branch-1.0-v2.patch, HBASE-13828-branch-1.1-v2.patch, HBASE-13828-v1.patch, HBASE-13828-v2.patch, HBASE-13828-v3.patch, HBASE-13828.patch We suffered a regression HBASE-13826 recently due to lack of testing coverage for group permissions for AC. With the recent perf boost provided by HBASE-13658, it wouldn't be a bad idea to add checks for group level users to applicable unit tests in TestAccessController. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HBASE-13855) NPE in PartitionedMobCompactor
[ https://issues.apache.org/jira/browse/HBASE-13855?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Anoop Sam John updated HBASE-13855: --- Priority: Critical (was: Major) NPE in PartitionedMobCompactor -- Key: HBASE-13855 URL: https://issues.apache.org/jira/browse/HBASE-13855 Project: HBase Issue Type: Sub-task Components: mob Affects Versions: hbase-11339 Reporter: Jingcheng Du Assignee: Jingcheng Du Priority: Critical Fix For: hbase-11339 Attachments: HBASE-13855.diff In PartitionedMobCompactor, mob files are split into partitions, the compactions of partitions run in parallel. The partitions share the same set of del files. There might be race conditions when open readers of del store files in each partition which can cause NPE. In this patch, we will pre-create the reader for each del store file to avoid this issue. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HBASE-13886) Return empty value when the mob file is corrupt instead of throwing exceptions
[ https://issues.apache.org/jira/browse/HBASE-13886?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14581520#comment-14581520 ] Jingcheng Du commented on HBASE-13886: -- Yes, got it. Thanks! Return empty value when the mob file is corrupt instead of throwing exceptions -- Key: HBASE-13886 URL: https://issues.apache.org/jira/browse/HBASE-13886 Project: HBase Issue Type: Sub-task Components: mob Affects Versions: hbase-11339 Reporter: Jingcheng Du Assignee: Jingcheng Du Fix For: hbase-11339 Attachments: HBASE-13886-V2.diff, HBASE-13886.diff Now in reading, CorruptHFileException is thrown if the target mob file is corrupt. We can return empty value for that mob cell instead of throwing exceptions. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HBASE-13828) Add group permissions testing coverage to AC.
[ https://issues.apache.org/jira/browse/HBASE-13828?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14581541#comment-14581541 ] Hudson commented on HBASE-13828: SUCCESS: Integrated in HBase-1.2 #145 (See [https://builds.apache.org/job/HBase-1.2/145/]) HBASE-13828 Add group permissions testing coverage to AC (apurtell: rev c4054de40c945ab8b2b39b096894b4679c8afb15) * hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestAccessController2.java * hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestNamespaceCommands.java * hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestAccessController.java Add group permissions testing coverage to AC. - Key: HBASE-13828 URL: https://issues.apache.org/jira/browse/HBASE-13828 Project: HBase Issue Type: Improvement Reporter: Srikanth Srungarapu Assignee: Ashish Singhi Fix For: 2.0.0, 0.98.13, 1.0.2, 1.2.0, 1.1.1 Attachments: HBASE-13828-0.98-v2.patch, HBASE-13828-branch-1-v2.patch, HBASE-13828-branch-1.0-v2.patch, HBASE-13828-branch-1.1-v2.patch, HBASE-13828-v1.patch, HBASE-13828-v2.patch, HBASE-13828-v3.patch, HBASE-13828.patch We suffered a regression HBASE-13826 recently due to lack of testing coverage for group permissions for AC. With the recent perf boost provided by HBASE-13658, it wouldn't be a bad idea to add checks for group level users to applicable unit tests in TestAccessController. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HBASE-13828) Add group permissions testing coverage to AC.
[ https://issues.apache.org/jira/browse/HBASE-13828?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14581569#comment-14581569 ] Hudson commented on HBASE-13828: FAILURE: Integrated in HBase-1.0 #958 (See [https://builds.apache.org/job/HBase-1.0/958/]) HBASE-13828 Add group permissions testing coverage to AC (apurtell: rev 904ec1e4c3e7556b6ce290180c62f22469a8a608) * hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestAccessController2.java * hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestNamespaceCommands.java * hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestAccessController.java Add group permissions testing coverage to AC. - Key: HBASE-13828 URL: https://issues.apache.org/jira/browse/HBASE-13828 Project: HBase Issue Type: Improvement Reporter: Srikanth Srungarapu Assignee: Ashish Singhi Fix For: 2.0.0, 0.98.13, 1.0.2, 1.2.0, 1.1.1 Attachments: HBASE-13828-0.98-v2.patch, HBASE-13828-branch-1-v2.patch, HBASE-13828-branch-1.0-v2.patch, HBASE-13828-branch-1.1-v2.patch, HBASE-13828-v1.patch, HBASE-13828-v2.patch, HBASE-13828-v3.patch, HBASE-13828.patch We suffered a regression HBASE-13826 recently due to lack of testing coverage for group permissions for AC. With the recent perf boost provided by HBASE-13658, it wouldn't be a bad idea to add checks for group level users to applicable unit tests in TestAccessController. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HBASE-13344) Add enforcer rule that matches our JDK support statement
[ https://issues.apache.org/jira/browse/HBASE-13344?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14581508#comment-14581508 ] Sean Busbey commented on HBASE-13344: - We could try to chase down why -source and -target arguments on the maven compiler plugin don't appear to be working for us. The docs for both 1.7 and 1.8 claim that javac will produce correct bytecode when given the args. Add enforcer rule that matches our JDK support statement Key: HBASE-13344 URL: https://issues.apache.org/jira/browse/HBASE-13344 Project: HBase Issue Type: Improvement Components: build Affects Versions: 2.0.0 Reporter: Sean Busbey Assignee: Matt Warhaftig Priority: Minor Labels: beginner, maven Fix For: 2.0.0, 0.94.28, 0.98.13, 1.0.2, 1.2.0, 1.1.1 Attachments: HBASE-13344-master.patch, HBASE-13344-master_addendum_v1.patch, HBASE-13344-master_v2.patch The [ref guide gives a list of JDKs that we expect our hbase versions to work with at runtime|http://hbase.apache.org/book.html#basic.prerequisites]. Let's add in the extra-enforcer-rules mojo and start using [the bytecode version rule|http://mojo.codehaus.org/extra-enforcer-rules/enforceBytecodeVersion.html] to make sure that the result of our builds on a given branch won't fail out because of a misconfigured target jdk version (or a dependency that targets a later jdk). -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HBASE-13886) Return empty value when the mob file is corrupt instead of throwing exceptions
[ https://issues.apache.org/jira/browse/HBASE-13886?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jingcheng Du updated HBASE-13886: - Attachment: HBASE-13886-V2.diff Upload a new patch to address the comments. Thanks. Return empty value when the mob file is corrupt instead of throwing exceptions -- Key: HBASE-13886 URL: https://issues.apache.org/jira/browse/HBASE-13886 Project: HBase Issue Type: Sub-task Components: mob Affects Versions: hbase-11339 Reporter: Jingcheng Du Assignee: Jingcheng Du Fix For: hbase-11339 Attachments: HBASE-13886-V2.diff, HBASE-13886.diff Now in reading, CorruptHFileException is thrown if the target mob file is corrupt. We can return empty value for that mob cell instead of throwing exceptions. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HBASE-13344) Add enforcer rule that matches our JDK support statement
[ https://issues.apache.org/jira/browse/HBASE-13344?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14581518#comment-14581518 ] Andrew Purtell commented on HBASE-13344: another issue besides bytecode version are types from later JREs that sneak in. Add enforcer rule that matches our JDK support statement Key: HBASE-13344 URL: https://issues.apache.org/jira/browse/HBASE-13344 Project: HBase Issue Type: Improvement Components: build Affects Versions: 2.0.0 Reporter: Sean Busbey Assignee: Matt Warhaftig Priority: Minor Labels: beginner, maven Fix For: 2.0.0, 0.94.28, 0.98.13, 1.0.2, 1.2.0, 1.1.1 Attachments: HBASE-13344-master.patch, HBASE-13344-master_addendum_v1.patch, HBASE-13344-master_v2.patch The [ref guide gives a list of JDKs that we expect our hbase versions to work with at runtime|http://hbase.apache.org/book.html#basic.prerequisites]. Let's add in the extra-enforcer-rules mojo and start using [the bytecode version rule|http://mojo.codehaus.org/extra-enforcer-rules/enforceBytecodeVersion.html] to make sure that the result of our builds on a given branch won't fail out because of a misconfigured target jdk version (or a dependency that targets a later jdk). -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HBASE-13344) Add enforcer rule that matches our JDK support statement
[ https://issues.apache.org/jira/browse/HBASE-13344?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14581528#comment-14581528 ] Sean Busbey commented on HBASE-13344: - We can fix that using the Animal Sniffer plugin. Add enforcer rule that matches our JDK support statement Key: HBASE-13344 URL: https://issues.apache.org/jira/browse/HBASE-13344 Project: HBase Issue Type: Improvement Components: build Affects Versions: 2.0.0 Reporter: Sean Busbey Assignee: Matt Warhaftig Priority: Minor Labels: beginner, maven Fix For: 2.0.0, 0.94.28, 0.98.13, 1.0.2, 1.2.0, 1.1.1 Attachments: HBASE-13344-master.patch, HBASE-13344-master_addendum_v1.patch, HBASE-13344-master_v2.patch The [ref guide gives a list of JDKs that we expect our hbase versions to work with at runtime|http://hbase.apache.org/book.html#basic.prerequisites]. Let's add in the extra-enforcer-rules mojo and start using [the bytecode version rule|http://mojo.codehaus.org/extra-enforcer-rules/enforceBytecodeVersion.html] to make sure that the result of our builds on a given branch won't fail out because of a misconfigured target jdk version (or a dependency that targets a later jdk). -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HBASE-13876) Improving performance of HeapMemoryManager
[ https://issues.apache.org/jira/browse/HBASE-13876?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Abhilash updated HBASE-13876: - Description: I am trying to improve the performance of DefaultHeapMemoryTuner by introducing some more checks. The current checks under which the DefaultHeapMemoryTuner works are very rare so I am trying to weaken these checks to improve its performance. Check current memstore size and current block cache size. For say if we are using less than 50% of currently available block cache size we say block cache is sufficient and same for memstore. This check will be very effective when server is either load heavy or write heavy. Earlier version just waited for number of evictions / number of flushes to be zero which are very rare. Otherwise based on percent change in number of cache misses and number of flushes we increase / decrease memory provided for caching / memstore. After doing so, on next call of HeapMemoryTuner we verify that last change has indeed decreased number of evictions / flush either of which it was expected to do. We also check that it does not make the other (evictions / flush) increase much. I am doing this analysis by comparing percent change (which is basically nothing but normalized derivative) of number of evictions and number of flushes during last two periods. The main motive for doing this was that if we have random reads then we will be having a lot of cache misses. But even after increasing block cache we wont be able to decrease number of cache misses and we will revert back and eventually we will not waste memory on block caches. This will also help us ignore random short term spikes in reads / writes. I have also tried to take care not to tune memory if do do not have enough hints as unnecessary tuning my slow down the system. was: I am trying to improve the performance of DefaultHeapMemoryTuner by introducing some more checks. The current checks under which the DefaultHeapMemoryTuner works are very rare so I am trying to weaken these checks to improve its performance. Check current memstore size and current block cache size. If we are using less than 50% of currently available block cache size we say block cache is sufficient and same for memstore. This check will be very effective when server is either load heavy or write heavy. Earlier version just waited for number of evictions / number of flushes to be zero which are very rare. Otherwise based on percent change in number of cache misses and number of flushes we increase / decrease memory provided for caching / memstore. After doing so, on next call of HeapMemoryTuner we verify that last change has indeed decreased number of evictions / flush ( combined). I am doing this analysis by comparing percent change (which is basically nothing but normalized derivative) of number of evictions and number of flushes during last two periods. The main motive for doing this was that if we have random reads then we will be having a lot of cache misses. But even after increasing block cache we wont be able to decrease number of cache misses and we will revert back and eventually we will not waste memory on block caches. This will also help us ignore random short term spikes in reads / writes. Improving performance of HeapMemoryManager -- Key: HBASE-13876 URL: https://issues.apache.org/jira/browse/HBASE-13876 Project: HBase Issue Type: Improvement Components: hbase, regionserver Affects Versions: 2.0.0, 1.0.1, 1.1.0, 1.1.1 Reporter: Abhilash Assignee: Abhilash Priority: Minor Attachments: HBASE-13876-v2.patch, HBASE-13876-v3.patch, HBASE-13876-v4.patch, HBASE-13876-v5.patch, HBASE-13876-v5.patch, HBASE-13876.patch I am trying to improve the performance of DefaultHeapMemoryTuner by introducing some more checks. The current checks under which the DefaultHeapMemoryTuner works are very rare so I am trying to weaken these checks to improve its performance. Check current memstore size and current block cache size. For say if we are using less than 50% of currently available block cache size we say block cache is sufficient and same for memstore. This check will be very effective when server is either load heavy or write heavy. Earlier version just waited for number of evictions / number of flushes to be zero which are very rare. Otherwise based on percent change in number of cache misses and number of flushes we increase / decrease memory provided for caching / memstore. After doing so, on next call of HeapMemoryTuner we verify that last change has indeed decreased number of evictions / flush either of which it was expected to do. We also check that it does not make the other (evictions / flush) increase much. I am
[jira] [Commented] (HBASE-13891) AM should handle RegionServerStoppedException during assignment
[ https://issues.apache.org/jira/browse/HBASE-13891?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14582706#comment-14582706 ] Enis Soztutar commented on HBASE-13891: --- bq. Probably the RegionServerStoppedException should be detected and the destination of the plan be added to the dead server list. Catching this makes sense, but it is not clear how to do handling. We should not have more than one source of cluster membership (see HBASE-13605). If we for example catch this and run SSH, it means that we are using both zk and the RPC failures as a way to detect cluster membership. If we can find a way to change the target for the region assignment, that may prevent this type of assignment loop. AM should handle RegionServerStoppedException during assignment --- Key: HBASE-13891 URL: https://issues.apache.org/jira/browse/HBASE-13891 Project: HBase Issue Type: Bug Components: master, Region Assignment Affects Versions: 1.1.0.1 Reporter: Nick Dimiduk I noticed the following in the master logs {noformat} 2015-06-11 11:04:55,278 WARN [AM.ZK.Worker-pool2-t337] master.AssignmentManager: Failed assignment of SYSTEM.SEQUENCE,\x8E\x00\x00\x00,1434010321127.d2be67cf43d6bd600c7f461701ca908f. to ip-172-31-32-232.ec2.internal,16020,1434020633773, trying to assign elsewhere instead; try=1 of 10 org.apache.hadoop.hbase.regionserver.RegionServerStoppedException: org.apache.hadoop.hbase.regionserver.RegionServerStoppedException: Server ip-172-31-32-232.ec2.internal,16020,1434020633773 not running, aborting at org.apache.hadoop.hbase.regionserver.RSRpcServices.checkOpen(RSRpcServices.java:980) at org.apache.hadoop.hbase.regionserver.RSRpcServices.openRegion(RSRpcServices.java:1382) at org.apache.hadoop.hbase.protobuf.generated.AdminProtos$AdminService$2.callBlockingMethod(AdminProtos.java:22117) at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2112) at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101) at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130) at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107) at java.lang.Thread.run(Thread.java:745) at sun.reflect.GeneratedConstructorAccessor26.newInstance(Unknown Source) at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) at java.lang.reflect.Constructor.newInstance(Constructor.java:526) at org.apache.hadoop.ipc.RemoteException.instantiateException(RemoteException.java:106) at org.apache.hadoop.ipc.RemoteException.unwrapRemoteException(RemoteException.java:95) at org.apache.hadoop.hbase.protobuf.ProtobufUtil.getRemoteException(ProtobufUtil.java:322) at org.apache.hadoop.hbase.master.ServerManager.sendRegionOpen(ServerManager.java:752) at org.apache.hadoop.hbase.master.AssignmentManager.assign(AssignmentManager.java:2136) at org.apache.hadoop.hbase.master.AssignmentManager.assign(AssignmentManager.java:1590) at org.apache.hadoop.hbase.master.AssignmentManager.assign(AssignmentManager.java:1568) at org.apache.hadoop.hbase.master.handler.ClosedRegionHandler.process(ClosedRegionHandler.java:106) at org.apache.hadoop.hbase.master.AssignmentManager.handleRegion(AssignmentManager.java:1063) at org.apache.hadoop.hbase.master.AssignmentManager$6.run(AssignmentManager.java:1511) at org.apache.hadoop.hbase.master.AssignmentManager$3.run(AssignmentManager.java:1295) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471) at java.util.concurrent.FutureTask.run(FutureTask.java:262) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) at java.lang.Thread.run(Thread.java:745) Caused by: org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.regionserver.RegionServerStoppedException): org.apache.hadoop.hbase.regionserver.RegionServerStoppedException: Server ip-172-31-32-232.ec2.internal,16020,1434020633773 not running, aborting at org.apache.hadoop.hbase.regionserver.RSRpcServices.checkOpen(RSRpcServices.java:980) at org.apache.hadoop.hbase.regionserver.RSRpcServices.openRegion(RSRpcServices.java:1382) at org.apache.hadoop.hbase.protobuf.generated.AdminProtos$AdminService$2.callBlockingMethod(AdminProtos.java:22117) at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2112) at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101) at
[jira] [Updated] (HBASE-13879) Add hbase.hstore.compactionThreshold to HConstants
[ https://issues.apache.org/jira/browse/HBASE-13879?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Gabor Liptak updated HBASE-13879: - Release Note: Create hbase.hstore.compaction.min in HConstants Status: Patch Available (was: Open) Add hbase.hstore.compactionThreshold to HConstants -- Key: HBASE-13879 URL: https://issues.apache.org/jira/browse/HBASE-13879 Project: HBase Issue Type: Improvement Reporter: Gabor Liptak Priority: Minor Attachments: HBASE-13879.1.patch -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HBASE-13879) Add hbase.hstore.compactionThreshold to HConstants
[ https://issues.apache.org/jira/browse/HBASE-13879?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Gabor Liptak updated HBASE-13879: - Attachment: HBASE-13879.1.patch Add hbase.hstore.compactionThreshold to HConstants -- Key: HBASE-13879 URL: https://issues.apache.org/jira/browse/HBASE-13879 Project: HBase Issue Type: Improvement Reporter: Gabor Liptak Priority: Minor Attachments: HBASE-13879.1.patch -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Created] (HBASE-13892) Scanner with all results filtered out results in NPE
Josh Elser created HBASE-13892: -- Summary: Scanner with all results filtered out results in NPE Key: HBASE-13892 URL: https://issues.apache.org/jira/browse/HBASE-13892 Project: HBase Issue Type: Bug Components: Client Affects Versions: 1.1.0 Reporter: Josh Elser Assignee: Josh Elser Priority: Critical Fix For: 1.2.0, 1.1.1 Saw a failure during some testing with region_mover.rb {code} NativeException: java.lang.NullPointerException: null __ensure__ at /usr/hdp/current/hbase-master/bin/region_mover.rb:110 isSuccessfulScan at /usr/hdp/current/hbase-master/bin/region_mover.rb:109 isSuccessfulScan at /usr/hdp/current/hbase-master/bin/region_mover.rb:104 unloadRegions at /usr/hdp/current/hbase-master/bin/region_mover.rb:328 {code} To try to get a real stacktrace, I wrote a simple test. Turns out, it was really simple to just produce the NPE within ClientScanner. Patch with fix and test incoming. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HBASE-13877) Interrupt to flush from TableFlushProcedure causes dataloss in ITBLL
[ https://issues.apache.org/jira/browse/HBASE-13877?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Enis Soztutar updated HBASE-13877: -- Attachment: hbase-13877_v3-branch-1.1.patch Here is a v3 patch which is a little bit more comprehensive. The semantics for what to do if flushCache() and close() fails is hard since the region cannot do much by itself. The contract for flushCache() and close() when DroppedSnapshotException is thrown was already that RS should abort. Now, the patch makes it more explicit, as well as adds safeguard so that region itself calls abort if rss is passed. Since flush() can be called from multiple different callers (MemstoreFlusher, snapshot, etc), we also have to guarantee that before DSE is thrown, we put the region in closing state so that no other writes / flushes can happen. This is because we cannot call {{close(true)}} in the flushCache() since we cannot promote our read lock to a write lock. The caller should receive the DSE, then abort himself and RSS, which then calls close(true). But there is a window of time before RSS calls close(true), so no other flushes should come in while the caller handles the exception. [~saint@gmail.com], [~Apache9] does the patch make sense? It also touches upon HBASE-10514. Interrupt to flush from TableFlushProcedure causes dataloss in ITBLL Key: HBASE-13877 URL: https://issues.apache.org/jira/browse/HBASE-13877 Project: HBase Issue Type: Bug Reporter: Enis Soztutar Assignee: Enis Soztutar Priority: Blocker Fix For: 2.0.0, 1.2.0, 1.1.1 Attachments: hbase-13877_v1.patch, hbase-13877_v2-branch-1.1.patch, hbase-13877_v3-branch-1.1.patch ITBLL with 1.25B rows failed for me (and Stack as reported in https://issues.apache.org/jira/browse/HBASE-13811?focusedCommentId=14577834page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-14577834) HBASE-13811 and HBASE-13853 fixed an issue with WAL edit filtering. The root cause this time seems to be different. It is due to procedure based flush interrupting the flush request in case the procedure is cancelled from an exception elsewhere. This leaves the memstore snapshot intact without aborting the server. The next flush, then flushes the previous memstore with the current seqId (as opposed to seqId from the memstore snapshot). This creates an hfile with larger seqId than what its contents are. Previous behavior in 0.98 and 1.0 (I believe) is that after flush prepare and interruption / exception will cause RS abort. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HBASE-13560) Large compaction queue should steal from small compaction queue when idle
[ https://issues.apache.org/jira/browse/HBASE-13560?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14582734#comment-14582734 ] Hudson commented on HBASE-13560: FAILURE: Integrated in HBase-TRUNK #6564 (See [https://builds.apache.org/job/HBase-TRUNK/6564/]) HBASE-13560 large compaction thread pool will steal jobs from small compaction pool when idle (eclark: rev 9d3422ed16004da1b0f9a874a98bd140b46b7a6f) * hbase-server/src/main/java/org/apache/hadoop/hbase/util/StealJobQueue.java * hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestCompaction.java * hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompactSplitThread.java * hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestStealJobQueue.java Large compaction queue should steal from small compaction queue when idle - Key: HBASE-13560 URL: https://issues.apache.org/jira/browse/HBASE-13560 Project: HBase Issue Type: Bug Components: Compaction Affects Versions: 2.0.0, 1.2.0 Reporter: Elliott Clark Assignee: Changgeng Li Fix For: 2.0.0, 1.2.0 Attachments: HBASE-13560.patch, queuestealwork-v1.patch, queuestealwork-v4.patch, queuestealwork-v5.patch, queuestealwork-v6.patch, queuestealwork-v7.patch If you tune compaction threads so that a server is never over commited when large and small compaction threads are busy then it should be possible to have the large compactions steal work. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HBASE-13877) Interrupt to flush from TableFlushProcedure causes dataloss in ITBLL
[ https://issues.apache.org/jira/browse/HBASE-13877?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14582742#comment-14582742 ] Duo Zhang commented on HBASE-13877: --- +1 on patch v3. Interrupt to flush from TableFlushProcedure causes dataloss in ITBLL Key: HBASE-13877 URL: https://issues.apache.org/jira/browse/HBASE-13877 Project: HBase Issue Type: Bug Reporter: Enis Soztutar Assignee: Enis Soztutar Priority: Blocker Fix For: 2.0.0, 1.2.0, 1.1.1 Attachments: hbase-13877_v1.patch, hbase-13877_v2-branch-1.1.patch, hbase-13877_v3-branch-1.1.patch ITBLL with 1.25B rows failed for me (and Stack as reported in https://issues.apache.org/jira/browse/HBASE-13811?focusedCommentId=14577834page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-14577834) HBASE-13811 and HBASE-13853 fixed an issue with WAL edit filtering. The root cause this time seems to be different. It is due to procedure based flush interrupting the flush request in case the procedure is cancelled from an exception elsewhere. This leaves the memstore snapshot intact without aborting the server. The next flush, then flushes the previous memstore with the current seqId (as opposed to seqId from the memstore snapshot). This creates an hfile with larger seqId than what its contents are. Previous behavior in 0.98 and 1.0 (I believe) is that after flush prepare and interruption / exception will cause RS abort. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HBASE-13891) AM should handle RegionServerStoppedException during assignment
[ https://issues.apache.org/jira/browse/HBASE-13891?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Nick Dimiduk updated HBASE-13891: - Attachment: 13891.patch Something like this? AM should handle RegionServerStoppedException during assignment --- Key: HBASE-13891 URL: https://issues.apache.org/jira/browse/HBASE-13891 Project: HBase Issue Type: Bug Components: master, Region Assignment Affects Versions: 1.1.0.1 Reporter: Nick Dimiduk Attachments: 13891.patch I noticed the following in the master logs {noformat} 2015-06-11 11:04:55,278 WARN [AM.ZK.Worker-pool2-t337] master.AssignmentManager: Failed assignment of SYSTEM.SEQUENCE,\x8E\x00\x00\x00,1434010321127.d2be67cf43d6bd600c7f461701ca908f. to ip-172-31-32-232.ec2.internal,16020,1434020633773, trying to assign elsewhere instead; try=1 of 10 org.apache.hadoop.hbase.regionserver.RegionServerStoppedException: org.apache.hadoop.hbase.regionserver.RegionServerStoppedException: Server ip-172-31-32-232.ec2.internal,16020,1434020633773 not running, aborting at org.apache.hadoop.hbase.regionserver.RSRpcServices.checkOpen(RSRpcServices.java:980) at org.apache.hadoop.hbase.regionserver.RSRpcServices.openRegion(RSRpcServices.java:1382) at org.apache.hadoop.hbase.protobuf.generated.AdminProtos$AdminService$2.callBlockingMethod(AdminProtos.java:22117) at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2112) at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101) at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130) at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107) at java.lang.Thread.run(Thread.java:745) at sun.reflect.GeneratedConstructorAccessor26.newInstance(Unknown Source) at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) at java.lang.reflect.Constructor.newInstance(Constructor.java:526) at org.apache.hadoop.ipc.RemoteException.instantiateException(RemoteException.java:106) at org.apache.hadoop.ipc.RemoteException.unwrapRemoteException(RemoteException.java:95) at org.apache.hadoop.hbase.protobuf.ProtobufUtil.getRemoteException(ProtobufUtil.java:322) at org.apache.hadoop.hbase.master.ServerManager.sendRegionOpen(ServerManager.java:752) at org.apache.hadoop.hbase.master.AssignmentManager.assign(AssignmentManager.java:2136) at org.apache.hadoop.hbase.master.AssignmentManager.assign(AssignmentManager.java:1590) at org.apache.hadoop.hbase.master.AssignmentManager.assign(AssignmentManager.java:1568) at org.apache.hadoop.hbase.master.handler.ClosedRegionHandler.process(ClosedRegionHandler.java:106) at org.apache.hadoop.hbase.master.AssignmentManager.handleRegion(AssignmentManager.java:1063) at org.apache.hadoop.hbase.master.AssignmentManager$6.run(AssignmentManager.java:1511) at org.apache.hadoop.hbase.master.AssignmentManager$3.run(AssignmentManager.java:1295) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471) at java.util.concurrent.FutureTask.run(FutureTask.java:262) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) at java.lang.Thread.run(Thread.java:745) Caused by: org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.regionserver.RegionServerStoppedException): org.apache.hadoop.hbase.regionserver.RegionServerStoppedException: Server ip-172-31-32-232.ec2.internal,16020,1434020633773 not running, aborting at org.apache.hadoop.hbase.regionserver.RSRpcServices.checkOpen(RSRpcServices.java:980) at org.apache.hadoop.hbase.regionserver.RSRpcServices.openRegion(RSRpcServices.java:1382) at org.apache.hadoop.hbase.protobuf.generated.AdminProtos$AdminService$2.callBlockingMethod(AdminProtos.java:22117) at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2112) at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101) at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130) at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107) at java.lang.Thread.run(Thread.java:745) at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1206) at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213) at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287) at
[jira] [Updated] (HBASE-13755) Provide single super user check implementation
[ https://issues.apache.org/jira/browse/HBASE-13755?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Srikanth Srungarapu updated HBASE-13755: Resolution: Fixed Fix Version/s: (was: 1.2.0) Hadoop Flags: Reviewed Status: Resolved (was: Patch Available) Fixed AC related merge issues and pushed to master. Looks like branch-1 has issues with VC and QOS related classes. Now that you have additional privileges, I guess you can push things yourself :). Once done, please don't forget to update fix version. Provide single super user check implementation -- Key: HBASE-13755 URL: https://issues.apache.org/jira/browse/HBASE-13755 Project: HBase Issue Type: Improvement Reporter: Anoop Sam John Assignee: Mikhail Antonov Fix For: 2.0.0 Attachments: HBASE-13755-v1.patch, HBASE-13755-v2.patch, HBASE-13755-v3.patch, HBASE-13755-v3.patch, HBASE-13755-v3.patch, HBASE-13755-v3.patch, HBASE-13755-v3.patch, HBASE-13755-v4.patch, HBASE-13755-v5.patch, HBASE-13755-v6.patch Followup for HBASE-13375. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HBASE-13879) Add hbase.hstore.compactionThreshold to HConstants
[ https://issues.apache.org/jira/browse/HBASE-13879?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14582860#comment-14582860 ] Hadoop QA commented on HBASE-13879: --- {color:green}+1 overall{color}. Here are the results of testing the latest attachment http://issues.apache.org/jira/secure/attachment/12739145/HBASE-13879.1.patch against master branch at commit 9d3422ed16004da1b0f9a874a98bd140b46b7a6f. ATTACHMENT ID: 12739145 {color:green}+1 @author{color}. The patch does not contain any @author tags. {color:green}+1 tests included{color}. The patch appears to include 127 new or modified tests. {color:green}+1 hadoop versions{color}. The patch compiles with all supported hadoop versions (2.4.1 2.5.2 2.6.0) {color:green}+1 javac{color}. The applied patch does not increase the total number of javac compiler warnings. {color:green}+1 protoc{color}. The applied patch does not increase the total number of protoc compiler warnings. {color:green}+1 javadoc{color}. The javadoc tool did not generate any warning messages. {color:green}+1 checkstyle{color}. The applied patch does not increase the total number of checkstyle errors {color:green}+1 findbugs{color}. The patch does not introduce any new Findbugs (version 2.0.3) warnings. {color:green}+1 release audit{color}. The applied patch does not increase the total number of release audit warnings. {color:green}+1 lineLengths{color}. The patch does not introduce lines longer than 100 {color:green}+1 site{color}. The mvn site goal succeeds with this patch. {color:green}+1 core tests{color}. The patch passed unit tests in . Test results: https://builds.apache.org/job/PreCommit-HBASE-Build/14382//testReport/ Release Findbugs (version 2.0.3)warnings: https://builds.apache.org/job/PreCommit-HBASE-Build/14382//artifact/patchprocess/newFindbugsWarnings.html Checkstyle Errors: https://builds.apache.org/job/PreCommit-HBASE-Build/14382//artifact/patchprocess/checkstyle-aggregate.html Console output: https://builds.apache.org/job/PreCommit-HBASE-Build/14382//console This message is automatically generated. Add hbase.hstore.compactionThreshold to HConstants -- Key: HBASE-13879 URL: https://issues.apache.org/jira/browse/HBASE-13879 Project: HBase Issue Type: Improvement Reporter: Gabor Liptak Priority: Minor Attachments: HBASE-13879.1.patch -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HBASE-13892) Scanner with all results filtered out results in NPE
[ https://issues.apache.org/jira/browse/HBASE-13892?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14582879#comment-14582879 ] Hadoop QA commented on HBASE-13892: --- {color:green}+1 overall{color}. Here are the results of testing the latest attachment http://issues.apache.org/jira/secure/attachment/12739148/HBASE-13892.patch against master branch at commit 9d3422ed16004da1b0f9a874a98bd140b46b7a6f. ATTACHMENT ID: 12739148 {color:green}+1 @author{color}. The patch does not contain any @author tags. {color:green}+1 tests included{color}. The patch appears to include 4 new or modified tests. {color:green}+1 hadoop versions{color}. The patch compiles with all supported hadoop versions (2.4.1 2.5.2 2.6.0) {color:green}+1 javac{color}. The applied patch does not increase the total number of javac compiler warnings. {color:green}+1 protoc{color}. The applied patch does not increase the total number of protoc compiler warnings. {color:green}+1 javadoc{color}. The javadoc tool did not generate any warning messages. {color:green}+1 checkstyle{color}. The applied patch does not increase the total number of checkstyle errors {color:green}+1 findbugs{color}. The patch does not introduce any new Findbugs (version 2.0.3) warnings. {color:green}+1 release audit{color}. The applied patch does not increase the total number of release audit warnings. {color:green}+1 lineLengths{color}. The patch does not introduce lines longer than 100 {color:green}+1 site{color}. The mvn site goal succeeds with this patch. {color:green}+1 core tests{color}. The patch passed unit tests in . Test results: https://builds.apache.org/job/PreCommit-HBASE-Build/14384//testReport/ Release Findbugs (version 2.0.3)warnings: https://builds.apache.org/job/PreCommit-HBASE-Build/14384//artifact/patchprocess/newFindbugsWarnings.html Checkstyle Errors: https://builds.apache.org/job/PreCommit-HBASE-Build/14384//artifact/patchprocess/checkstyle-aggregate.html Console output: https://builds.apache.org/job/PreCommit-HBASE-Build/14384//console This message is automatically generated. Scanner with all results filtered out results in NPE Key: HBASE-13892 URL: https://issues.apache.org/jira/browse/HBASE-13892 Project: HBase Issue Type: Bug Components: Client Affects Versions: 1.1.0 Reporter: Josh Elser Assignee: Josh Elser Priority: Critical Fix For: 2.0.0, 0.98.14, 1.0.2, 1.2.0, 1.1.1 Attachments: HBASE-13892.patch Saw a failure during some testing with region_mover.rb {code} NativeException: java.lang.NullPointerException: null __ensure__ at /usr/hdp/current/hbase-master/bin/region_mover.rb:110 isSuccessfulScan at /usr/hdp/current/hbase-master/bin/region_mover.rb:109 isSuccessfulScan at /usr/hdp/current/hbase-master/bin/region_mover.rb:104 unloadRegions at /usr/hdp/current/hbase-master/bin/region_mover.rb:328 {code} To try to get a real stacktrace, I wrote a simple test. Turns out, it was really simple to just produce the NPE within ClientScanner. {code} java.lang.NullPointerException: null at org.apache.hadoop.hbase.client.ClientScanner.getResultsToAddToCache(ClientScanner.java:576) at org.apache.hadoop.hbase.client.ClientScanner.loadCache(ClientScanner.java:492) at org.apache.hadoop.hbase.client.ClientScanner.next(ClientScanner.java:364) {code} Patch with fix and test incoming. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HBASE-13892) Scanner with all results filtered out results in NPE
[ https://issues.apache.org/jira/browse/HBASE-13892?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14582892#comment-14582892 ] Hudson commented on HBASE-13892: FAILURE: Integrated in HBase-0.98 #1027 (See [https://builds.apache.org/job/HBase-0.98/1027/]) HBASE-13892 NPE in ClientScanner on null results array (apurtell: rev 3f31327135b784eceeafb4d417a19d441cfbd712) * hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java Scanner with all results filtered out results in NPE Key: HBASE-13892 URL: https://issues.apache.org/jira/browse/HBASE-13892 Project: HBase Issue Type: Bug Components: Client Affects Versions: 1.1.0 Reporter: Josh Elser Assignee: Josh Elser Priority: Critical Fix For: 2.0.0, 0.98.14, 1.0.2, 1.2.0, 1.1.1 Attachments: HBASE-13892.patch Saw a failure during some testing with region_mover.rb {code} NativeException: java.lang.NullPointerException: null __ensure__ at /usr/hdp/current/hbase-master/bin/region_mover.rb:110 isSuccessfulScan at /usr/hdp/current/hbase-master/bin/region_mover.rb:109 isSuccessfulScan at /usr/hdp/current/hbase-master/bin/region_mover.rb:104 unloadRegions at /usr/hdp/current/hbase-master/bin/region_mover.rb:328 {code} To try to get a real stacktrace, I wrote a simple test. Turns out, it was really simple to just produce the NPE within ClientScanner. {code} java.lang.NullPointerException: null at org.apache.hadoop.hbase.client.ClientScanner.getResultsToAddToCache(ClientScanner.java:576) at org.apache.hadoop.hbase.client.ClientScanner.loadCache(ClientScanner.java:492) at org.apache.hadoop.hbase.client.ClientScanner.next(ClientScanner.java:364) {code} Patch with fix and test incoming. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HBASE-13876) Improving performance of HeapMemoryManager
[ https://issues.apache.org/jira/browse/HBASE-13876?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14582893#comment-14582893 ] Hadoop QA commented on HBASE-13876: --- {color:green}+1 overall{color}. Here are the results of testing the latest attachment http://issues.apache.org/jira/secure/attachment/12739151/HBASE-13876-v6.patch against master branch at commit 9d3422ed16004da1b0f9a874a98bd140b46b7a6f. ATTACHMENT ID: 12739151 {color:green}+1 @author{color}. The patch does not contain any @author tags. {color:green}+1 tests included{color}. The patch appears to include 3 new or modified tests. {color:green}+1 hadoop versions{color}. The patch compiles with all supported hadoop versions (2.4.1 2.5.2 2.6.0) {color:green}+1 javac{color}. The applied patch does not increase the total number of javac compiler warnings. {color:green}+1 protoc{color}. The applied patch does not increase the total number of protoc compiler warnings. {color:green}+1 javadoc{color}. The javadoc tool did not generate any warning messages. {color:green}+1 checkstyle{color}. The applied patch does not increase the total number of checkstyle errors {color:green}+1 findbugs{color}. The patch does not introduce any new Findbugs (version 2.0.3) warnings. {color:green}+1 release audit{color}. The applied patch does not increase the total number of release audit warnings. {color:green}+1 lineLengths{color}. The patch does not introduce lines longer than 100 {color:green}+1 site{color}. The mvn site goal succeeds with this patch. {color:green}+1 core tests{color}. The patch passed unit tests in . Test results: https://builds.apache.org/job/PreCommit-HBASE-Build/14385//testReport/ Release Findbugs (version 2.0.3)warnings: https://builds.apache.org/job/PreCommit-HBASE-Build/14385//artifact/patchprocess/newFindbugsWarnings.html Checkstyle Errors: https://builds.apache.org/job/PreCommit-HBASE-Build/14385//artifact/patchprocess/checkstyle-aggregate.html Console output: https://builds.apache.org/job/PreCommit-HBASE-Build/14385//console This message is automatically generated. Improving performance of HeapMemoryManager -- Key: HBASE-13876 URL: https://issues.apache.org/jira/browse/HBASE-13876 Project: HBase Issue Type: Improvement Components: hbase, regionserver Affects Versions: 2.0.0, 1.0.1, 1.1.0, 1.1.1 Reporter: Abhilash Assignee: Abhilash Priority: Minor Attachments: HBASE-13876-v2.patch, HBASE-13876-v3.patch, HBASE-13876-v4.patch, HBASE-13876-v5.patch, HBASE-13876-v5.patch, HBASE-13876-v6.patch, HBASE-13876.patch I am trying to improve the performance of DefaultHeapMemoryTuner by introducing some more checks. The current checks under which the DefaultHeapMemoryTuner works are very rare so I am trying to weaken these checks to improve its performance. Check current memstore size and current block cache size. For say if we are using less than 50% of currently available block cache size we say block cache is sufficient and same for memstore. This check will be very effective when server is either load heavy or write heavy. Earlier version just waited for number of evictions / number of flushes to be zero which are very rare. Otherwise based on percent change in number of cache misses and number of flushes we increase / decrease memory provided for caching / memstore. After doing so, on next call of HeapMemoryTuner we verify that last change has indeed decreased number of evictions / flush either of which it was expected to do. We also check that it does not make the other (evictions / flush) increase much. I am doing this analysis by comparing percent change (which is basically nothing but normalized derivative) of number of evictions and number of flushes during last two periods. The main motive for doing this was that if we have random reads then we will be having a lot of cache misses. But even after increasing block cache we wont be able to decrease number of cache misses and we will revert back and eventually we will not waste memory on block caches. This will also help us ignore random short term spikes in reads / writes. I have also tried to take care not to tune memory if do do not have enough hints as unnecessary tuning my slow down the system. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HBASE-13892) Scanner with all results filtered out results in NPE
[ https://issues.apache.org/jira/browse/HBASE-13892?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14582902#comment-14582902 ] Hudson commented on HBASE-13892: FAILURE: Integrated in HBase-1.0 #959 (See [https://builds.apache.org/job/HBase-1.0/959/]) HBASE-13892 NPE in ClientScanner on null results array (apurtell: rev 71baa89ec45393ff5459567b069be6085ea16b8e) * hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java Scanner with all results filtered out results in NPE Key: HBASE-13892 URL: https://issues.apache.org/jira/browse/HBASE-13892 Project: HBase Issue Type: Bug Components: Client Affects Versions: 1.1.0 Reporter: Josh Elser Assignee: Josh Elser Priority: Critical Fix For: 2.0.0, 0.98.14, 1.0.2, 1.2.0, 1.1.1 Attachments: HBASE-13892.patch Saw a failure during some testing with region_mover.rb {code} NativeException: java.lang.NullPointerException: null __ensure__ at /usr/hdp/current/hbase-master/bin/region_mover.rb:110 isSuccessfulScan at /usr/hdp/current/hbase-master/bin/region_mover.rb:109 isSuccessfulScan at /usr/hdp/current/hbase-master/bin/region_mover.rb:104 unloadRegions at /usr/hdp/current/hbase-master/bin/region_mover.rb:328 {code} To try to get a real stacktrace, I wrote a simple test. Turns out, it was really simple to just produce the NPE within ClientScanner. {code} java.lang.NullPointerException: null at org.apache.hadoop.hbase.client.ClientScanner.getResultsToAddToCache(ClientScanner.java:576) at org.apache.hadoop.hbase.client.ClientScanner.loadCache(ClientScanner.java:492) at org.apache.hadoop.hbase.client.ClientScanner.next(ClientScanner.java:364) {code} Patch with fix and test incoming. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HBASE-13560) Large compaction queue should steal from small compaction queue when idle
[ https://issues.apache.org/jira/browse/HBASE-13560?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14582708#comment-14582708 ] Hudson commented on HBASE-13560: FAILURE: Integrated in HBase-1.2 #146 (See [https://builds.apache.org/job/HBase-1.2/146/]) HBASE-13560 large compaction thread pool will steal jobs from small compaction pool when idle (eclark: rev abf1aa603cbab69ea5a2cb6628a699899cf2e4ef) * hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestStealJobQueue.java * hbase-server/src/main/java/org/apache/hadoop/hbase/util/StealJobQueue.java * hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestCompaction.java * hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompactSplitThread.java Large compaction queue should steal from small compaction queue when idle - Key: HBASE-13560 URL: https://issues.apache.org/jira/browse/HBASE-13560 Project: HBase Issue Type: Bug Components: Compaction Affects Versions: 2.0.0, 1.2.0 Reporter: Elliott Clark Assignee: Changgeng Li Fix For: 2.0.0, 1.2.0 Attachments: HBASE-13560.patch, queuestealwork-v1.patch, queuestealwork-v4.patch, queuestealwork-v5.patch, queuestealwork-v6.patch, queuestealwork-v7.patch If you tune compaction threads so that a server is never over commited when large and small compaction threads are busy then it should be possible to have the large compactions steal work. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Resolved] (HBASE-13435) Scan with PrefixFilter, Range filter, column filter, or all 3 returns OutOfOrderScannerNextException
[ https://issues.apache.org/jira/browse/HBASE-13435?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Andrew Purtell resolved HBASE-13435. Resolution: Duplicate We have several JIRAs that duplicate this. Search for 'OutOfOrderScannerNextException' Scan with PrefixFilter, Range filter, column filter, or all 3 returns OutOfOrderScannerNextException Key: HBASE-13435 URL: https://issues.apache.org/jira/browse/HBASE-13435 Project: HBase Issue Type: Bug Reporter: William Watson We've run this with an hbase shell prefix filter, tried with column, range filters, and limits, and tried doing a pig script -- which we knew was going to be less performant but thought it could work with the same, simple purpose. We wanted to select a specific user's data from a few days (14 ish) worth of data. We also tried selecting a few hours worth of data as a work around, to no avail. In pig, we switched it to just give us all the data for the two week time range. The errors look like RPC timeouts, but we don't feel it should be happening and that pig/hbase/both should be able to handle these queries, if you will. The error we get in both the hbase shell and in pig boils down to possible RPC timeout?. Literally says ? in the message. We saw this stack overflow, but it's not very helpful. I also saw a few hbase tickets, none of which are super helpful and none indicate that this was an issue fixed in hbase 0.99 or anything newer that what we have. http://stackoverflow.com/questions/26437830/hbase-shell-outoforderscannernextexception-error-on-scanner-count-calls Here are the down and dirty deets: Pig script: {code} hbase_records = LOAD 'hbase://impression_event_production_hbase' USING org.apache.pig.backend.hadoop.hbase.HBaseStorage( 'cf1:uid:chararray,cf1:ts:chararray,cf1:data_regime_id:chararray,cf1:ago:chararray,cf1:ao:chararray,cf1:aca:chararray,cf1:si:chararray,cf1:ci:chararray,cf1:kv0:chararray,cf1:g_id:chararray,cf1:h_id:chararray,cf1:cg:chararray,cf1:kv1:chararray,cf1:kv2:chararray,cf1:kv3:chararray,cf1:kv4:chararray,cf1:kv5:chararray,cf1:kv6:chararray,cf1:kv7:chararray,cf1:kv8:chararray,cf1:kv9:chararray', '-loadKey=false -minTimestamp=142729920 -maxTimestamp=1428551999000') AS (uid,ts,data_regime_id,ago,ao,aca,si,ci,kv0,g_id,h_id,cg,kv1,kv2,kv3,kv4,kv5,kv6,kv7,kv8,kv9); store hbase_records into 'output_place'; {code} Error: {code} 2015-04-08 20:18:35,316 [main] INFO org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher - Failed! 2015-04-08 20:18:35,610 [main] ERROR org.apache.pig.tools.grunt.GruntParser - ERROR 2997: Unable to recreate exception from backed error: Error: org.apache.hadoop.hbase.DoNotRetryIOException: Failed after retry of OutOfOrderScannerNextException: was there a rpc timeout? at org.apache.hadoop.hbase.client.ClientScanner.next(ClientScanner.java:403) at org.apache.hadoop.hbase.mapreduce.TableRecordReaderImpl.nextKeyValue(TableRecordReaderImpl.java:232) at org.apache.hadoop.hbase.mapreduce.TableRecordReader.nextKeyValue(TableRecordReader.java:138) at org.apache.pig.backend.hadoop.hbase.HBaseTableInputFormat$HBaseTableRecordReader.nextKeyValue(HBaseTableInputFormat.java:162) at org.apache.pig.backend.hadoop.hbase.HBaseStorage.getNext(HBaseStorage.java:645) at org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigRecordReader.nextKeyValue(PigRecordReader.java:204) at org.apache.hadoop.mapred.MapTask$NewTrackingRecordReader.nextKeyValue(MapTask.java:533) at org.apache.hadoop.mapreduce.task.MapContextImpl.nextKeyValue(MapContextImpl.java:80) at org.apache.hadoop.mapreduce.lib.map.WrappedMapper$Context.nextKeyValue(WrappedMapper.java:91) at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:144) at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:764) at org.apache.hadoop.mapred.MapTask.run(MapTask.java:340) at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:167) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:415) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1557) at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:162) Caused by: org.apache.hadoop.hbase.exceptions.OutOfOrderScannerNextException: org.apache.hadoop.hbase.exceptions.OutOfOrderScannerNextException: Expected nextCallSeq: 1 But the nextCallSeq got from client: 0; request=scanner_id: 4919882396333524452 number_of_rows: 100 close_scanner: false next_call_seq: 0 at org.apache.hadoop.hbase.regionserver.HRegionServer.scan(HRegionServer.java:3110) at
[jira] [Commented] (HBASE-13378) RegionScannerImpl synchronized for READ_UNCOMMITTED Isolation Levels
[ https://issues.apache.org/jira/browse/HBASE-13378?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14582744#comment-14582744 ] Lars Hofhansl commented on HBASE-13378: --- Follow the Money... Or {{HRegion.getSmallestReadPoint()}} :) RegionScannerImpl synchronized for READ_UNCOMMITTED Isolation Levels Key: HBASE-13378 URL: https://issues.apache.org/jira/browse/HBASE-13378 Project: HBase Issue Type: New Feature Reporter: John Leach Assignee: John Leach Priority: Minor Attachments: HBASE-13378.patch, HBASE-13378.txt Original Estimate: 2h Time Spent: 2h Remaining Estimate: 0h This block of code below coupled with the close method could be changed so that READ_UNCOMMITTED does not synchronize. {CODE:JAVA} // synchronize on scannerReadPoints so that nobody calculates // getSmallestReadPoint, before scannerReadPoints is updated. IsolationLevel isolationLevel = scan.getIsolationLevel(); synchronized(scannerReadPoints) { this.readPt = getReadpoint(isolationLevel); scannerReadPoints.put(this, this.readPt); } {CODE} This hotspots for me under heavy get requests. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HBASE-13892) Scanner with all results filtered out results in NPE
[ https://issues.apache.org/jira/browse/HBASE-13892?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Josh Elser updated HBASE-13892: --- Status: Patch Available (was: Open) Scanner with all results filtered out results in NPE Key: HBASE-13892 URL: https://issues.apache.org/jira/browse/HBASE-13892 Project: HBase Issue Type: Bug Components: Client Affects Versions: 1.1.0 Reporter: Josh Elser Assignee: Josh Elser Priority: Critical Fix For: 1.2.0, 1.1.1 Attachments: HBASE-13892.patch Saw a failure during some testing with region_mover.rb {code} NativeException: java.lang.NullPointerException: null __ensure__ at /usr/hdp/current/hbase-master/bin/region_mover.rb:110 isSuccessfulScan at /usr/hdp/current/hbase-master/bin/region_mover.rb:109 isSuccessfulScan at /usr/hdp/current/hbase-master/bin/region_mover.rb:104 unloadRegions at /usr/hdp/current/hbase-master/bin/region_mover.rb:328 {code To try to get a real stacktrace, I wrote a simple test. Turns out, it was really simple to just produce the NPE within ClientScanner. {code} java.lang.NullPointerException: null at org.apache.hadoop.hbase.client.ClientScanner.getResultsToAddToCache(ClientScanner.java:576) at org.apache.hadoop.hbase.client.ClientScanner.loadCache(ClientScanner.java:492) at org.apache.hadoop.hbase.client.ClientScanner.next(ClientScanner.java:364) {code} Patch with fix and test incoming. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HBASE-13892) Scanner with all results filtered out results in NPE
[ https://issues.apache.org/jira/browse/HBASE-13892?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Josh Elser updated HBASE-13892: --- Description: Saw a failure during some testing with region_mover.rb {code} NativeException: java.lang.NullPointerException: null __ensure__ at /usr/hdp/current/hbase-master/bin/region_mover.rb:110 isSuccessfulScan at /usr/hdp/current/hbase-master/bin/region_mover.rb:109 isSuccessfulScan at /usr/hdp/current/hbase-master/bin/region_mover.rb:104 unloadRegions at /usr/hdp/current/hbase-master/bin/region_mover.rb:328 {code} To try to get a real stacktrace, I wrote a simple test. Turns out, it was really simple to just produce the NPE within ClientScanner. {code} java.lang.NullPointerException: null at org.apache.hadoop.hbase.client.ClientScanner.getResultsToAddToCache(ClientScanner.java:576) at org.apache.hadoop.hbase.client.ClientScanner.loadCache(ClientScanner.java:492) at org.apache.hadoop.hbase.client.ClientScanner.next(ClientScanner.java:364) {code} Patch with fix and test incoming. was: Saw a failure during some testing with region_mover.rb {code} NativeException: java.lang.NullPointerException: null __ensure__ at /usr/hdp/current/hbase-master/bin/region_mover.rb:110 isSuccessfulScan at /usr/hdp/current/hbase-master/bin/region_mover.rb:109 isSuccessfulScan at /usr/hdp/current/hbase-master/bin/region_mover.rb:104 unloadRegions at /usr/hdp/current/hbase-master/bin/region_mover.rb:328 {code To try to get a real stacktrace, I wrote a simple test. Turns out, it was really simple to just produce the NPE within ClientScanner. {code} java.lang.NullPointerException: null at org.apache.hadoop.hbase.client.ClientScanner.getResultsToAddToCache(ClientScanner.java:576) at org.apache.hadoop.hbase.client.ClientScanner.loadCache(ClientScanner.java:492) at org.apache.hadoop.hbase.client.ClientScanner.next(ClientScanner.java:364) {code} Patch with fix and test incoming. Scanner with all results filtered out results in NPE Key: HBASE-13892 URL: https://issues.apache.org/jira/browse/HBASE-13892 Project: HBase Issue Type: Bug Components: Client Affects Versions: 1.1.0 Reporter: Josh Elser Assignee: Josh Elser Priority: Critical Fix For: 1.2.0, 1.1.1 Attachments: HBASE-13892.patch Saw a failure during some testing with region_mover.rb {code} NativeException: java.lang.NullPointerException: null __ensure__ at /usr/hdp/current/hbase-master/bin/region_mover.rb:110 isSuccessfulScan at /usr/hdp/current/hbase-master/bin/region_mover.rb:109 isSuccessfulScan at /usr/hdp/current/hbase-master/bin/region_mover.rb:104 unloadRegions at /usr/hdp/current/hbase-master/bin/region_mover.rb:328 {code} To try to get a real stacktrace, I wrote a simple test. Turns out, it was really simple to just produce the NPE within ClientScanner. {code} java.lang.NullPointerException: null at org.apache.hadoop.hbase.client.ClientScanner.getResultsToAddToCache(ClientScanner.java:576) at org.apache.hadoop.hbase.client.ClientScanner.loadCache(ClientScanner.java:492) at org.apache.hadoop.hbase.client.ClientScanner.next(ClientScanner.java:364) {code} Patch with fix and test incoming. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HBASE-13892) Scanner with all results filtered out results in NPE
[ https://issues.apache.org/jira/browse/HBASE-13892?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14582756#comment-14582756 ] Enis Soztutar commented on HBASE-13892: --- I'm assuming only 1.1+ is affected? Scanner with all results filtered out results in NPE Key: HBASE-13892 URL: https://issues.apache.org/jira/browse/HBASE-13892 Project: HBase Issue Type: Bug Components: Client Affects Versions: 1.1.0 Reporter: Josh Elser Assignee: Josh Elser Priority: Critical Fix For: 1.2.0, 1.1.1 Attachments: HBASE-13892.patch Saw a failure during some testing with region_mover.rb {code} NativeException: java.lang.NullPointerException: null __ensure__ at /usr/hdp/current/hbase-master/bin/region_mover.rb:110 isSuccessfulScan at /usr/hdp/current/hbase-master/bin/region_mover.rb:109 isSuccessfulScan at /usr/hdp/current/hbase-master/bin/region_mover.rb:104 unloadRegions at /usr/hdp/current/hbase-master/bin/region_mover.rb:328 {code} To try to get a real stacktrace, I wrote a simple test. Turns out, it was really simple to just produce the NPE within ClientScanner. {code} java.lang.NullPointerException: null at org.apache.hadoop.hbase.client.ClientScanner.getResultsToAddToCache(ClientScanner.java:576) at org.apache.hadoop.hbase.client.ClientScanner.loadCache(ClientScanner.java:492) at org.apache.hadoop.hbase.client.ClientScanner.next(ClientScanner.java:364) {code} Patch with fix and test incoming. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HBASE-13892) Scanner with all results filtered out results in NPE
[ https://issues.apache.org/jira/browse/HBASE-13892?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14582755#comment-14582755 ] Enis Soztutar commented on HBASE-13892: --- Nice catch! +1. Scanner with all results filtered out results in NPE Key: HBASE-13892 URL: https://issues.apache.org/jira/browse/HBASE-13892 Project: HBase Issue Type: Bug Components: Client Affects Versions: 1.1.0 Reporter: Josh Elser Assignee: Josh Elser Priority: Critical Fix For: 1.2.0, 1.1.1 Attachments: HBASE-13892.patch Saw a failure during some testing with region_mover.rb {code} NativeException: java.lang.NullPointerException: null __ensure__ at /usr/hdp/current/hbase-master/bin/region_mover.rb:110 isSuccessfulScan at /usr/hdp/current/hbase-master/bin/region_mover.rb:109 isSuccessfulScan at /usr/hdp/current/hbase-master/bin/region_mover.rb:104 unloadRegions at /usr/hdp/current/hbase-master/bin/region_mover.rb:328 {code} To try to get a real stacktrace, I wrote a simple test. Turns out, it was really simple to just produce the NPE within ClientScanner. {code} java.lang.NullPointerException: null at org.apache.hadoop.hbase.client.ClientScanner.getResultsToAddToCache(ClientScanner.java:576) at org.apache.hadoop.hbase.client.ClientScanner.loadCache(ClientScanner.java:492) at org.apache.hadoop.hbase.client.ClientScanner.next(ClientScanner.java:364) {code} Patch with fix and test incoming. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HBASE-13892) Scanner with all results filtered out results in NPE
[ https://issues.apache.org/jira/browse/HBASE-13892?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14582759#comment-14582759 ] Ted Yu commented on HBASE-13892: lgtm nit: {code} 6429ResultScanner s = table.getScanner(scan); {code} s should be closed. Scanner with all results filtered out results in NPE Key: HBASE-13892 URL: https://issues.apache.org/jira/browse/HBASE-13892 Project: HBase Issue Type: Bug Components: Client Affects Versions: 1.1.0 Reporter: Josh Elser Assignee: Josh Elser Priority: Critical Fix For: 2.0.0, 1.2.0, 1.1.1 Attachments: HBASE-13892.patch Saw a failure during some testing with region_mover.rb {code} NativeException: java.lang.NullPointerException: null __ensure__ at /usr/hdp/current/hbase-master/bin/region_mover.rb:110 isSuccessfulScan at /usr/hdp/current/hbase-master/bin/region_mover.rb:109 isSuccessfulScan at /usr/hdp/current/hbase-master/bin/region_mover.rb:104 unloadRegions at /usr/hdp/current/hbase-master/bin/region_mover.rb:328 {code} To try to get a real stacktrace, I wrote a simple test. Turns out, it was really simple to just produce the NPE within ClientScanner. {code} java.lang.NullPointerException: null at org.apache.hadoop.hbase.client.ClientScanner.getResultsToAddToCache(ClientScanner.java:576) at org.apache.hadoop.hbase.client.ClientScanner.loadCache(ClientScanner.java:492) at org.apache.hadoop.hbase.client.ClientScanner.next(ClientScanner.java:364) {code} Patch with fix and test incoming. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HBASE-13892) Scanner with all results filtered out results in NPE
[ https://issues.apache.org/jira/browse/HBASE-13892?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Andrew Purtell updated HBASE-13892: --- Fix Version/s: 2.0.0 +1 Yikes, will commit shortly. I double checked 1.0 and 0.98, the code is different there and the new test passes when ported back. Will still take the new test for those branches. Scanner with all results filtered out results in NPE Key: HBASE-13892 URL: https://issues.apache.org/jira/browse/HBASE-13892 Project: HBase Issue Type: Bug Components: Client Affects Versions: 1.1.0 Reporter: Josh Elser Assignee: Josh Elser Priority: Critical Fix For: 2.0.0, 1.2.0, 1.1.1 Attachments: HBASE-13892.patch Saw a failure during some testing with region_mover.rb {code} NativeException: java.lang.NullPointerException: null __ensure__ at /usr/hdp/current/hbase-master/bin/region_mover.rb:110 isSuccessfulScan at /usr/hdp/current/hbase-master/bin/region_mover.rb:109 isSuccessfulScan at /usr/hdp/current/hbase-master/bin/region_mover.rb:104 unloadRegions at /usr/hdp/current/hbase-master/bin/region_mover.rb:328 {code} To try to get a real stacktrace, I wrote a simple test. Turns out, it was really simple to just produce the NPE within ClientScanner. {code} java.lang.NullPointerException: null at org.apache.hadoop.hbase.client.ClientScanner.getResultsToAddToCache(ClientScanner.java:576) at org.apache.hadoop.hbase.client.ClientScanner.loadCache(ClientScanner.java:492) at org.apache.hadoop.hbase.client.ClientScanner.next(ClientScanner.java:364) {code} Patch with fix and test incoming. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HBASE-13892) Scanner with all results filtered out results in NPE
[ https://issues.apache.org/jira/browse/HBASE-13892?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14582758#comment-14582758 ] Josh Elser commented on HBASE-13892: I think this might have been introduced in HBASE-11544 (cc/ [~jonathan.lawlor]). Scanner with all results filtered out results in NPE Key: HBASE-13892 URL: https://issues.apache.org/jira/browse/HBASE-13892 Project: HBase Issue Type: Bug Components: Client Affects Versions: 1.1.0 Reporter: Josh Elser Assignee: Josh Elser Priority: Critical Fix For: 2.0.0, 1.2.0, 1.1.1 Attachments: HBASE-13892.patch Saw a failure during some testing with region_mover.rb {code} NativeException: java.lang.NullPointerException: null __ensure__ at /usr/hdp/current/hbase-master/bin/region_mover.rb:110 isSuccessfulScan at /usr/hdp/current/hbase-master/bin/region_mover.rb:109 isSuccessfulScan at /usr/hdp/current/hbase-master/bin/region_mover.rb:104 unloadRegions at /usr/hdp/current/hbase-master/bin/region_mover.rb:328 {code} To try to get a real stacktrace, I wrote a simple test. Turns out, it was really simple to just produce the NPE within ClientScanner. {code} java.lang.NullPointerException: null at org.apache.hadoop.hbase.client.ClientScanner.getResultsToAddToCache(ClientScanner.java:576) at org.apache.hadoop.hbase.client.ClientScanner.loadCache(ClientScanner.java:492) at org.apache.hadoop.hbase.client.ClientScanner.next(ClientScanner.java:364) {code} Patch with fix and test incoming. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HBASE-13892) Scanner with all results filtered out results in NPE
[ https://issues.apache.org/jira/browse/HBASE-13892?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14582764#comment-14582764 ] Andrew Purtell commented on HBASE-13892: bq. s should be closed. Will fix that on commit. I have them staged. Anything else? Scanner with all results filtered out results in NPE Key: HBASE-13892 URL: https://issues.apache.org/jira/browse/HBASE-13892 Project: HBase Issue Type: Bug Components: Client Affects Versions: 1.1.0 Reporter: Josh Elser Assignee: Josh Elser Priority: Critical Fix For: 2.0.0, 1.2.0, 1.1.1 Attachments: HBASE-13892.patch Saw a failure during some testing with region_mover.rb {code} NativeException: java.lang.NullPointerException: null __ensure__ at /usr/hdp/current/hbase-master/bin/region_mover.rb:110 isSuccessfulScan at /usr/hdp/current/hbase-master/bin/region_mover.rb:109 isSuccessfulScan at /usr/hdp/current/hbase-master/bin/region_mover.rb:104 unloadRegions at /usr/hdp/current/hbase-master/bin/region_mover.rb:328 {code} To try to get a real stacktrace, I wrote a simple test. Turns out, it was really simple to just produce the NPE within ClientScanner. {code} java.lang.NullPointerException: null at org.apache.hadoop.hbase.client.ClientScanner.getResultsToAddToCache(ClientScanner.java:576) at org.apache.hadoop.hbase.client.ClientScanner.loadCache(ClientScanner.java:492) at org.apache.hadoop.hbase.client.ClientScanner.next(ClientScanner.java:364) {code} Patch with fix and test incoming. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HBASE-13892) Scanner with all results filtered out results in NPE
[ https://issues.apache.org/jira/browse/HBASE-13892?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14582772#comment-14582772 ] Nick Dimiduk commented on HBASE-13892: -- +1 Scanner with all results filtered out results in NPE Key: HBASE-13892 URL: https://issues.apache.org/jira/browse/HBASE-13892 Project: HBase Issue Type: Bug Components: Client Affects Versions: 1.1.0 Reporter: Josh Elser Assignee: Josh Elser Priority: Critical Fix For: 2.0.0, 1.2.0, 1.1.1 Attachments: HBASE-13892.patch Saw a failure during some testing with region_mover.rb {code} NativeException: java.lang.NullPointerException: null __ensure__ at /usr/hdp/current/hbase-master/bin/region_mover.rb:110 isSuccessfulScan at /usr/hdp/current/hbase-master/bin/region_mover.rb:109 isSuccessfulScan at /usr/hdp/current/hbase-master/bin/region_mover.rb:104 unloadRegions at /usr/hdp/current/hbase-master/bin/region_mover.rb:328 {code} To try to get a real stacktrace, I wrote a simple test. Turns out, it was really simple to just produce the NPE within ClientScanner. {code} java.lang.NullPointerException: null at org.apache.hadoop.hbase.client.ClientScanner.getResultsToAddToCache(ClientScanner.java:576) at org.apache.hadoop.hbase.client.ClientScanner.loadCache(ClientScanner.java:492) at org.apache.hadoop.hbase.client.ClientScanner.next(ClientScanner.java:364) {code} Patch with fix and test incoming. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HBASE-13876) Improving performance of HeapMemoryManager
[ https://issues.apache.org/jira/browse/HBASE-13876?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Abhilash updated HBASE-13876: - Attachment: HBASE-13876-v6.patch Improving performance of HeapMemoryManager -- Key: HBASE-13876 URL: https://issues.apache.org/jira/browse/HBASE-13876 Project: HBase Issue Type: Improvement Components: hbase, regionserver Affects Versions: 2.0.0, 1.0.1, 1.1.0, 1.1.1 Reporter: Abhilash Assignee: Abhilash Priority: Minor Attachments: HBASE-13876-v2.patch, HBASE-13876-v3.patch, HBASE-13876-v4.patch, HBASE-13876-v5.patch, HBASE-13876-v5.patch, HBASE-13876-v6.patch, HBASE-13876.patch I am trying to improve the performance of DefaultHeapMemoryTuner by introducing some more checks. The current checks under which the DefaultHeapMemoryTuner works are very rare so I am trying to weaken these checks to improve its performance. Check current memstore size and current block cache size. For say if we are using less than 50% of currently available block cache size we say block cache is sufficient and same for memstore. This check will be very effective when server is either load heavy or write heavy. Earlier version just waited for number of evictions / number of flushes to be zero which are very rare. Otherwise based on percent change in number of cache misses and number of flushes we increase / decrease memory provided for caching / memstore. After doing so, on next call of HeapMemoryTuner we verify that last change has indeed decreased number of evictions / flush either of which it was expected to do. We also check that it does not make the other (evictions / flush) increase much. I am doing this analysis by comparing percent change (which is basically nothing but normalized derivative) of number of evictions and number of flushes during last two periods. The main motive for doing this was that if we have random reads then we will be having a lot of cache misses. But even after increasing block cache we wont be able to decrease number of cache misses and we will revert back and eventually we will not waste memory on block caches. This will also help us ignore random short term spikes in reads / writes. I have also tried to take care not to tune memory if do do not have enough hints as unnecessary tuning my slow down the system. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HBASE-13876) Improving performance of HeapMemoryManager
[ https://issues.apache.org/jira/browse/HBASE-13876?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14582778#comment-14582778 ] Abhilash commented on HBASE-13876: -- Using stats from few past lookup periods(configurable) to decide tuner step. Improving performance of HeapMemoryManager -- Key: HBASE-13876 URL: https://issues.apache.org/jira/browse/HBASE-13876 Project: HBase Issue Type: Improvement Components: hbase, regionserver Affects Versions: 2.0.0, 1.0.1, 1.1.0, 1.1.1 Reporter: Abhilash Assignee: Abhilash Priority: Minor Attachments: HBASE-13876-v2.patch, HBASE-13876-v3.patch, HBASE-13876-v4.patch, HBASE-13876-v5.patch, HBASE-13876-v5.patch, HBASE-13876-v6.patch, HBASE-13876.patch I am trying to improve the performance of DefaultHeapMemoryTuner by introducing some more checks. The current checks under which the DefaultHeapMemoryTuner works are very rare so I am trying to weaken these checks to improve its performance. Check current memstore size and current block cache size. For say if we are using less than 50% of currently available block cache size we say block cache is sufficient and same for memstore. This check will be very effective when server is either load heavy or write heavy. Earlier version just waited for number of evictions / number of flushes to be zero which are very rare. Otherwise based on percent change in number of cache misses and number of flushes we increase / decrease memory provided for caching / memstore. After doing so, on next call of HeapMemoryTuner we verify that last change has indeed decreased number of evictions / flush either of which it was expected to do. We also check that it does not make the other (evictions / flush) increase much. I am doing this analysis by comparing percent change (which is basically nothing but normalized derivative) of number of evictions and number of flushes during last two periods. The main motive for doing this was that if we have random reads then we will be having a lot of cache misses. But even after increasing block cache we wont be able to decrease number of cache misses and we will revert back and eventually we will not waste memory on block caches. This will also help us ignore random short term spikes in reads / writes. I have also tried to take care not to tune memory if do do not have enough hints as unnecessary tuning my slow down the system. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HBASE-13877) Interrupt to flush from TableFlushProcedure causes dataloss in ITBLL
[ https://issues.apache.org/jira/browse/HBASE-13877?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14582858#comment-14582858 ] Hadoop QA commented on HBASE-13877: --- {color:red}-1 overall{color}. Here are the results of testing the latest attachment http://issues.apache.org/jira/secure/attachment/12739146/hbase-13877_v3-branch-1.1.patch against branch-1.1 branch at commit 9d3422ed16004da1b0f9a874a98bd140b46b7a6f. ATTACHMENT ID: 12739146 {color:green}+1 @author{color}. The patch does not contain any @author tags. {color:green}+1 tests included{color}. The patch appears to include 6 new or modified tests. {color:green}+1 hadoop versions{color}. The patch compiles with all supported hadoop versions (2.4.1 2.5.2 2.6.0) {color:green}+1 javac{color}. The applied patch does not increase the total number of javac compiler warnings. {color:green}+1 protoc{color}. The applied patch does not increase the total number of protoc compiler warnings. {color:green}+1 javadoc{color}. The javadoc tool did not generate any warning messages. {color:green}+1 checkstyle{color}. The applied patch does not increase the total number of checkstyle errors {color:green}+1 findbugs{color}. The patch does not introduce any new Findbugs (version 2.0.3) warnings. {color:green}+1 release audit{color}. The applied patch does not increase the total number of release audit warnings. {color:green}+1 lineLengths{color}. The patch does not introduce lines longer than 100 {color:green}+1 site{color}. The mvn site goal succeeds with this patch. {color:red}-1 core tests{color}. The patch failed these unit tests: org.apache.hadoop.hbase.regionserver.wal.TestSecureWALReplay org.apache.hadoop.hbase.regionserver.wal.TestWALReplay org.apache.hadoop.hbase.regionserver.wal.TestWALReplayCompressed Test results: https://builds.apache.org/job/PreCommit-HBASE-Build/14383//testReport/ Release Findbugs (version 2.0.3)warnings: https://builds.apache.org/job/PreCommit-HBASE-Build/14383//artifact/patchprocess/newFindbugsWarnings.html Checkstyle Errors: https://builds.apache.org/job/PreCommit-HBASE-Build/14383//artifact/patchprocess/checkstyle-aggregate.html Console output: https://builds.apache.org/job/PreCommit-HBASE-Build/14383//console This message is automatically generated. Interrupt to flush from TableFlushProcedure causes dataloss in ITBLL Key: HBASE-13877 URL: https://issues.apache.org/jira/browse/HBASE-13877 Project: HBase Issue Type: Bug Reporter: Enis Soztutar Assignee: Enis Soztutar Priority: Blocker Fix For: 2.0.0, 1.2.0, 1.1.1 Attachments: hbase-13877_v1.patch, hbase-13877_v2-branch-1.1.patch, hbase-13877_v3-branch-1.1.patch ITBLL with 1.25B rows failed for me (and Stack as reported in https://issues.apache.org/jira/browse/HBASE-13811?focusedCommentId=14577834page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-14577834) HBASE-13811 and HBASE-13853 fixed an issue with WAL edit filtering. The root cause this time seems to be different. It is due to procedure based flush interrupting the flush request in case the procedure is cancelled from an exception elsewhere. This leaves the memstore snapshot intact without aborting the server. The next flush, then flushes the previous memstore with the current seqId (as opposed to seqId from the memstore snapshot). This creates an hfile with larger seqId than what its contents are. Previous behavior in 0.98 and 1.0 (I believe) is that after flush prepare and interruption / exception will cause RS abort. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HBASE-13892) Scanner with all results filtered out results in NPE
[ https://issues.apache.org/jira/browse/HBASE-13892?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Josh Elser updated HBASE-13892: --- Description: Saw a failure during some testing with region_mover.rb {code} NativeException: java.lang.NullPointerException: null __ensure__ at /usr/hdp/current/hbase-master/bin/region_mover.rb:110 isSuccessfulScan at /usr/hdp/current/hbase-master/bin/region_mover.rb:109 isSuccessfulScan at /usr/hdp/current/hbase-master/bin/region_mover.rb:104 unloadRegions at /usr/hdp/current/hbase-master/bin/region_mover.rb:328 {code To try to get a real stacktrace, I wrote a simple test. Turns out, it was really simple to just produce the NPE within ClientScanner. {code} java.lang.NullPointerException: null at org.apache.hadoop.hbase.client.ClientScanner.getResultsToAddToCache(ClientScanner.java:576) at org.apache.hadoop.hbase.client.ClientScanner.loadCache(ClientScanner.java:492) at org.apache.hadoop.hbase.client.ClientScanner.next(ClientScanner.java:364) {code} Patch with fix and test incoming. was: Saw a failure during some testing with region_mover.rb {code} NativeException: java.lang.NullPointerException: null __ensure__ at /usr/hdp/current/hbase-master/bin/region_mover.rb:110 isSuccessfulScan at /usr/hdp/current/hbase-master/bin/region_mover.rb:109 isSuccessfulScan at /usr/hdp/current/hbase-master/bin/region_mover.rb:104 unloadRegions at /usr/hdp/current/hbase-master/bin/region_mover.rb:328 {code} To try to get a real stacktrace, I wrote a simple test. Turns out, it was really simple to just produce the NPE within ClientScanner. Patch with fix and test incoming. Scanner with all results filtered out results in NPE Key: HBASE-13892 URL: https://issues.apache.org/jira/browse/HBASE-13892 Project: HBase Issue Type: Bug Components: Client Affects Versions: 1.1.0 Reporter: Josh Elser Assignee: Josh Elser Priority: Critical Fix For: 1.2.0, 1.1.1 Saw a failure during some testing with region_mover.rb {code} NativeException: java.lang.NullPointerException: null __ensure__ at /usr/hdp/current/hbase-master/bin/region_mover.rb:110 isSuccessfulScan at /usr/hdp/current/hbase-master/bin/region_mover.rb:109 isSuccessfulScan at /usr/hdp/current/hbase-master/bin/region_mover.rb:104 unloadRegions at /usr/hdp/current/hbase-master/bin/region_mover.rb:328 {code To try to get a real stacktrace, I wrote a simple test. Turns out, it was really simple to just produce the NPE within ClientScanner. {code} java.lang.NullPointerException: null at org.apache.hadoop.hbase.client.ClientScanner.getResultsToAddToCache(ClientScanner.java:576) at org.apache.hadoop.hbase.client.ClientScanner.loadCache(ClientScanner.java:492) at org.apache.hadoop.hbase.client.ClientScanner.next(ClientScanner.java:364) {code} Patch with fix and test incoming. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HBASE-13892) Scanner with all results filtered out results in NPE
[ https://issues.apache.org/jira/browse/HBASE-13892?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Josh Elser updated HBASE-13892: --- Attachment: HBASE-13892.patch Scanner with all results filtered out results in NPE Key: HBASE-13892 URL: https://issues.apache.org/jira/browse/HBASE-13892 Project: HBase Issue Type: Bug Components: Client Affects Versions: 1.1.0 Reporter: Josh Elser Assignee: Josh Elser Priority: Critical Fix For: 1.2.0, 1.1.1 Attachments: HBASE-13892.patch Saw a failure during some testing with region_mover.rb {code} NativeException: java.lang.NullPointerException: null __ensure__ at /usr/hdp/current/hbase-master/bin/region_mover.rb:110 isSuccessfulScan at /usr/hdp/current/hbase-master/bin/region_mover.rb:109 isSuccessfulScan at /usr/hdp/current/hbase-master/bin/region_mover.rb:104 unloadRegions at /usr/hdp/current/hbase-master/bin/region_mover.rb:328 {code To try to get a real stacktrace, I wrote a simple test. Turns out, it was really simple to just produce the NPE within ClientScanner. {code} java.lang.NullPointerException: null at org.apache.hadoop.hbase.client.ClientScanner.getResultsToAddToCache(ClientScanner.java:576) at org.apache.hadoop.hbase.client.ClientScanner.loadCache(ClientScanner.java:492) at org.apache.hadoop.hbase.client.ClientScanner.next(ClientScanner.java:364) {code} Patch with fix and test incoming. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HBASE-13892) Scanner with all results filtered out results in NPE
[ https://issues.apache.org/jira/browse/HBASE-13892?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Andrew Purtell updated HBASE-13892: --- Resolution: Fixed Fix Version/s: 1.0.2 0.98.14 Hadoop Flags: Reviewed Status: Resolved (was: Patch Available) Committed to = 0.98: new test plus fix to 1.1+, only new test to 0.98 and 1.0. New test passes on all branches. I amended the tests to use try-with-resources (for 0.98, try-finally) to clean up resources as Ted suggested. Happy to help with the branch porting mechanics [~elserj], thanks for finding and fixing this. Scanner with all results filtered out results in NPE Key: HBASE-13892 URL: https://issues.apache.org/jira/browse/HBASE-13892 Project: HBase Issue Type: Bug Components: Client Affects Versions: 1.1.0 Reporter: Josh Elser Assignee: Josh Elser Priority: Critical Fix For: 2.0.0, 0.98.14, 1.0.2, 1.2.0, 1.1.1 Attachments: HBASE-13892.patch Saw a failure during some testing with region_mover.rb {code} NativeException: java.lang.NullPointerException: null __ensure__ at /usr/hdp/current/hbase-master/bin/region_mover.rb:110 isSuccessfulScan at /usr/hdp/current/hbase-master/bin/region_mover.rb:109 isSuccessfulScan at /usr/hdp/current/hbase-master/bin/region_mover.rb:104 unloadRegions at /usr/hdp/current/hbase-master/bin/region_mover.rb:328 {code} To try to get a real stacktrace, I wrote a simple test. Turns out, it was really simple to just produce the NPE within ClientScanner. {code} java.lang.NullPointerException: null at org.apache.hadoop.hbase.client.ClientScanner.getResultsToAddToCache(ClientScanner.java:576) at org.apache.hadoop.hbase.client.ClientScanner.loadCache(ClientScanner.java:492) at org.apache.hadoop.hbase.client.ClientScanner.next(ClientScanner.java:364) {code} Patch with fix and test incoming. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HBASE-13855) NPE in PartitionedMobCompactor
[ https://issues.apache.org/jira/browse/HBASE-13855?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14581623#comment-14581623 ] Anoop Sam John commented on HBASE-13855: This is a critical issue and am going to commit it now. [~jmhsieh] pls check once the test caused NPE for you before this patch. NPE in PartitionedMobCompactor -- Key: HBASE-13855 URL: https://issues.apache.org/jira/browse/HBASE-13855 Project: HBase Issue Type: Sub-task Components: mob Affects Versions: hbase-11339 Reporter: Jingcheng Du Assignee: Jingcheng Du Priority: Critical Fix For: hbase-11339 Attachments: HBASE-13855.diff In PartitionedMobCompactor, mob files are split into partitions, the compactions of partitions run in parallel. The partitions share the same set of del files. There might be race conditions when open readers of del store files in each partition which can cause NPE. In this patch, we will pre-create the reader for each del store file to avoid this issue. -- This message was sent by Atlassian JIRA (v6.3.4#6332)