[jira] [Commented] (HBASE-19275) TestSnapshotFileCache never worked properly

2019-05-09 Thread Hudson (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-19275?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16836898#comment-16836898
 ] 

Hudson commented on HBASE-19275:


Results for branch branch-1
[build #815 on 
builds.a.o|https://builds.apache.org/job/HBase%20Nightly/job/branch-1/815/]: 
(x) *{color:red}-1 overall{color}*

details (if available):

(x) {color:red}-1 general checks{color}
-- For more information [see general 
report|https://builds.apache.org/job/HBase%20Nightly/job/branch-1/815//General_Nightly_Build_Report/]


(x) {color:red}-1 jdk7 checks{color}
-- Something went wrong running this stage, please [check relevant console 
output|https://builds.apache.org/job/HBase%20Nightly/job/branch-1/815//console].


(x) {color:red}-1 jdk8 hadoop2 checks{color}
-- For more information [see jdk8 (hadoop2) 
report|https://builds.apache.org/job/HBase%20Nightly/job/branch-1/815//JDK8_Nightly_Build_Report_(Hadoop2)/]




(/) {color:green}+1 source release artifact{color}
-- See build output for details.


> TestSnapshotFileCache never worked properly
> ---
>
> Key: HBASE-19275
> URL: https://issues.apache.org/jira/browse/HBASE-19275
> Project: HBase
>  Issue Type: Sub-task
>Affects Versions: 3.0.0, 1.4.0, 1.5.0, 2.0.0
>Reporter: Andrew Purtell
>Assignee: Xu Cang
>Priority: Major
> Fix For: 3.0.0, 2.2.0, 1.4.8, 2.1.1, 2.0.3
>
> Attachments: HBASE-19275-branch-1.patch, 
> HBASE-19275-master.001.patch, HBASE-19275-master.001.patch
>
>
> Error-prone noticed we were asking Iterables.contains() questions with the 
> wrong type in TestSnapshotFileCache. I've attached a fixed version of the 
> test. The results suggest the cache is not evicting entries properly. 
> {noformat}
> java.lang.AssertionError: Cache found 
> 'hdfs://localhost:52867/user/apurtell/test-data/8ce04c85-ce4b-4844-b454-5303482ade95/data/default/snapshot1/9e49edd0ab41657fb0c6ebb4d9dfad15/cf/f132e5b06f66443f8003363ed1535aac',
>  but it shouldn't have.
>   at org.junit.Assert.fail(Assert.java:88)
>   at org.junit.Assert.assertTrue(Assert.java:41)
>   at org.junit.Assert.assertFalse(Assert.java:64)
>   at 
> org.apache.hadoop.hbase.master.snapshot.TestSnapshotFileCache.createAndTestSnapshot(TestSnapshotFileCache.java:260)
>   at 
> org.apache.hadoop.hbase.master.snapshot.TestSnapshotFileCache.createAndTestSnapshotV1(TestSnapshotFileCache.java:206)
>   at 
> org.apache.hadoop.hbase.master.snapshot.TestSnapshotFileCache.testReloadModifiedDirectory(TestSnapshotFileCache.java:102)
> {noformat}
> {noformat}
> java.lang.AssertionError: Cache found 
> 'hdfs://localhost:52867/user/apurtell/test-data/8ce04c85-ce4b-4844-b454-5303482ade95/data/default/snapshot1a/2e81adb9212c98cff970eafa006fc40b/cf/a2ec478d850e4e348359699c53b732c4',
>  but it shouldn't have.
>   at org.junit.Assert.fail(Assert.java:88)
>   at org.junit.Assert.assertTrue(Assert.java:41)
>   at org.junit.Assert.assertFalse(Assert.java:64)
>   at 
> org.apache.hadoop.hbase.master.snapshot.TestSnapshotFileCache.createAndTestSnapshot(TestSnapshotFileCache.java:260)
>   at 
> org.apache.hadoop.hbase.master.snapshot.TestSnapshotFileCache.createAndTestSnapshotV1(TestSnapshotFileCache.java:206)
>   at 
> org.apache.hadoop.hbase.master.snapshot.TestSnapshotFileCache.testLoadAndDelete(TestSnapshotFileCache.java:88)
> {noformat}
> These changes are part of HBASE-19239
> I've disabled the offending test cases with @Ignore in that patch, but they 
> should be reenabled and fixed. 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-20724) Sometimes some compacted storefiles are still opened after region failover

2019-05-09 Thread Hudson (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-20724?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16836894#comment-16836894
 ] 

Hudson commented on HBASE-20724:


Results for branch branch-1
[build #815 on 
builds.a.o|https://builds.apache.org/job/HBase%20Nightly/job/branch-1/815/]: 
(x) *{color:red}-1 overall{color}*

details (if available):

(x) {color:red}-1 general checks{color}
-- For more information [see general 
report|https://builds.apache.org/job/HBase%20Nightly/job/branch-1/815//General_Nightly_Build_Report/]


(x) {color:red}-1 jdk7 checks{color}
-- Something went wrong running this stage, please [check relevant console 
output|https://builds.apache.org/job/HBase%20Nightly/job/branch-1/815//console].


(x) {color:red}-1 jdk8 hadoop2 checks{color}
-- For more information [see jdk8 (hadoop2) 
report|https://builds.apache.org/job/HBase%20Nightly/job/branch-1/815//JDK8_Nightly_Build_Report_(Hadoop2)/]




(/) {color:green}+1 source release artifact{color}
-- See build output for details.


> Sometimes some compacted storefiles are still opened after region failover
> --
>
> Key: HBASE-20724
> URL: https://issues.apache.org/jira/browse/HBASE-20724
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 3.0.0, 1.3.0, 1.4.0, 1.5.0, 2.0.0
>Reporter: Francis Liu
>Assignee: Guanghao Zhang
>Priority: Critical
> Fix For: 3.0.0, 2.2.0, 2.3.0
>
> Attachments: HBASE-20724.branch-2.001.patch, 
> HBASE-20724.master.001.patch, HBASE-20724.master.002.patch, 
> HBASE-20724.master.003.patch, HBASE-20724.master.004.patch, 
> HBASE-20724.master.005.patch, HBASE-20724.master.006.patch, 
> HBASE-20724.master.007.patch, HBASE-20724.master.008.patch, 
> HBASE-20724.master.009.patch, HBASE-20724.master.010.patch, 
> HBASE-20724.master.011.patch, HBASE-20724.master.012.patch, 
> HBASE-20724.master.013.patch, HBASE-20724.master.013.patch, 
> HBASE-20724.master.014.patch
>
>
> It is important that compacted storefiles of a given compaction execution are 
> wholly opened or archived to insure data consistency. ie a storefile 
> containing delete tombstones can be archived while older storefiles 
> containing cells that were supposed to be deleted are left unarchived thereby 
> undeleting those cells.
> When a server fails compaction markers (in the wal edit) are used to 
> determine which storefiles are compacted and should be excluded during region 
> open (during failover). But the WALs containing compaction markers can be 
> prematurely archived even though there are still compacted storefiles for 
> that particular compaction event that hasn't been archived yet. Thus losing 
> compaction information that needs to be replayed in the event of an RS crash. 
> This is because hlog archiving logic only keeps track of flushed storefiles 
> and not compacted ones.
> https://issues.apache.org/jira/browse/HBASE-20704?focusedCommentId=16507680=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-16507680



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-22389) Revert HBASE-19275 TestSnapshotFileCache never worked properly on branch-1

2019-05-09 Thread Hudson (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-22389?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16836897#comment-16836897
 ] 

Hudson commented on HBASE-22389:


Results for branch branch-1
[build #815 on 
builds.a.o|https://builds.apache.org/job/HBase%20Nightly/job/branch-1/815/]: 
(x) *{color:red}-1 overall{color}*

details (if available):

(x) {color:red}-1 general checks{color}
-- For more information [see general 
report|https://builds.apache.org/job/HBase%20Nightly/job/branch-1/815//General_Nightly_Build_Report/]


(x) {color:red}-1 jdk7 checks{color}
-- Something went wrong running this stage, please [check relevant console 
output|https://builds.apache.org/job/HBase%20Nightly/job/branch-1/815//console].


(x) {color:red}-1 jdk8 hadoop2 checks{color}
-- For more information [see jdk8 (hadoop2) 
report|https://builds.apache.org/job/HBase%20Nightly/job/branch-1/815//JDK8_Nightly_Build_Report_(Hadoop2)/]




(/) {color:green}+1 source release artifact{color}
-- See build output for details.


> Revert HBASE-19275 TestSnapshotFileCache never worked properly on branch-1
> --
>
> Key: HBASE-22389
> URL: https://issues.apache.org/jira/browse/HBASE-22389
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 1.5.0
>Reporter: Zach York
>Assignee: Zach York
>Priority: Major
> Fix For: 1.5.0, 1.4.11
>
>
> This commit seems broken on branch-1. 
> https://github.com/apache/hbase/commit/d80d3fa45487fedf203c7144e8267753e4d53529#diff-ae046012dd310fc116a729bf36c7df81
>  seems to have fixed the issue that the other HBASE-19275 patches fixed, but 
> when HBASE-19275 was applied to branch-1 it appears to have reverted this 
> code: 
> https://github.com/apache/hbase/commit/e8e4beacb039299a9e56cd355e3ada08784a7acb#diff-ae046012dd310fc116a729bf36c7df81
>  meaning the tests always pass again.
> [~apurtell] [~xucang]



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-22330) Backport HBASE-20724 (Sometimes some compacted storefiles are still opened after region failover) to branch-1

2019-05-09 Thread Hudson (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-22330?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16836893#comment-16836893
 ] 

Hudson commented on HBASE-22330:


Results for branch branch-1
[build #815 on 
builds.a.o|https://builds.apache.org/job/HBase%20Nightly/job/branch-1/815/]: 
(x) *{color:red}-1 overall{color}*

details (if available):

(x) {color:red}-1 general checks{color}
-- For more information [see general 
report|https://builds.apache.org/job/HBase%20Nightly/job/branch-1/815//General_Nightly_Build_Report/]


(x) {color:red}-1 jdk7 checks{color}
-- Something went wrong running this stage, please [check relevant console 
output|https://builds.apache.org/job/HBase%20Nightly/job/branch-1/815//console].


(x) {color:red}-1 jdk8 hadoop2 checks{color}
-- For more information [see jdk8 (hadoop2) 
report|https://builds.apache.org/job/HBase%20Nightly/job/branch-1/815//JDK8_Nightly_Build_Report_(Hadoop2)/]




(/) {color:green}+1 source release artifact{color}
-- See build output for details.


> Backport HBASE-20724 (Sometimes some compacted storefiles are still opened 
> after region failover) to branch-1
> -
>
> Key: HBASE-22330
> URL: https://issues.apache.org/jira/browse/HBASE-22330
> Project: HBase
>  Issue Type: Sub-task
>  Components: Compaction, regionserver
>Affects Versions: 1.5.0, 1.4.9, 1.3.4
>Reporter: Andrew Purtell
>Assignee: Abhishek Singh Chouhan
>Priority: Major
> Fix For: 1.5.0, 1.3.5, 1.4.11
>
> Attachments: HBASE-22330-addendum.branch-1.patch, 
> HBASE-22330.branch-1.001.patch, HBASE-22330.branch-1.002.patch, 
> HBASE-22330.branch-1.3.001.patch
>
>
> There appears to be a race condition between close and split which when 
> combined with a side effect of HBASE-20704, leads to the parent region store 
> files getting archived and cleared while daughter regions still have 
> references to those parent region store files.
> Here is the timeline of events observed for an affected region:
>  # RS1 faces ZooKeeper connectivity issue for master node and starts shutting 
> itself down. As part of this it starts to close the store and clean up the 
> compacted files (File A)
>  # Master starts bulk assigning regions and assign parent region to RS2
>  # Region opens on RS2 and ends up opening compacted store file(s) (suspect 
> this is due to HBASE-20724)
>  # Now split happens and daughter regions open on RS2 and try to run a 
> compaction as part of post open
>  # Split request at this point is complete. However now archiving proceeds on 
> RS1 and ends up archiving the store file that is referenced by the daughter. 
> Compaction fails due to FileNotFoundException and all subsequent attempts to 
> open the region will fail until manual resolution.
> We think having HBASE-20724 would help in such situations since we won't end 
> up loading compacted store files in the first place. 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-22390) backport HBASE-22190 to branch-1

2019-05-09 Thread Hudson (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-22390?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16836895#comment-16836895
 ] 

Hudson commented on HBASE-22390:


Results for branch branch-1
[build #815 on 
builds.a.o|https://builds.apache.org/job/HBase%20Nightly/job/branch-1/815/]: 
(x) *{color:red}-1 overall{color}*

details (if available):

(x) {color:red}-1 general checks{color}
-- For more information [see general 
report|https://builds.apache.org/job/HBase%20Nightly/job/branch-1/815//General_Nightly_Build_Report/]


(x) {color:red}-1 jdk7 checks{color}
-- Something went wrong running this stage, please [check relevant console 
output|https://builds.apache.org/job/HBase%20Nightly/job/branch-1/815//console].


(x) {color:red}-1 jdk8 hadoop2 checks{color}
-- For more information [see jdk8 (hadoop2) 
report|https://builds.apache.org/job/HBase%20Nightly/job/branch-1/815//JDK8_Nightly_Build_Report_(Hadoop2)/]




(/) {color:green}+1 source release artifact{color}
-- See build output for details.


> backport HBASE-22190 to branch-1
> 
>
> Key: HBASE-22390
> URL: https://issues.apache.org/jira/browse/HBASE-22390
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 1.5.0, 1.4.10
>Reporter: Zach York
>Assignee: Zach York
>Priority: Major
> Fix For: 1.5.0, 1.4.11
>
>
> HBASE-22190 was only applied to branch-2+ while the bug exists in branch-1 as 
> well.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-22190) SnapshotFileCache may fail to load the correct snapshot file list when there is an on-going snapshot operation

2019-05-09 Thread Hudson (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-22190?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16836896#comment-16836896
 ] 

Hudson commented on HBASE-22190:


Results for branch branch-1
[build #815 on 
builds.a.o|https://builds.apache.org/job/HBase%20Nightly/job/branch-1/815/]: 
(x) *{color:red}-1 overall{color}*

details (if available):

(x) {color:red}-1 general checks{color}
-- For more information [see general 
report|https://builds.apache.org/job/HBase%20Nightly/job/branch-1/815//General_Nightly_Build_Report/]


(x) {color:red}-1 jdk7 checks{color}
-- Something went wrong running this stage, please [check relevant console 
output|https://builds.apache.org/job/HBase%20Nightly/job/branch-1/815//console].


(x) {color:red}-1 jdk8 hadoop2 checks{color}
-- For more information [see jdk8 (hadoop2) 
report|https://builds.apache.org/job/HBase%20Nightly/job/branch-1/815//JDK8_Nightly_Build_Report_(Hadoop2)/]




(/) {color:green}+1 source release artifact{color}
-- See build output for details.


> SnapshotFileCache may fail to load the correct snapshot file list when there 
> is an on-going snapshot operation
> --
>
> Key: HBASE-22190
> URL: https://issues.apache.org/jira/browse/HBASE-22190
> Project: HBase
>  Issue Type: Bug
>  Components: snapshots
>Reporter: Duo Zhang
>Assignee: Duo Zhang
>Priority: Blocker
> Fix For: 3.0.0, 2.2.0, 2.3.0, 2.1.5
>
>
> And it seems that it is not only a test issue, we do delete the files under 
> the archive directory, which is incorrect.
> Need to find out why, this maybe a serious bug.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-22389) Revert HBASE-19275 TestSnapshotFileCache never worked properly on branch-1

2019-05-09 Thread Hudson (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-22389?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16836890#comment-16836890
 ] 

Hudson commented on HBASE-22389:


Results for branch branch-1.4
[build #787 on 
builds.a.o|https://builds.apache.org/job/HBase%20Nightly/job/branch-1.4/787/]: 
(x) *{color:red}-1 overall{color}*

details (if available):

(x) {color:red}-1 general checks{color}
-- For more information [see general 
report|https://builds.apache.org/job/HBase%20Nightly/job/branch-1.4/787//General_Nightly_Build_Report/]


(x) {color:red}-1 jdk7 checks{color}
-- For more information [see jdk7 
report|https://builds.apache.org/job/HBase%20Nightly/job/branch-1.4/787//JDK7_Nightly_Build_Report/]


(x) {color:red}-1 jdk8 hadoop2 checks{color}
-- Something went wrong running this stage, please [check relevant console 
output|https://builds.apache.org/job/HBase%20Nightly/job/branch-1.4/787//console].




(/) {color:green}+1 source release artifact{color}
-- See build output for details.


> Revert HBASE-19275 TestSnapshotFileCache never worked properly on branch-1
> --
>
> Key: HBASE-22389
> URL: https://issues.apache.org/jira/browse/HBASE-22389
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 1.5.0
>Reporter: Zach York
>Assignee: Zach York
>Priority: Major
> Fix For: 1.5.0, 1.4.11
>
>
> This commit seems broken on branch-1. 
> https://github.com/apache/hbase/commit/d80d3fa45487fedf203c7144e8267753e4d53529#diff-ae046012dd310fc116a729bf36c7df81
>  seems to have fixed the issue that the other HBASE-19275 patches fixed, but 
> when HBASE-19275 was applied to branch-1 it appears to have reverted this 
> code: 
> https://github.com/apache/hbase/commit/e8e4beacb039299a9e56cd355e3ada08784a7acb#diff-ae046012dd310fc116a729bf36c7df81
>  meaning the tests always pass again.
> [~apurtell] [~xucang]



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-22390) backport HBASE-22190 to branch-1

2019-05-09 Thread Hudson (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-22390?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16836888#comment-16836888
 ] 

Hudson commented on HBASE-22390:


Results for branch branch-1.4
[build #787 on 
builds.a.o|https://builds.apache.org/job/HBase%20Nightly/job/branch-1.4/787/]: 
(x) *{color:red}-1 overall{color}*

details (if available):

(x) {color:red}-1 general checks{color}
-- For more information [see general 
report|https://builds.apache.org/job/HBase%20Nightly/job/branch-1.4/787//General_Nightly_Build_Report/]


(x) {color:red}-1 jdk7 checks{color}
-- For more information [see jdk7 
report|https://builds.apache.org/job/HBase%20Nightly/job/branch-1.4/787//JDK7_Nightly_Build_Report/]


(x) {color:red}-1 jdk8 hadoop2 checks{color}
-- Something went wrong running this stage, please [check relevant console 
output|https://builds.apache.org/job/HBase%20Nightly/job/branch-1.4/787//console].




(/) {color:green}+1 source release artifact{color}
-- See build output for details.


> backport HBASE-22190 to branch-1
> 
>
> Key: HBASE-22390
> URL: https://issues.apache.org/jira/browse/HBASE-22390
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 1.5.0, 1.4.10
>Reporter: Zach York
>Assignee: Zach York
>Priority: Major
> Fix For: 1.5.0, 1.4.11
>
>
> HBASE-22190 was only applied to branch-2+ while the bug exists in branch-1 as 
> well.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-22190) SnapshotFileCache may fail to load the correct snapshot file list when there is an on-going snapshot operation

2019-05-09 Thread Hudson (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-22190?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16836889#comment-16836889
 ] 

Hudson commented on HBASE-22190:


Results for branch branch-1.4
[build #787 on 
builds.a.o|https://builds.apache.org/job/HBase%20Nightly/job/branch-1.4/787/]: 
(x) *{color:red}-1 overall{color}*

details (if available):

(x) {color:red}-1 general checks{color}
-- For more information [see general 
report|https://builds.apache.org/job/HBase%20Nightly/job/branch-1.4/787//General_Nightly_Build_Report/]


(x) {color:red}-1 jdk7 checks{color}
-- For more information [see jdk7 
report|https://builds.apache.org/job/HBase%20Nightly/job/branch-1.4/787//JDK7_Nightly_Build_Report/]


(x) {color:red}-1 jdk8 hadoop2 checks{color}
-- Something went wrong running this stage, please [check relevant console 
output|https://builds.apache.org/job/HBase%20Nightly/job/branch-1.4/787//console].




(/) {color:green}+1 source release artifact{color}
-- See build output for details.


> SnapshotFileCache may fail to load the correct snapshot file list when there 
> is an on-going snapshot operation
> --
>
> Key: HBASE-22190
> URL: https://issues.apache.org/jira/browse/HBASE-22190
> Project: HBase
>  Issue Type: Bug
>  Components: snapshots
>Reporter: Duo Zhang
>Assignee: Duo Zhang
>Priority: Blocker
> Fix For: 3.0.0, 2.2.0, 2.3.0, 2.1.5
>
>
> And it seems that it is not only a test issue, we do delete the files under 
> the archive directory, which is incorrect.
> Need to find out why, this maybe a serious bug.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-20724) Sometimes some compacted storefiles are still opened after region failover

2019-05-09 Thread Hudson (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-20724?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16836887#comment-16836887
 ] 

Hudson commented on HBASE-20724:


Results for branch branch-1.4
[build #787 on 
builds.a.o|https://builds.apache.org/job/HBase%20Nightly/job/branch-1.4/787/]: 
(x) *{color:red}-1 overall{color}*

details (if available):

(x) {color:red}-1 general checks{color}
-- For more information [see general 
report|https://builds.apache.org/job/HBase%20Nightly/job/branch-1.4/787//General_Nightly_Build_Report/]


(x) {color:red}-1 jdk7 checks{color}
-- For more information [see jdk7 
report|https://builds.apache.org/job/HBase%20Nightly/job/branch-1.4/787//JDK7_Nightly_Build_Report/]


(x) {color:red}-1 jdk8 hadoop2 checks{color}
-- Something went wrong running this stage, please [check relevant console 
output|https://builds.apache.org/job/HBase%20Nightly/job/branch-1.4/787//console].




(/) {color:green}+1 source release artifact{color}
-- See build output for details.


> Sometimes some compacted storefiles are still opened after region failover
> --
>
> Key: HBASE-20724
> URL: https://issues.apache.org/jira/browse/HBASE-20724
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 3.0.0, 1.3.0, 1.4.0, 1.5.0, 2.0.0
>Reporter: Francis Liu
>Assignee: Guanghao Zhang
>Priority: Critical
> Fix For: 3.0.0, 2.2.0, 2.3.0
>
> Attachments: HBASE-20724.branch-2.001.patch, 
> HBASE-20724.master.001.patch, HBASE-20724.master.002.patch, 
> HBASE-20724.master.003.patch, HBASE-20724.master.004.patch, 
> HBASE-20724.master.005.patch, HBASE-20724.master.006.patch, 
> HBASE-20724.master.007.patch, HBASE-20724.master.008.patch, 
> HBASE-20724.master.009.patch, HBASE-20724.master.010.patch, 
> HBASE-20724.master.011.patch, HBASE-20724.master.012.patch, 
> HBASE-20724.master.013.patch, HBASE-20724.master.013.patch, 
> HBASE-20724.master.014.patch
>
>
> It is important that compacted storefiles of a given compaction execution are 
> wholly opened or archived to insure data consistency. ie a storefile 
> containing delete tombstones can be archived while older storefiles 
> containing cells that were supposed to be deleted are left unarchived thereby 
> undeleting those cells.
> When a server fails compaction markers (in the wal edit) are used to 
> determine which storefiles are compacted and should be excluded during region 
> open (during failover). But the WALs containing compaction markers can be 
> prematurely archived even though there are still compacted storefiles for 
> that particular compaction event that hasn't been archived yet. Thus losing 
> compaction information that needs to be replayed in the event of an RS crash. 
> This is because hlog archiving logic only keeps track of flushed storefiles 
> and not compacted ones.
> https://issues.apache.org/jira/browse/HBASE-20704?focusedCommentId=16507680=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-16507680



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-19275) TestSnapshotFileCache never worked properly

2019-05-09 Thread Hudson (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-19275?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16836891#comment-16836891
 ] 

Hudson commented on HBASE-19275:


Results for branch branch-1.4
[build #787 on 
builds.a.o|https://builds.apache.org/job/HBase%20Nightly/job/branch-1.4/787/]: 
(x) *{color:red}-1 overall{color}*

details (if available):

(x) {color:red}-1 general checks{color}
-- For more information [see general 
report|https://builds.apache.org/job/HBase%20Nightly/job/branch-1.4/787//General_Nightly_Build_Report/]


(x) {color:red}-1 jdk7 checks{color}
-- For more information [see jdk7 
report|https://builds.apache.org/job/HBase%20Nightly/job/branch-1.4/787//JDK7_Nightly_Build_Report/]


(x) {color:red}-1 jdk8 hadoop2 checks{color}
-- Something went wrong running this stage, please [check relevant console 
output|https://builds.apache.org/job/HBase%20Nightly/job/branch-1.4/787//console].




(/) {color:green}+1 source release artifact{color}
-- See build output for details.


> TestSnapshotFileCache never worked properly
> ---
>
> Key: HBASE-19275
> URL: https://issues.apache.org/jira/browse/HBASE-19275
> Project: HBase
>  Issue Type: Sub-task
>Affects Versions: 3.0.0, 1.4.0, 1.5.0, 2.0.0
>Reporter: Andrew Purtell
>Assignee: Xu Cang
>Priority: Major
> Fix For: 3.0.0, 2.2.0, 1.4.8, 2.1.1, 2.0.3
>
> Attachments: HBASE-19275-branch-1.patch, 
> HBASE-19275-master.001.patch, HBASE-19275-master.001.patch
>
>
> Error-prone noticed we were asking Iterables.contains() questions with the 
> wrong type in TestSnapshotFileCache. I've attached a fixed version of the 
> test. The results suggest the cache is not evicting entries properly. 
> {noformat}
> java.lang.AssertionError: Cache found 
> 'hdfs://localhost:52867/user/apurtell/test-data/8ce04c85-ce4b-4844-b454-5303482ade95/data/default/snapshot1/9e49edd0ab41657fb0c6ebb4d9dfad15/cf/f132e5b06f66443f8003363ed1535aac',
>  but it shouldn't have.
>   at org.junit.Assert.fail(Assert.java:88)
>   at org.junit.Assert.assertTrue(Assert.java:41)
>   at org.junit.Assert.assertFalse(Assert.java:64)
>   at 
> org.apache.hadoop.hbase.master.snapshot.TestSnapshotFileCache.createAndTestSnapshot(TestSnapshotFileCache.java:260)
>   at 
> org.apache.hadoop.hbase.master.snapshot.TestSnapshotFileCache.createAndTestSnapshotV1(TestSnapshotFileCache.java:206)
>   at 
> org.apache.hadoop.hbase.master.snapshot.TestSnapshotFileCache.testReloadModifiedDirectory(TestSnapshotFileCache.java:102)
> {noformat}
> {noformat}
> java.lang.AssertionError: Cache found 
> 'hdfs://localhost:52867/user/apurtell/test-data/8ce04c85-ce4b-4844-b454-5303482ade95/data/default/snapshot1a/2e81adb9212c98cff970eafa006fc40b/cf/a2ec478d850e4e348359699c53b732c4',
>  but it shouldn't have.
>   at org.junit.Assert.fail(Assert.java:88)
>   at org.junit.Assert.assertTrue(Assert.java:41)
>   at org.junit.Assert.assertFalse(Assert.java:64)
>   at 
> org.apache.hadoop.hbase.master.snapshot.TestSnapshotFileCache.createAndTestSnapshot(TestSnapshotFileCache.java:260)
>   at 
> org.apache.hadoop.hbase.master.snapshot.TestSnapshotFileCache.createAndTestSnapshotV1(TestSnapshotFileCache.java:206)
>   at 
> org.apache.hadoop.hbase.master.snapshot.TestSnapshotFileCache.testLoadAndDelete(TestSnapshotFileCache.java:88)
> {noformat}
> These changes are part of HBASE-19239
> I've disabled the offending test cases with @Ignore in that patch, but they 
> should be reenabled and fixed. 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-22330) Backport HBASE-20724 (Sometimes some compacted storefiles are still opened after region failover) to branch-1

2019-05-09 Thread Hudson (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-22330?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16836886#comment-16836886
 ] 

Hudson commented on HBASE-22330:


Results for branch branch-1.4
[build #787 on 
builds.a.o|https://builds.apache.org/job/HBase%20Nightly/job/branch-1.4/787/]: 
(x) *{color:red}-1 overall{color}*

details (if available):

(x) {color:red}-1 general checks{color}
-- For more information [see general 
report|https://builds.apache.org/job/HBase%20Nightly/job/branch-1.4/787//General_Nightly_Build_Report/]


(x) {color:red}-1 jdk7 checks{color}
-- For more information [see jdk7 
report|https://builds.apache.org/job/HBase%20Nightly/job/branch-1.4/787//JDK7_Nightly_Build_Report/]


(x) {color:red}-1 jdk8 hadoop2 checks{color}
-- Something went wrong running this stage, please [check relevant console 
output|https://builds.apache.org/job/HBase%20Nightly/job/branch-1.4/787//console].




(/) {color:green}+1 source release artifact{color}
-- See build output for details.


> Backport HBASE-20724 (Sometimes some compacted storefiles are still opened 
> after region failover) to branch-1
> -
>
> Key: HBASE-22330
> URL: https://issues.apache.org/jira/browse/HBASE-22330
> Project: HBase
>  Issue Type: Sub-task
>  Components: Compaction, regionserver
>Affects Versions: 1.5.0, 1.4.9, 1.3.4
>Reporter: Andrew Purtell
>Assignee: Abhishek Singh Chouhan
>Priority: Major
> Fix For: 1.5.0, 1.3.5, 1.4.11
>
> Attachments: HBASE-22330-addendum.branch-1.patch, 
> HBASE-22330.branch-1.001.patch, HBASE-22330.branch-1.002.patch, 
> HBASE-22330.branch-1.3.001.patch
>
>
> There appears to be a race condition between close and split which when 
> combined with a side effect of HBASE-20704, leads to the parent region store 
> files getting archived and cleared while daughter regions still have 
> references to those parent region store files.
> Here is the timeline of events observed for an affected region:
>  # RS1 faces ZooKeeper connectivity issue for master node and starts shutting 
> itself down. As part of this it starts to close the store and clean up the 
> compacted files (File A)
>  # Master starts bulk assigning regions and assign parent region to RS2
>  # Region opens on RS2 and ends up opening compacted store file(s) (suspect 
> this is due to HBASE-20724)
>  # Now split happens and daughter regions open on RS2 and try to run a 
> compaction as part of post open
>  # Split request at this point is complete. However now archiving proceeds on 
> RS1 and ends up archiving the store file that is referenced by the daughter. 
> Compaction fails due to FileNotFoundException and all subsequent attempts to 
> open the region will fail until manual resolution.
> We think having HBASE-20724 would help in such situations since we won't end 
> up loading compacted store files in the first place. 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-20724) Sometimes some compacted storefiles are still opened after region failover

2019-05-09 Thread Hudson (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-20724?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16836881#comment-16836881
 ] 

Hudson commented on HBASE-20724:


Results for branch branch-1.3
[build #761 on 
builds.a.o|https://builds.apache.org/job/HBase%20Nightly/job/branch-1.3/761/]: 
(x) *{color:red}-1 overall{color}*

details (if available):

(/) {color:green}+1 general checks{color}
-- For more information [see general 
report|https://builds.apache.org/job/HBase%20Nightly/job/branch-1.3/761//General_Nightly_Build_Report/]


(/) {color:green}+1 jdk7 checks{color}
-- For more information [see jdk7 
report|https://builds.apache.org/job/HBase%20Nightly/job/branch-1.3/761//JDK7_Nightly_Build_Report/]


(x) {color:red}-1 jdk8 hadoop2 checks{color}
-- Something went wrong running this stage, please [check relevant console 
output|https://builds.apache.org/job/HBase%20Nightly/job/branch-1.3/761//console].




(/) {color:green}+1 source release artifact{color}
-- See build output for details.


> Sometimes some compacted storefiles are still opened after region failover
> --
>
> Key: HBASE-20724
> URL: https://issues.apache.org/jira/browse/HBASE-20724
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 3.0.0, 1.3.0, 1.4.0, 1.5.0, 2.0.0
>Reporter: Francis Liu
>Assignee: Guanghao Zhang
>Priority: Critical
> Fix For: 3.0.0, 2.2.0, 2.3.0
>
> Attachments: HBASE-20724.branch-2.001.patch, 
> HBASE-20724.master.001.patch, HBASE-20724.master.002.patch, 
> HBASE-20724.master.003.patch, HBASE-20724.master.004.patch, 
> HBASE-20724.master.005.patch, HBASE-20724.master.006.patch, 
> HBASE-20724.master.007.patch, HBASE-20724.master.008.patch, 
> HBASE-20724.master.009.patch, HBASE-20724.master.010.patch, 
> HBASE-20724.master.011.patch, HBASE-20724.master.012.patch, 
> HBASE-20724.master.013.patch, HBASE-20724.master.013.patch, 
> HBASE-20724.master.014.patch
>
>
> It is important that compacted storefiles of a given compaction execution are 
> wholly opened or archived to insure data consistency. ie a storefile 
> containing delete tombstones can be archived while older storefiles 
> containing cells that were supposed to be deleted are left unarchived thereby 
> undeleting those cells.
> When a server fails compaction markers (in the wal edit) are used to 
> determine which storefiles are compacted and should be excluded during region 
> open (during failover). But the WALs containing compaction markers can be 
> prematurely archived even though there are still compacted storefiles for 
> that particular compaction event that hasn't been archived yet. Thus losing 
> compaction information that needs to be replayed in the event of an RS crash. 
> This is because hlog archiving logic only keeps track of flushed storefiles 
> and not compacted ones.
> https://issues.apache.org/jira/browse/HBASE-20704?focusedCommentId=16507680=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-16507680



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-22330) Backport HBASE-20724 (Sometimes some compacted storefiles are still opened after region failover) to branch-1

2019-05-09 Thread Hudson (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-22330?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16836880#comment-16836880
 ] 

Hudson commented on HBASE-22330:


Results for branch branch-1.3
[build #761 on 
builds.a.o|https://builds.apache.org/job/HBase%20Nightly/job/branch-1.3/761/]: 
(x) *{color:red}-1 overall{color}*

details (if available):

(/) {color:green}+1 general checks{color}
-- For more information [see general 
report|https://builds.apache.org/job/HBase%20Nightly/job/branch-1.3/761//General_Nightly_Build_Report/]


(/) {color:green}+1 jdk7 checks{color}
-- For more information [see jdk7 
report|https://builds.apache.org/job/HBase%20Nightly/job/branch-1.3/761//JDK7_Nightly_Build_Report/]


(x) {color:red}-1 jdk8 hadoop2 checks{color}
-- Something went wrong running this stage, please [check relevant console 
output|https://builds.apache.org/job/HBase%20Nightly/job/branch-1.3/761//console].




(/) {color:green}+1 source release artifact{color}
-- See build output for details.


> Backport HBASE-20724 (Sometimes some compacted storefiles are still opened 
> after region failover) to branch-1
> -
>
> Key: HBASE-22330
> URL: https://issues.apache.org/jira/browse/HBASE-22330
> Project: HBase
>  Issue Type: Sub-task
>  Components: Compaction, regionserver
>Affects Versions: 1.5.0, 1.4.9, 1.3.4
>Reporter: Andrew Purtell
>Assignee: Abhishek Singh Chouhan
>Priority: Major
> Fix For: 1.5.0, 1.3.5, 1.4.11
>
> Attachments: HBASE-22330-addendum.branch-1.patch, 
> HBASE-22330.branch-1.001.patch, HBASE-22330.branch-1.002.patch, 
> HBASE-22330.branch-1.3.001.patch
>
>
> There appears to be a race condition between close and split which when 
> combined with a side effect of HBASE-20704, leads to the parent region store 
> files getting archived and cleared while daughter regions still have 
> references to those parent region store files.
> Here is the timeline of events observed for an affected region:
>  # RS1 faces ZooKeeper connectivity issue for master node and starts shutting 
> itself down. As part of this it starts to close the store and clean up the 
> compacted files (File A)
>  # Master starts bulk assigning regions and assign parent region to RS2
>  # Region opens on RS2 and ends up opening compacted store file(s) (suspect 
> this is due to HBASE-20724)
>  # Now split happens and daughter regions open on RS2 and try to run a 
> compaction as part of post open
>  # Split request at this point is complete. However now archiving proceeds on 
> RS1 and ends up archiving the store file that is referenced by the daughter. 
> Compaction fails due to FileNotFoundException and all subsequent attempts to 
> open the region will fail until manual resolution.
> We think having HBASE-20724 would help in such situations since we won't end 
> up loading compacted store files in the first place. 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-22274) Cell size limit check on append should consider cell's previous size.

2019-05-09 Thread Xu Cang (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-22274?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16836874#comment-16836874
 ] 

Xu Cang commented on HBASE-22274:
-

pushed to branch-1.4

> Cell size limit check on append should consider cell's previous size.
> -
>
> Key: HBASE-22274
> URL: https://issues.apache.org/jira/browse/HBASE-22274
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 3.0.0, 2.0.0, 1.3.5
>Reporter: Xu Cang
>Assignee: Xu Cang
>Priority: Minor
> Attachments: HBASE-22274-branch-1.001.patch, 
> HBASE-22274-branch-1.002.patch, HBASE-22274-branch-1.003.patch, 
> HBASE-22274-branch-1.003.patch, HBASE-22274-branch-1.004.patch, 
> HBASE-22274-branch-1.005.patch, HBASE-22274-master.001.patch, 
> HBASE-22274-master.002.patch, HBASE-22274-master.002.patch, 
> HBASE-22274-master.003.patch
>
>
> Now we have cell size limit check based on this parameter 
> *hbase.server.keyvalue.maxsize* 
> One case was missing: appending to a cell only take append op's cell size 
> into account against this limit check. we should check against the potential 
> final cell size after the append.'
> It's easy to reproduce this :
>  
> Apply this diff
>  
> {code:java}
> diff --git 
> a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java
>  
> b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java
>  index 5a285ef6ba..8633177ebe 100644 --- 
> a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java
>  +++ 
> b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java
>  @@ -6455,7 +6455,7 
> - t.append(new Append(ROW).addColumn(FAMILY, QUALIFIER, new byte[10 * 
> 1024])); 
> + t.append(new Append(ROW).addColumn(FAMILY, QUALIFIER, new byte[2 * 1024])); 
> {code}
>  
> Fix is to add this check in #reckonDeltas in HRegion class, where we have 
> already got the appended cell's size. 
> Will throw DoNotRetryIOException if checks is failed.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-22274) Cell size limit check on append should consider cell's previous size.

2019-05-09 Thread Xu Cang (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-22274?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16836868#comment-16836868
 ] 

Xu Cang commented on HBASE-22274:
-

https://issues.apache.org/jira/browse/HBASE-18043 was the one implemented cell 
size limitation.

It wasn't applied to branch-1.3

So I will not push my fix to branch-1.3

(for internal use we can directly apply the patch from branch-1)

> Cell size limit check on append should consider cell's previous size.
> -
>
> Key: HBASE-22274
> URL: https://issues.apache.org/jira/browse/HBASE-22274
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 3.0.0, 2.0.0, 1.3.5
>Reporter: Xu Cang
>Assignee: Xu Cang
>Priority: Minor
> Attachments: HBASE-22274-branch-1.001.patch, 
> HBASE-22274-branch-1.002.patch, HBASE-22274-branch-1.003.patch, 
> HBASE-22274-branch-1.003.patch, HBASE-22274-branch-1.004.patch, 
> HBASE-22274-branch-1.005.patch, HBASE-22274-master.001.patch, 
> HBASE-22274-master.002.patch, HBASE-22274-master.002.patch, 
> HBASE-22274-master.003.patch
>
>
> Now we have cell size limit check based on this parameter 
> *hbase.server.keyvalue.maxsize* 
> One case was missing: appending to a cell only take append op's cell size 
> into account against this limit check. we should check against the potential 
> final cell size after the append.'
> It's easy to reproduce this :
>  
> Apply this diff
>  
> {code:java}
> diff --git 
> a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java
>  
> b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java
>  index 5a285ef6ba..8633177ebe 100644 --- 
> a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java
>  +++ 
> b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java
>  @@ -6455,7 +6455,7 
> - t.append(new Append(ROW).addColumn(FAMILY, QUALIFIER, new byte[10 * 
> 1024])); 
> + t.append(new Append(ROW).addColumn(FAMILY, QUALIFIER, new byte[2 * 1024])); 
> {code}
>  
> Fix is to add this check in #reckonDeltas in HRegion class, where we have 
> already got the appended cell's size. 
> Will throw DoNotRetryIOException if checks is failed.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-22274) Cell size limit check on append should consider cell's previous size.

2019-05-09 Thread Xu Cang (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-22274?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16836860#comment-16836860
 ] 

Xu Cang commented on HBASE-22274:
-

pushed patch to branch-1.

> Cell size limit check on append should consider cell's previous size.
> -
>
> Key: HBASE-22274
> URL: https://issues.apache.org/jira/browse/HBASE-22274
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 3.0.0, 2.0.0, 1.3.5
>Reporter: Xu Cang
>Assignee: Xu Cang
>Priority: Minor
> Attachments: HBASE-22274-branch-1.001.patch, 
> HBASE-22274-branch-1.002.patch, HBASE-22274-branch-1.003.patch, 
> HBASE-22274-branch-1.003.patch, HBASE-22274-branch-1.004.patch, 
> HBASE-22274-branch-1.005.patch, HBASE-22274-master.001.patch, 
> HBASE-22274-master.002.patch, HBASE-22274-master.002.patch, 
> HBASE-22274-master.003.patch
>
>
> Now we have cell size limit check based on this parameter 
> *hbase.server.keyvalue.maxsize* 
> One case was missing: appending to a cell only take append op's cell size 
> into account against this limit check. we should check against the potential 
> final cell size after the append.'
> It's easy to reproduce this :
>  
> Apply this diff
>  
> {code:java}
> diff --git 
> a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java
>  
> b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java
>  index 5a285ef6ba..8633177ebe 100644 --- 
> a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java
>  +++ 
> b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java
>  @@ -6455,7 +6455,7 
> - t.append(new Append(ROW).addColumn(FAMILY, QUALIFIER, new byte[10 * 
> 1024])); 
> + t.append(new Append(ROW).addColumn(FAMILY, QUALIFIER, new byte[2 * 1024])); 
> {code}
>  
> Fix is to add this check in #reckonDeltas in HRegion class, where we have 
> already got the appended cell's size. 
> Will throw DoNotRetryIOException if checks is failed.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-22274) Cell size limit check on append should consider cell's previous size.

2019-05-09 Thread Xu Cang (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-22274?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16836864#comment-16836864
 ] 

Xu Cang commented on HBASE-22274:
-

pushed to master branch

> Cell size limit check on append should consider cell's previous size.
> -
>
> Key: HBASE-22274
> URL: https://issues.apache.org/jira/browse/HBASE-22274
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 3.0.0, 2.0.0, 1.3.5
>Reporter: Xu Cang
>Assignee: Xu Cang
>Priority: Minor
> Attachments: HBASE-22274-branch-1.001.patch, 
> HBASE-22274-branch-1.002.patch, HBASE-22274-branch-1.003.patch, 
> HBASE-22274-branch-1.003.patch, HBASE-22274-branch-1.004.patch, 
> HBASE-22274-branch-1.005.patch, HBASE-22274-master.001.patch, 
> HBASE-22274-master.002.patch, HBASE-22274-master.002.patch, 
> HBASE-22274-master.003.patch
>
>
> Now we have cell size limit check based on this parameter 
> *hbase.server.keyvalue.maxsize* 
> One case was missing: appending to a cell only take append op's cell size 
> into account against this limit check. we should check against the potential 
> final cell size after the append.'
> It's easy to reproduce this :
>  
> Apply this diff
>  
> {code:java}
> diff --git 
> a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java
>  
> b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java
>  index 5a285ef6ba..8633177ebe 100644 --- 
> a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java
>  +++ 
> b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java
>  @@ -6455,7 +6455,7 
> - t.append(new Append(ROW).addColumn(FAMILY, QUALIFIER, new byte[10 * 
> 1024])); 
> + t.append(new Append(ROW).addColumn(FAMILY, QUALIFIER, new byte[2 * 1024])); 
> {code}
>  
> Fix is to add this check in #reckonDeltas in HRegion class, where we have 
> already got the appended cell's size. 
> Will throw DoNotRetryIOException if checks is failed.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-20724) Sometimes some compacted storefiles are still opened after region failover

2019-05-09 Thread Hudson (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-20724?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16836850#comment-16836850
 ] 

Hudson commented on HBASE-20724:


SUCCESS: Integrated in Jenkins build HBase-1.3-IT #554 (See 
[https://builds.apache.org/job/HBase-1.3-IT/554/])
Addendum HBASE-22330 Backport HBASE-20724 (Sometimes some compacted 
(abhishek.chouhan: 
[https://github.com/apache/hbase/commit/10f00547627076d79d77cf58dd2deaece2287084])
* (edit) 
hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestCleanupCompactedFileAfterFailover.java


> Sometimes some compacted storefiles are still opened after region failover
> --
>
> Key: HBASE-20724
> URL: https://issues.apache.org/jira/browse/HBASE-20724
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 3.0.0, 1.3.0, 1.4.0, 1.5.0, 2.0.0
>Reporter: Francis Liu
>Assignee: Guanghao Zhang
>Priority: Critical
> Fix For: 3.0.0, 2.2.0, 2.3.0
>
> Attachments: HBASE-20724.branch-2.001.patch, 
> HBASE-20724.master.001.patch, HBASE-20724.master.002.patch, 
> HBASE-20724.master.003.patch, HBASE-20724.master.004.patch, 
> HBASE-20724.master.005.patch, HBASE-20724.master.006.patch, 
> HBASE-20724.master.007.patch, HBASE-20724.master.008.patch, 
> HBASE-20724.master.009.patch, HBASE-20724.master.010.patch, 
> HBASE-20724.master.011.patch, HBASE-20724.master.012.patch, 
> HBASE-20724.master.013.patch, HBASE-20724.master.013.patch, 
> HBASE-20724.master.014.patch
>
>
> It is important that compacted storefiles of a given compaction execution are 
> wholly opened or archived to insure data consistency. ie a storefile 
> containing delete tombstones can be archived while older storefiles 
> containing cells that were supposed to be deleted are left unarchived thereby 
> undeleting those cells.
> When a server fails compaction markers (in the wal edit) are used to 
> determine which storefiles are compacted and should be excluded during region 
> open (during failover). But the WALs containing compaction markers can be 
> prematurely archived even though there are still compacted storefiles for 
> that particular compaction event that hasn't been archived yet. Thus losing 
> compaction information that needs to be replayed in the event of an RS crash. 
> This is because hlog archiving logic only keeps track of flushed storefiles 
> and not compacted ones.
> https://issues.apache.org/jira/browse/HBASE-20704?focusedCommentId=16507680=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-16507680



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-22330) Backport HBASE-20724 (Sometimes some compacted storefiles are still opened after region failover) to branch-1

2019-05-09 Thread Hudson (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-22330?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16836849#comment-16836849
 ] 

Hudson commented on HBASE-22330:


SUCCESS: Integrated in Jenkins build HBase-1.3-IT #554 (See 
[https://builds.apache.org/job/HBase-1.3-IT/554/])
Addendum HBASE-22330 Backport HBASE-20724 (Sometimes some compacted 
(abhishek.chouhan: 
[https://github.com/apache/hbase/commit/10f00547627076d79d77cf58dd2deaece2287084])
* (edit) 
hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestCleanupCompactedFileAfterFailover.java


> Backport HBASE-20724 (Sometimes some compacted storefiles are still opened 
> after region failover) to branch-1
> -
>
> Key: HBASE-22330
> URL: https://issues.apache.org/jira/browse/HBASE-22330
> Project: HBase
>  Issue Type: Sub-task
>  Components: Compaction, regionserver
>Affects Versions: 1.5.0, 1.4.9, 1.3.4
>Reporter: Andrew Purtell
>Assignee: Abhishek Singh Chouhan
>Priority: Major
> Fix For: 1.5.0, 1.3.5, 1.4.11
>
> Attachments: HBASE-22330-addendum.branch-1.patch, 
> HBASE-22330.branch-1.001.patch, HBASE-22330.branch-1.002.patch, 
> HBASE-22330.branch-1.3.001.patch
>
>
> There appears to be a race condition between close and split which when 
> combined with a side effect of HBASE-20704, leads to the parent region store 
> files getting archived and cleared while daughter regions still have 
> references to those parent region store files.
> Here is the timeline of events observed for an affected region:
>  # RS1 faces ZooKeeper connectivity issue for master node and starts shutting 
> itself down. As part of this it starts to close the store and clean up the 
> compacted files (File A)
>  # Master starts bulk assigning regions and assign parent region to RS2
>  # Region opens on RS2 and ends up opening compacted store file(s) (suspect 
> this is due to HBASE-20724)
>  # Now split happens and daughter regions open on RS2 and try to run a 
> compaction as part of post open
>  # Split request at this point is complete. However now archiving proceeds on 
> RS1 and ends up archiving the store file that is referenced by the daughter. 
> Compaction fails due to FileNotFoundException and all subsequent attempts to 
> open the region will fail until manual resolution.
> We think having HBASE-20724 would help in such situations since we won't end 
> up loading compacted store files in the first place. 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-22391) Fix flaky tests from TestFromClientSide

2019-05-09 Thread HBase QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-22391?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16836838#comment-16836838
 ] 

HBase QA commented on HBASE-22391:
--

| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  0m 
49s{color} | {color:blue} Docker mode activated. {color} |
|| || || || {color:brown} Prechecks {color} ||
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m  
1s{color} | {color:blue} Findbugs executables are not available. {color} |
| {color:green}+1{color} | {color:green} hbaseanti {color} | {color:green}  0m  
0s{color} | {color:green} Patch does not have any anti-patterns. {color} |
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:green}+1{color} | {color:green} test4tests {color} | {color:green}  0m 
 0s{color} | {color:green} The patch appears to include 1 new or modified test 
files. {color} |
|| || || || {color:brown} branch-1 Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  8m 
19s{color} | {color:green} branch-1 passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
40s{color} | {color:green} branch-1 passed with JDK v1.8.0_212 {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
41s{color} | {color:green} branch-1 passed with JDK v1.7.0_222 {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  1m 
27s{color} | {color:green} branch-1 passed {color} |
| {color:green}+1{color} | {color:green} shadedjars {color} | {color:green}  2m 
53s{color} | {color:green} branch has no errors when building our shaded 
downstream artifacts. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
36s{color} | {color:green} branch-1 passed with JDK v1.8.0_212 {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
38s{color} | {color:green} branch-1 passed with JDK v1.7.0_222 {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  1m 
42s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
39s{color} | {color:green} the patch passed with JDK v1.8.0_212 {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  0m 
39s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
41s{color} | {color:green} the patch passed with JDK v1.7.0_222 {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  0m 
41s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} checkstyle {color} | {color:red}  1m 
24s{color} | {color:red} hbase-server: The patch generated 2 new + 221 
unchanged - 1 fixed = 223 total (was 222) {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} shadedjars {color} | {color:green}  2m 
56s{color} | {color:green} patch has no errors when building our shaded 
downstream artifacts. {color} |
| {color:green}+1{color} | {color:green} hadoopcheck {color} | {color:green}  
1m 43s{color} | {color:green} Patch does not cause any errors with Hadoop 
2.7.4. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
28s{color} | {color:green} the patch passed with JDK v1.8.0_212 {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
37s{color} | {color:green} the patch passed with JDK v1.7.0_222 {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} unit {color} | {color:green}130m 
39s{color} | {color:green} hbase-server in the patch passed. {color} |
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
24s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black}157m 48s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Docker | Client=17.05.0-ce Server=17.05.0-ce base: 
https://builds.apache.org/job/PreCommit-HBASE-Build/281/artifact/patchprocess/Dockerfile
 |
| JIRA Issue | HBASE-22391 |
| JIRA Patch URL | 
https://issues.apache.org/jira/secure/attachment/12968338/HBASE-22391-branch-1.001.patch
 |
| Optional Tests |  dupname  asflicense  javac  javadoc  unit  findbugs  
shadedjars  hadoopcheck  hbaseanti  checkstyle  compile  |
| uname | Linux 8e8ba9a61157 

[jira] [Commented] (HBASE-22274) Cell size limit check on append should consider cell's previous size.

2019-05-09 Thread HBase QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-22274?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16836835#comment-16836835
 ] 

HBase QA commented on HBASE-22274:
--

| (/) *{color:green}+1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  0m 
51s{color} | {color:blue} Docker mode activated. {color} |
|| || || || {color:brown} Prechecks {color} ||
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m  
0s{color} | {color:blue} Findbugs executables are not available. {color} |
| {color:green}+1{color} | {color:green} hbaseanti {color} | {color:green}  0m  
0s{color} | {color:green} Patch does not have any anti-patterns. {color} |
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:green}+1{color} | {color:green} test4tests {color} | {color:green}  0m 
 0s{color} | {color:green} The patch appears to include 1 new or modified test 
files. {color} |
|| || || || {color:brown} branch-1 Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  1m 
48s{color} | {color:green} branch-1 passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
38s{color} | {color:green} branch-1 passed with JDK v1.8.0_212 {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
41s{color} | {color:green} branch-1 passed with JDK v1.7.0_222 {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  1m 
29s{color} | {color:green} branch-1 passed {color} |
| {color:green}+1{color} | {color:green} shadedjars {color} | {color:green}  2m 
47s{color} | {color:green} branch has no errors when building our shaded 
downstream artifacts. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
28s{color} | {color:green} branch-1 passed with JDK v1.8.0_212 {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
37s{color} | {color:green} branch-1 passed with JDK v1.7.0_222 {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  1m 
41s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
38s{color} | {color:green} the patch passed with JDK v1.8.0_212 {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  0m 
38s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
40s{color} | {color:green} the patch passed with JDK v1.7.0_222 {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  0m 
40s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  1m 
27s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} shadedjars {color} | {color:green}  2m 
45s{color} | {color:green} patch has no errors when building our shaded 
downstream artifacts. {color} |
| {color:green}+1{color} | {color:green} hadoopcheck {color} | {color:green}  
1m 42s{color} | {color:green} Patch does not cause any errors with Hadoop 
2.7.4. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
29s{color} | {color:green} the patch passed with JDK v1.8.0_212 {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
38s{color} | {color:green} the patch passed with JDK v1.7.0_222 {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} unit {color} | {color:green}129m 
45s{color} | {color:green} hbase-server in the patch passed. {color} |
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
21s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black}150m  1s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Docker | Client=17.05.0-ce Server=17.05.0-ce base: 
https://builds.apache.org/job/PreCommit-HBASE-Build/280/artifact/patchprocess/Dockerfile
 |
| JIRA Issue | HBASE-22274 |
| JIRA Patch URL | 
https://issues.apache.org/jira/secure/attachment/12968337/HBASE-22274-branch-1.005.patch
 |
| Optional Tests |  dupname  asflicense  javac  javadoc  unit  findbugs  
shadedjars  hadoopcheck  hbaseanti  checkstyle  compile  |
| uname | Linux 8abef0aa581d 4.4.0-143-generic #169~14.04.2-Ubuntu SMP Wed Feb 
13 15:00:41 UTC 

[jira] [Commented] (HBASE-18373) fstat unimplemented error on Rasbian Jessie

2019-05-09 Thread Toshihiro Suzuki (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-18373?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16836830#comment-16836830
 ] 

Toshihiro Suzuki commented on HBASE-18373:
--

Looks this issue is not HBase side problem, but an issue with JRuby and 
Oracle's JVM. 

See the following for details:
https://github.com/jruby/jruby/issues/2913
https://github.com/elastic/logstash/issues/3127#issuecomment-101068714
https://github.com/elastic/logstash/issues/2962

Closing as not a problem.

> fstat unimplemented error on Rasbian Jessie
> ---
>
> Key: HBASE-18373
> URL: https://issues.apache.org/jira/browse/HBASE-18373
> Project: HBase
>  Issue Type: Bug
>  Components: shell
>Affects Versions: 2.0.0-alpha-1
> Environment: Raspberry Pi 3
> Rasbian Jessie (latest version)
>Reporter: Nathan Green
>Priority: Blocker
>
> I have a Hbase distributed cluster setup on my Raspberry PI cluster (1 x 
> Namenode & 4 x Datanode)
> The stable version of Hbase shell works fine... unfortunately I had hadoop 
> issues (mismatch of jar files gr) so I have installed latest Hbase & 
> installed matching version of Hadoop (2.7.1).
> HBase itself is up and running and all good.  Hbase shell however gives this 
> error:
> pi@pidoop1:/opt/hbase/lib/ruby/irb $ hbase shell
> 2017-07-13 12:25:11,722 WARN  [main] util.NativeCodeLoader: Unable to load 
> native-hadoop library for your platform... using builtin-java classes where 
> applicable
> SLF4J: Class path contains multiple SLF4J bindings.
> SLF4J: Found binding in 
> [jar:file:/opt/hbase/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in 
> [jar:file:/opt/hadoop/share/hadoop/common/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an 
> explanation.
> SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
> HBase Shell
> Use "help" to get list of supported commands.
> Use "exit" to quit this interactive shell.
> Version 2.0.0-alpha-1, rc830a0f47f58d4892dd3300032c8244d6278aecc, Wed Jun  7 
> 22:05:17 PDT 2017
> Took 0.0300 seconds
> NotImplementedError: fstat unimplemented unsupported or native support failed 
> to load; see http://wiki.jruby.org/Native-Libraries
>   initialize at org/jruby/RubyIO.java:1013
> open at org/jruby/RubyIO.java:1154
>   initialize at 
> uri:classloader:/META-INF/jruby.home/lib/ruby/stdlib/irb/input-method.rb:141
>   initialize at 
> uri:classloader:/META-INF/jruby.home/lib/ruby/stdlib/irb/context.rb:70
>   initialize at 
> uri:classloader:/META-INF/jruby.home/lib/ruby/stdlib/irb.rb:426
>   initialize at /opt/hbase/lib/ruby/irb/hirb.rb:43
>start at /opt/hbase/bin/hirb.rb:194
>at /opt/hbase/bin/hirb.rb:206



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] [hbase] infraio commented on a change in pull request #221: HBASE-22365 Region may be opened on two RegionServers

2019-05-09 Thread GitBox
infraio commented on a change in pull request #221: HBASE-22365 Region may be 
opened on two RegionServers
URL: https://github.com/apache/hbase/pull/221#discussion_r282722269
 
 

 ##
 File path: 
hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/AssignmentManager.java
 ##
 @@ -1575,18 +1562,36 @@ void regionClosing(RegionStateNode regionNode) throws 
IOException {
 metrics.incrementOperationCounter();
   }
 
-  // should be called within the synchronized block of RegionStateNode
-  // The parameter 'normally' means whether we are closed cleanly, if it is 
true, then it means that
-  // we are closed due to a RS crash.
-  void regionClosed(RegionStateNode regionNode, boolean normally) throws 
IOException {
-RegionState.State state = regionNode.getState();
+  // for open and close, they will first be persist to the procedure store in
+  // RegionRemoteProcedureBase. So here we will first change the in memory 
state as it is considered
+  // as succeeded if the persistence to procedure store is succeeded, and then 
when the
+  // RegionRemoteProcedureBase is woken up, we will persist the 
RegionStateNode to hbase:meta.
+
+  // should be called under the RegionStateNode lock
+  void regionOpenedNoPersistence(RegionStateNode regionNode) throws 
IOException {
+regionNode.transitionState(State.OPEN, STATES_EXPECTED_ON_OPEN);
+RegionInfo regionInfo = regionNode.getRegionInfo();
+regionStates.addRegionToServer(regionNode);
+regionStates.removeFromFailedOpen(regionInfo);
+  }
+
+  // should be called under the RegionStateNode lock
+  void regionClosedNoPersistence(RegionStateNode regionNode) throws 
IOException {
 
 Review comment:
   ditto.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [hbase] infraio commented on a change in pull request #221: HBASE-22365 Region may be opened on two RegionServers

2019-05-09 Thread GitBox
infraio commented on a change in pull request #221: HBASE-22365 Region may be 
opened on two RegionServers
URL: https://github.com/apache/hbase/pull/221#discussion_r282724012
 
 

 ##
 File path: 
hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/RegionRemoteProcedureBase.java
 ##
 @@ -175,7 +176,7 @@ void reportTransition(MasterProcedureEnv env, 
RegionStateNode regionNode, Server
   throw new UnexpectedStateException("Received report from " + serverName 
+ ", expected " +
 targetServer + ", " + regionNode + ", proc=" + this);
 }
-reportTransition(regionNode, transitionCode, seqId);
+reportTransition(env, regionNode, transitionCode, seqId);
 // this state means we have received the report from RS, does not mean the 
result is fine, as we
 
 Review comment:
   Here we will first change the memory state and then persist to procedure 
store. If persist to procedure failed, need to change the memory state back?


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Resolved] (HBASE-18373) fstat unimplemented error on Rasbian Jessie

2019-05-09 Thread Toshihiro Suzuki (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-18373?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Toshihiro Suzuki resolved HBASE-18373.
--
Resolution: Not A Problem

> fstat unimplemented error on Rasbian Jessie
> ---
>
> Key: HBASE-18373
> URL: https://issues.apache.org/jira/browse/HBASE-18373
> Project: HBase
>  Issue Type: Bug
>  Components: shell
>Affects Versions: 2.0.0-alpha-1
> Environment: Raspberry Pi 3
> Rasbian Jessie (latest version)
>Reporter: Nathan Green
>Priority: Blocker
>
> I have a Hbase distributed cluster setup on my Raspberry PI cluster (1 x 
> Namenode & 4 x Datanode)
> The stable version of Hbase shell works fine... unfortunately I had hadoop 
> issues (mismatch of jar files gr) so I have installed latest Hbase & 
> installed matching version of Hadoop (2.7.1).
> HBase itself is up and running and all good.  Hbase shell however gives this 
> error:
> pi@pidoop1:/opt/hbase/lib/ruby/irb $ hbase shell
> 2017-07-13 12:25:11,722 WARN  [main] util.NativeCodeLoader: Unable to load 
> native-hadoop library for your platform... using builtin-java classes where 
> applicable
> SLF4J: Class path contains multiple SLF4J bindings.
> SLF4J: Found binding in 
> [jar:file:/opt/hbase/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in 
> [jar:file:/opt/hadoop/share/hadoop/common/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an 
> explanation.
> SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
> HBase Shell
> Use "help" to get list of supported commands.
> Use "exit" to quit this interactive shell.
> Version 2.0.0-alpha-1, rc830a0f47f58d4892dd3300032c8244d6278aecc, Wed Jun  7 
> 22:05:17 PDT 2017
> Took 0.0300 seconds
> NotImplementedError: fstat unimplemented unsupported or native support failed 
> to load; see http://wiki.jruby.org/Native-Libraries
>   initialize at org/jruby/RubyIO.java:1013
> open at org/jruby/RubyIO.java:1154
>   initialize at 
> uri:classloader:/META-INF/jruby.home/lib/ruby/stdlib/irb/input-method.rb:141
>   initialize at 
> uri:classloader:/META-INF/jruby.home/lib/ruby/stdlib/irb/context.rb:70
>   initialize at 
> uri:classloader:/META-INF/jruby.home/lib/ruby/stdlib/irb.rb:426
>   initialize at /opt/hbase/lib/ruby/irb/hirb.rb:43
>start at /opt/hbase/bin/hirb.rb:194
>at /opt/hbase/bin/hirb.rb:206



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] [hbase] infraio commented on a change in pull request #221: HBASE-22365 Region may be opened on two RegionServers

2019-05-09 Thread GitBox
infraio commented on a change in pull request #221: HBASE-22365 Region may be 
opened on two RegionServers
URL: https://github.com/apache/hbase/pull/221#discussion_r282722122
 
 

 ##
 File path: 
hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/AssignmentManager.java
 ##
 @@ -1575,18 +1562,36 @@ void regionClosing(RegionStateNode regionNode) throws 
IOException {
 metrics.incrementOperationCounter();
   }
 
-  // should be called within the synchronized block of RegionStateNode
-  // The parameter 'normally' means whether we are closed cleanly, if it is 
true, then it means that
-  // we are closed due to a RS crash.
-  void regionClosed(RegionStateNode regionNode, boolean normally) throws 
IOException {
-RegionState.State state = regionNode.getState();
+  // for open and close, they will first be persist to the procedure store in
+  // RegionRemoteProcedureBase. So here we will first change the in memory 
state as it is considered
+  // as succeeded if the persistence to procedure store is succeeded, and then 
when the
+  // RegionRemoteProcedureBase is woken up, we will persist the 
RegionStateNode to hbase:meta.
+
+  // should be called under the RegionStateNode lock
+  void regionOpenedNoPersistence(RegionStateNode regionNode) throws 
IOException {
 
 Review comment:
   regionOpenedWithoutPersistToMeta or regionOpenedNoPersistToMeta? It will be 
better to point "persist to meta".


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (HBASE-21991) Fix MetaMetrics issues - [Race condition, Faulty remove logic], few improvements

2019-05-09 Thread Sean Busbey (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21991?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16836826#comment-16836826
 ] 

Sean Busbey commented on HBASE-21991:
-

addendum looks good to me.

I posted to dev@ here:

https://s.apache.org/oQVF

If someone has a pressing need for this in a branch for an RC please let me 
know. Otherwise I'll wait a day or so for feedback on the dev posting and then 
take care of pushing the addendum and release noting things.

> Fix MetaMetrics issues - [Race condition, Faulty remove logic], few 
> improvements
> 
>
> Key: HBASE-21991
> URL: https://issues.apache.org/jira/browse/HBASE-21991
> Project: HBase
>  Issue Type: Bug
>  Components: Coprocessors, metrics
>Reporter: Sakthi
>Assignee: Sakthi
>Priority: Major
> Fix For: 3.0.0, 1.5.0, 2.2.0, 1.4.10, 2.3.0
>
> Attachments: hbase-21991.addendum.patch, 
> hbase-21991.branch-1.001.patch, hbase-21991.branch-1.002.patch, 
> hbase-21991.master.001.patch, hbase-21991.master.002.patch, 
> hbase-21991.master.003.patch, hbase-21991.master.004.patch, 
> hbase-21991.master.005.patch, hbase-21991.master.006.patch
>
>
> Here is a list of the issues related to the MetaMetrics implementation:
> +*Bugs*+:
>  # [_Lossy counting for top-k_] *Faulty remove logic of non-eligible meters*: 
> Under certain conditions, we might end up storing/exposing all the meters 
> rather than top-k-ish
>  # MetaMetrics can throw NPE resulting in aborting of the RS because of a 
> *Race Condition*.
> +*Improvements*+:
>  # With high number of regions in the cluster, exposure of metrics for each 
> region blows up the JMX from ~140 Kbs to 100+ Mbs depending on the number of 
> regions. It's better to use *lossy counting to maintain top-k for region 
> metrics* as well.
>  # As the lossy meters do not represent actual counts, I think, it'll be 
> better to *rename the meters to include "lossy" in the name*. It would be 
> more informative while monitoring the metrics and there would be less 
> confusion regarding actual counts to lossy counts.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] [hbase] Apache-HBase commented on issue #120: HBASE-20494 Updated the version of metrics-core to 3.2.6

2019-05-09 Thread GitBox
Apache-HBase commented on issue #120: HBASE-20494 Updated the version of 
metrics-core to 3.2.6
URL: https://github.com/apache/hbase/pull/120#issuecomment-491126822
 
 
   :confetti_ball: **+1 overall**
   
   
   
   
   
   
   | Vote | Subsystem | Runtime | Comment |
   |::|--:|:|:|
   | 0 | reexec | 36 | Docker mode activated. |
   ||| _ Prechecks _ |
   | +1 | @author | 0 | The patch does not contain any @author tags. |
   | -0 | test4tests | 0 | The patch doesn't appear to include any new or 
modified tests.  Please justify why no new tests are needed for this patch. 
Also please list what manual steps were performed to verify this patch. |
   ||| _ master Compile Tests _ |
   | +1 | mvninstall | 254 | master passed |
   | +1 | compile | 159 | master passed |
   | +1 | shadedjars | 258 | branch has no errors when building our shaded 
downstream artifacts. |
   | +1 | javadoc | 157 | master passed |
   ||| _ Patch Compile Tests _ |
   | +1 | mvninstall | 237 | the patch passed |
   | +1 | compile | 159 | the patch passed |
   | +1 | javac | 159 | the patch passed |
   | +1 | whitespace | 0 | The patch has no whitespace issues. |
   | +1 | xml | 1 | The patch has no ill-formed XML file. |
   | +1 | shadedjars | 258 | patch has no errors when building our shaded 
downstream artifacts. |
   | +1 | hadoopcheck | 505 | Patch does not cause any errors with Hadoop 2.7.4 
or 3.0.0. |
   | +1 | javadoc | 157 | the patch passed |
   ||| _ Other Tests _ |
   | +1 | unit | 18973 | root in the patch passed. |
   | +1 | asflicense | 39 | The patch does not generate ASF License warnings. |
   | | | 21260 | |
   
   
   | Subsystem | Report/Notes |
   |--:|:-|
   | Docker | Client=17.05.0-ce Server=17.05.0-ce base: 
https://builds.apache.org/job/HBase-PreCommit-GitHub-PR/job/PR-120/13/artifact/out/Dockerfile
 |
   | GITHUB PR | https://github.com/apache/hbase/pull/120 |
   | Optional Tests |  dupname  asflicense  javac  javadoc  unit  shadedjars  
hadoopcheck  xml  compile  |
   | uname | Linux 438ffa5e40b8 3.13.0-153-generic #203-Ubuntu SMP Thu Jun 14 
08:52:28 UTC 2018 x86_64 GNU/Linux |
   | Build tool | maven |
   | Personality | /testptch/patchprocess/precommit/personality/provided.sh |
   | git revision | master / 4d64dd2e82 |
   | maven | version: Apache Maven 3.5.4 
(1edded0938998edf8bf061f1ceb3cfdeccf443fe; 2018-06-17T18:33:14Z) |
   | Default Java | 1.8.0_181 |
   |  Test Results | 
https://builds.apache.org/job/HBase-PreCommit-GitHub-PR/job/PR-120/13/testReport/
 |
   | Max. process+thread count | 4994 (vs. ulimit of 1) |
   | modules | C: . U: . |
   | Console output | 
https://builds.apache.org/job/HBase-PreCommit-GitHub-PR/job/PR-120/13/console |
   | Powered by | Apache Yetus 0.9.0 http://yetus.apache.org |
   
   
   This message was automatically generated.
   
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (HBASE-22365) Region may be opened on two RegionServers

2019-05-09 Thread Guanghao Zhang (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-22365?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16836822#comment-16836822
 ] 

Guanghao Zhang commented on HBASE-22365:


Try the patch on our internal branch and the ITBLL which have 3 billions rows 
passed. :)

> Region may be opened on two RegionServers
> -
>
> Key: HBASE-22365
> URL: https://issues.apache.org/jira/browse/HBASE-22365
> Project: HBase
>  Issue Type: Improvement
>Affects Versions: 3.0.0, 2.2.0, 2.3.0
>Reporter: Guanghao Zhang
>Assignee: Duo Zhang
>Priority: Blocker
> Attachments: HBASE-22365-UT.patch
>
>
> Found this problem when run ITBLL with our internal branch which is based on 
> branch-2.2. So mark this as a blocker for 2.2.0. A region 
> 7ebdca9cd09e26074749b546586e2156 is moved from RS-st99 to RS-st98 and the 
> TRSP succeed. Meanwhile, RS-st99 crashed and schedule a new SCP for RS-st99. 
> So SCP initialized subprocedures for 7ebdca9cd09e26074749b546586e2156, too. 
> Then the 7ebdca9cd09e26074749b546586e2156 was assigned to two RegionServers.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-21536) Fix completebulkload usage instructions

2019-05-09 Thread HBase QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21536?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16836821#comment-16836821
 ] 

HBase QA commented on HBASE-21536:
--

| (/) *{color:green}+1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  0m 
25s{color} | {color:blue} Docker mode activated. {color} |
|| || || || {color:brown} Prechecks {color} ||
| {color:blue}0{color} | {color:blue} shelldocs {color} | {color:blue}  0m  
1s{color} | {color:blue} Shelldocs was not available. {color} |
| {color:green}+1{color} | {color:green} hbaseanti {color} | {color:green}  0m  
0s{color} | {color:green} Patch does not have any anti-patterns. {color} |
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:orange}-0{color} | {color:orange} test4tests {color} | {color:orange}  
0m  0s{color} | {color:orange} The patch doesn't appear to include any new or 
modified tests. Please justify why no new tests are needed for this patch. Also 
please list what manual steps were performed to verify this patch. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
23s{color} | {color:blue} Maven dependency ordering for branch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  3m 
55s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  2m 
36s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  2m 
 9s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} refguide {color} | {color:blue} 11m 
37s{color} | {color:blue} branch has no errors when building the reference 
guide. See footer for rendered docs, which you should manually inspect. {color} 
|
| {color:green}+1{color} | {color:green} shadedjars {color} | {color:green}  4m 
18s{color} | {color:green} branch has no errors when building our shaded 
downstream artifacts. {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m  
0s{color} | {color:blue} Skipped patched modules with no Java source: . {color} 
|
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  3m 
13s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  3m 
10s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
13s{color} | {color:blue} Maven dependency ordering for patch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  3m 
54s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  2m 
38s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  2m 
38s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  2m 
13s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} shellcheck {color} | {color:green}  0m 
 2s{color} | {color:green} There were no new shellcheck issues. {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:blue}0{color} | {color:blue} refguide {color} | {color:blue}  7m 
22s{color} | {color:blue} patch has no errors when building the reference 
guide. See footer for rendered docs, which you should manually inspect. {color} 
|
| {color:green}+1{color} | {color:green} shadedjars {color} | {color:green}  4m 
20s{color} | {color:green} patch has no errors when building our shaded 
downstream artifacts. {color} |
| {color:green}+1{color} | {color:green} hadoopcheck {color} | {color:green}  
8m  3s{color} | {color:green} Patch does not cause any errors with Hadoop 2.7.4 
or 3.0.0. {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m  
0s{color} | {color:blue} Skipped patched modules with no Java source: . {color} 
|
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  3m 
42s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  3m 
10s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} unit {color} | {color:green}210m 
15s{color} | {color:green} root in the patch passed. {color} |
| {color:green}+1{color} | 

[jira] [Commented] (HBASE-22254) refactor and improve decommissioning logic

2019-05-09 Thread Sergey Shelukhin (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-22254?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16836817#comment-16836817
 ] 

Sergey Shelukhin commented on HBASE-22254:
--

Btw, these APIs were added in HBASE-17370 but the book wasn't updated, it still 
relies on znode creation there... I wonder if the book should be updated w/this 
patch

> refactor and improve decommissioning logic
> --
>
> Key: HBASE-22254
> URL: https://issues.apache.org/jira/browse/HBASE-22254
> Project: HBase
>  Issue Type: Improvement
>Affects Versions: 3.0.0
>Reporter: Sergey Shelukhin
>Assignee: Sergey Shelukhin
>Priority: Major
> Attachments: HBASE-22254.01.patch, HBASE-22254.02.patch, 
> HBASE-22254.patch
>
>
> Making some changes needed to support better decommissioning on large 
> clusters and with container mode; to test those and add clarify I moved parts 
> of decommissioning logic from HMaster, Draining tracker, and ServerManager 
> into a separate class.
> Features added/improvements:
> 1) More resilient off-loading; right now off-loading fails for a subset of 
> regions in case of a single region failure; is never done on master restart, 
> etc.
> 2) Option to kill RS after off-loading (good for container mode HBase, e.g. 
> on YARN).
> 3) Option to specify machine names only to decommission, for the API to be 
> usable for an external system that doesn't care about HBase server names, or 
> e.g. multiple RS in containers on the same node.
> 4) Option to replace existing decommissioning list instead of adding to it 
> (the same; to avoid additionally remembering what was previously sent to 
> HBase).
> 5) Tests, comments ;)



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HBASE-22254) refactor and improve decommissioning logic

2019-05-09 Thread Sergey Shelukhin (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-22254?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergey Shelukhin updated HBASE-22254:
-
Status: Patch Available  (was: Open)

Fixed the test, also made the client changes to make new API features usable 
not just directly via a request.

> refactor and improve decommissioning logic
> --
>
> Key: HBASE-22254
> URL: https://issues.apache.org/jira/browse/HBASE-22254
> Project: HBase
>  Issue Type: Improvement
>Affects Versions: 3.0.0
>Reporter: Sergey Shelukhin
>Assignee: Sergey Shelukhin
>Priority: Major
> Attachments: HBASE-22254.01.patch, HBASE-22254.02.patch, 
> HBASE-22254.patch
>
>
> Making some changes needed to support better decommissioning on large 
> clusters and with container mode; to test those and add clarify I moved parts 
> of decommissioning logic from HMaster, Draining tracker, and ServerManager 
> into a separate class.
> Features added/improvements:
> 1) More resilient off-loading; right now off-loading fails for a subset of 
> regions in case of a single region failure; is never done on master restart, 
> etc.
> 2) Option to kill RS after off-loading (good for container mode HBase, e.g. 
> on YARN).
> 3) Option to specify machine names only to decommission, for the API to be 
> usable for an external system that doesn't care about HBase server names, or 
> e.g. multiple RS in containers on the same node.
> 4) Option to replace existing decommissioning list instead of adding to it 
> (the same; to avoid additionally remembering what was previously sent to 
> HBase).
> 5) Tests, comments ;)



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HBASE-22254) refactor and improve decommissioning logic

2019-05-09 Thread Sergey Shelukhin (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-22254?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergey Shelukhin updated HBASE-22254:
-
Attachment: HBASE-22254.02.patch

> refactor and improve decommissioning logic
> --
>
> Key: HBASE-22254
> URL: https://issues.apache.org/jira/browse/HBASE-22254
> Project: HBase
>  Issue Type: Improvement
>Affects Versions: 3.0.0
>Reporter: Sergey Shelukhin
>Assignee: Sergey Shelukhin
>Priority: Major
> Attachments: HBASE-22254.01.patch, HBASE-22254.02.patch, 
> HBASE-22254.patch
>
>
> Making some changes needed to support better decommissioning on large 
> clusters and with container mode; to test those and add clarify I moved parts 
> of decommissioning logic from HMaster, Draining tracker, and ServerManager 
> into a separate class.
> Features added/improvements:
> 1) More resilient off-loading; right now off-loading fails for a subset of 
> regions in case of a single region failure; is never done on master restart, 
> etc.
> 2) Option to kill RS after off-loading (good for container mode HBase, e.g. 
> on YARN).
> 3) Option to specify machine names only to decommission, for the API to be 
> usable for an external system that doesn't care about HBase server names, or 
> e.g. multiple RS in containers on the same node.
> 4) Option to replace existing decommissioning list instead of adding to it 
> (the same; to avoid additionally remembering what was previously sent to 
> HBase).
> 5) Tests, comments ;)



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-22389) Revert HBASE-19275 TestSnapshotFileCache never worked properly on branch-1

2019-05-09 Thread Zach York (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-22389?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16836807#comment-16836807
 ] 

Zach York commented on HBASE-22389:
---

Thanks for the reviews [~apurtell]!

> Revert HBASE-19275 TestSnapshotFileCache never worked properly on branch-1
> --
>
> Key: HBASE-22389
> URL: https://issues.apache.org/jira/browse/HBASE-22389
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 1.5.0
>Reporter: Zach York
>Assignee: Zach York
>Priority: Major
> Fix For: 1.5.0, 1.4.11
>
>
> This commit seems broken on branch-1. 
> https://github.com/apache/hbase/commit/d80d3fa45487fedf203c7144e8267753e4d53529#diff-ae046012dd310fc116a729bf36c7df81
>  seems to have fixed the issue that the other HBASE-19275 patches fixed, but 
> when HBASE-19275 was applied to branch-1 it appears to have reverted this 
> code: 
> https://github.com/apache/hbase/commit/e8e4beacb039299a9e56cd355e3ada08784a7acb#diff-ae046012dd310fc116a729bf36c7df81
>  meaning the tests always pass again.
> [~apurtell] [~xucang]



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] [hbase] Apache-HBase commented on issue #229: HBASE-22389 Revert "HBASE-19275 TestSnapshotFileCache never worked pr…

2019-05-09 Thread GitBox
Apache-HBase commented on issue #229: HBASE-22389 Revert "HBASE-19275 
TestSnapshotFileCache never worked pr…
URL: https://github.com/apache/hbase/pull/229#issuecomment-491112303
 
 
   :broken_heart: **-1 overall**
   
   
   
   
   
   
   | Vote | Subsystem | Runtime | Comment |
   |::|--:|:|:|
   | 0 | reexec | 0 | Docker mode activated. |
   | -1 | patch | 9 | https://github.com/apache/hbase/pull/229 does not apply 
to branch-1. Rebase required? Wrong Branch? See 
https://yetus.apache.org/documentation/in-progress/precommit-patchnames for 
help. |
   
   
   | Subsystem | Report/Notes |
   |--:|:-|
   | GITHUB PR | https://github.com/apache/hbase/pull/229 |
   | Console output | 
https://builds.apache.org/job/HBase-PreCommit-GitHub-PR/job/PR-229/2/console |
   | Powered by | Apache Yetus 0.9.0 http://yetus.apache.org |
   
   
   This message was automatically generated.
   
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Resolved] (HBASE-22389) Revert HBASE-19275 TestSnapshotFileCache never worked properly on branch-1

2019-05-09 Thread Zach York (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-22389?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zach York resolved HBASE-22389.
---
   Resolution: Fixed
Fix Version/s: 1.4.11
   1.5.0

> Revert HBASE-19275 TestSnapshotFileCache never worked properly on branch-1
> --
>
> Key: HBASE-22389
> URL: https://issues.apache.org/jira/browse/HBASE-22389
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 1.5.0
>Reporter: Zach York
>Assignee: Zach York
>Priority: Major
> Fix For: 1.5.0, 1.4.11
>
>
> This commit seems broken on branch-1. 
> https://github.com/apache/hbase/commit/d80d3fa45487fedf203c7144e8267753e4d53529#diff-ae046012dd310fc116a729bf36c7df81
>  seems to have fixed the issue that the other HBASE-19275 patches fixed, but 
> when HBASE-19275 was applied to branch-1 it appears to have reverted this 
> code: 
> https://github.com/apache/hbase/commit/e8e4beacb039299a9e56cd355e3ada08784a7acb#diff-ae046012dd310fc116a729bf36c7df81
>  meaning the tests always pass again.
> [~apurtell] [~xucang]



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] [hbase] z-york commented on issue #229: HBASE-22389 Revert "HBASE-19275 TestSnapshotFileCache never worked pr…

2019-05-09 Thread GitBox
z-york commented on issue #229: HBASE-22389 Revert "HBASE-19275 
TestSnapshotFileCache never worked pr…
URL: https://github.com/apache/hbase/pull/229#issuecomment-491110739
 
 
   Reran TestSnapshotFileCache after #230 and confirmed it passes now. Fixed 
checkstyle.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [hbase] z-york merged pull request #229: HBASE-22389 Revert "HBASE-19275 TestSnapshotFileCache never worked pr…

2019-05-09 Thread GitBox
z-york merged pull request #229: HBASE-22389 Revert "HBASE-19275 
TestSnapshotFileCache never worked pr…
URL: https://github.com/apache/hbase/pull/229
 
 
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [hbase] z-york commented on issue #230: HBASE-22390 Backport HBASE-22190 to branch-1

2019-05-09 Thread GitBox
z-york commented on issue #230: HBASE-22390 Backport HBASE-22190 to branch-1
URL: https://github.com/apache/hbase/pull/230#issuecomment-491108669
 
 
   pushed to branch-1 and branch-1.4. closing.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Updated] (HBASE-22391) Fix flaky tests from TestFromClientSide

2019-05-09 Thread Xu Cang (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-22391?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xu Cang updated HBASE-22391:

Issue Type: Bug  (was: New Feature)

> Fix flaky tests from TestFromClientSide
> ---
>
> Key: HBASE-22391
> URL: https://issues.apache.org/jira/browse/HBASE-22391
> Project: HBase
>  Issue Type: Bug
>  Components: test
>Affects Versions: 3.0.0, 2.0.5, 1.5.1
>Reporter: Xu Cang
>Assignee: Xu Cang
>Priority: Minor
> Fix For: 3.0.0, 1.5.0, 2.3.0, 2.1.5, 2.2.1, 1.3.5, 1.4.11
>
> Attachments: HBASE-22391-branch-1.001.patch
>
>
> tests in TestFromClientSide.java in general are flaky due to the reason that 
> after createTable, they did not wait for table to be ready before adding data 
> into table.
> Found this issue when working on HBASE-22274
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] [hbase] z-york closed pull request #230: HBASE-22390 Backport HBASE-22190 to branch-1

2019-05-09 Thread GitBox
z-york closed pull request #230: HBASE-22390 Backport HBASE-22190 to branch-1
URL: https://github.com/apache/hbase/pull/230
 
 
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Updated] (HBASE-22391) Fix flaky tests from TestFromClientSide

2019-05-09 Thread Xu Cang (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-22391?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xu Cang updated HBASE-22391:

Attachment: HBASE-22391-branch-1.001.patch
Status: Patch Available  (was: Open)

> Fix flaky tests from TestFromClientSide
> ---
>
> Key: HBASE-22391
> URL: https://issues.apache.org/jira/browse/HBASE-22391
> Project: HBase
>  Issue Type: New Feature
>  Components: test
>Affects Versions: 2.0.5, 3.0.0, 1.5.1
>Reporter: Xu Cang
>Assignee: Xu Cang
>Priority: Minor
> Fix For: 3.0.0, 1.5.0, 2.3.0, 2.1.5, 2.2.1, 1.3.5, 1.4.11
>
> Attachments: HBASE-22391-branch-1.001.patch
>
>
> tests in TestFromClientSide.java in general are flaky due to the reason that 
> after createTable, they did not wait for table to be ready before adding data 
> into table.
> Found this issue when working on HBASE-22274
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HBASE-22390) backport HBASE-22190 to branch-1

2019-05-09 Thread Zach York (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-22390?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zach York updated HBASE-22390:
--
   Resolution: Fixed
Fix Version/s: 1.4.11
   1.5.0
   Status: Resolved  (was: Patch Available)

> backport HBASE-22190 to branch-1
> 
>
> Key: HBASE-22390
> URL: https://issues.apache.org/jira/browse/HBASE-22390
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 1.5.0, 1.4.10
>Reporter: Zach York
>Assignee: Zach York
>Priority: Major
> Fix For: 1.5.0, 1.4.11
>
>
> HBASE-22190 was only applied to branch-2+ while the bug exists in branch-1 as 
> well.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HBASE-22274) Cell size limit check on append should consider cell's previous size.

2019-05-09 Thread Xu Cang (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-22274?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xu Cang updated HBASE-22274:

Attachment: HBASE-22274-branch-1.005.patch

> Cell size limit check on append should consider cell's previous size.
> -
>
> Key: HBASE-22274
> URL: https://issues.apache.org/jira/browse/HBASE-22274
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 3.0.0, 2.0.0, 1.3.5
>Reporter: Xu Cang
>Assignee: Xu Cang
>Priority: Minor
> Attachments: HBASE-22274-branch-1.001.patch, 
> HBASE-22274-branch-1.002.patch, HBASE-22274-branch-1.003.patch, 
> HBASE-22274-branch-1.003.patch, HBASE-22274-branch-1.004.patch, 
> HBASE-22274-branch-1.005.patch, HBASE-22274-master.001.patch, 
> HBASE-22274-master.002.patch, HBASE-22274-master.002.patch, 
> HBASE-22274-master.003.patch
>
>
> Now we have cell size limit check based on this parameter 
> *hbase.server.keyvalue.maxsize* 
> One case was missing: appending to a cell only take append op's cell size 
> into account against this limit check. we should check against the potential 
> final cell size after the append.'
> It's easy to reproduce this :
>  
> Apply this diff
>  
> {code:java}
> diff --git 
> a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java
>  
> b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java
>  index 5a285ef6ba..8633177ebe 100644 --- 
> a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java
>  +++ 
> b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java
>  @@ -6455,7 +6455,7 
> - t.append(new Append(ROW).addColumn(FAMILY, QUALIFIER, new byte[10 * 
> 1024])); 
> + t.append(new Append(ROW).addColumn(FAMILY, QUALIFIER, new byte[2 * 1024])); 
> {code}
>  
> Fix is to add this check in #reckonDeltas in HRegion class, where we have 
> already got the appended cell's size. 
> Will throw DoNotRetryIOException if checks is failed.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-17370) Fix or provide shell scripts to drain and decommission region server

2019-05-09 Thread Sergey Shelukhin (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-17370?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16836783#comment-16836783
 ] 

Sergey Shelukhin commented on HBASE-17370:
--

Should the book be updated for this? Looks like it still suggests creating 
znodes to decommission region servers.

> Fix or provide shell scripts to drain and decommission region server
> 
>
> Key: HBASE-17370
> URL: https://issues.apache.org/jira/browse/HBASE-17370
> Project: HBase
>  Issue Type: Sub-task
>Reporter: Jerry He
>Assignee: Nihal Jain
>Priority: Major
>  Labels: operability
> Fix For: 3.0.0, 2.2.0
>
> Attachments: HBASE-17370.branch-2.001.patch, 
> HBASE-17370.master.001.patch, HBASE-17370.master.002.patch
>
>
> 1. Update the existing shell scripts to use the new drain related API.
> 2  Or provide new shell scripts.
> 3. Provide a 'decommission' shell tool that puts the server in drain mode and 
> offload the server.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] [hbase] z-york commented on issue #230: HBASE-22390 Backport HBASE-22190 to branch-1

2019-05-09 Thread GitBox
z-york commented on issue #230: HBASE-22390 Backport HBASE-22190 to branch-1
URL: https://github.com/apache/hbase/pull/230#issuecomment-491104124
 
 
   all these tests pass locally. Fixed checkstyle as well. Committing.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (HBASE-22274) Cell size limit check on append should consider cell's previous size.

2019-05-09 Thread Andrew Purtell (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-22274?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16836780#comment-16836780
 ] 

Andrew Purtell commented on HBASE-22274:


bq. Reading code in this class "RSRpcServices", I can see the pattern people 
followed before was : we do not log client originated error or informational 
messages by default. 

I accept this argument, thank you, sounds good to me.

+1, with the test changed moved out


> Cell size limit check on append should consider cell's previous size.
> -
>
> Key: HBASE-22274
> URL: https://issues.apache.org/jira/browse/HBASE-22274
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 3.0.0, 2.0.0, 1.3.5
>Reporter: Xu Cang
>Assignee: Xu Cang
>Priority: Minor
> Attachments: HBASE-22274-branch-1.001.patch, 
> HBASE-22274-branch-1.002.patch, HBASE-22274-branch-1.003.patch, 
> HBASE-22274-branch-1.003.patch, HBASE-22274-branch-1.004.patch, 
> HBASE-22274-master.001.patch, HBASE-22274-master.002.patch, 
> HBASE-22274-master.002.patch, HBASE-22274-master.003.patch
>
>
> Now we have cell size limit check based on this parameter 
> *hbase.server.keyvalue.maxsize* 
> One case was missing: appending to a cell only take append op's cell size 
> into account against this limit check. we should check against the potential 
> final cell size after the append.'
> It's easy to reproduce this :
>  
> Apply this diff
>  
> {code:java}
> diff --git 
> a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java
>  
> b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java
>  index 5a285ef6ba..8633177ebe 100644 --- 
> a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java
>  +++ 
> b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java
>  @@ -6455,7 +6455,7 
> - t.append(new Append(ROW).addColumn(FAMILY, QUALIFIER, new byte[10 * 
> 1024])); 
> + t.append(new Append(ROW).addColumn(FAMILY, QUALIFIER, new byte[2 * 1024])); 
> {code}
>  
> Fix is to add this check in #reckonDeltas in HRegion class, where we have 
> already got the appended cell's size. 
> Will throw DoNotRetryIOException if checks is failed.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Comment Edited] (HBASE-22274) Cell size limit check on append should consider cell's previous size.

2019-05-09 Thread Xu Cang (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-22274?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16836777#comment-16836777
 ] 

Xu Cang edited comment on HBASE-22274 at 5/9/19 11:30 PM:
--

When cell size exceeds the limit, we always throw exception like this

 
{code:java}
throw new DoNotRetryIOException(msg);{code}
And this will be highly likely end up in client's log.

Reading code in this class "RSRpcServices", I can see the pattern people 
followed before was : we do not log client originated error or informational 
messages by default.  Such as this one

 
{code:java}
if (LOG.isDebugEnabled()) {
 LOG.debug(
 "Server shutting down and client tried to access missing scanner " + 
scannerName);
}{code}
So, I am OK with keeping the current logging strategy for cell limit one. As 
long as server rejects it, client will handle this case from his end. Open to 
discuss more about this if you think otherwise, [~apurtell]


was (Author: xucang):
When cell size exceeds the limit, we always throw exception like this

 
{code:java}
throw new DoNotRetryIOException(msg);{code}
And this will be highly likely end up in client's log.

Reading doe in this class "RSRpcServices", I can see the pattern people 
followed before was : we do not log client originated error or informational 
messages by default.  Such as this one

 
{code:java}
if (LOG.isDebugEnabled()) {
 LOG.debug(
 "Server shutting down and client tried to access missing scanner " + 
scannerName);
}{code}
So, I am OK with keeping the current logging strategy for cell limit one. As 
long as server rejects it, client will handle this case from his end. Open to 
discuss more about this if you think otherwise, [~apurtell]

> Cell size limit check on append should consider cell's previous size.
> -
>
> Key: HBASE-22274
> URL: https://issues.apache.org/jira/browse/HBASE-22274
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 3.0.0, 2.0.0, 1.3.5
>Reporter: Xu Cang
>Assignee: Xu Cang
>Priority: Minor
> Attachments: HBASE-22274-branch-1.001.patch, 
> HBASE-22274-branch-1.002.patch, HBASE-22274-branch-1.003.patch, 
> HBASE-22274-branch-1.003.patch, HBASE-22274-branch-1.004.patch, 
> HBASE-22274-master.001.patch, HBASE-22274-master.002.patch, 
> HBASE-22274-master.002.patch, HBASE-22274-master.003.patch
>
>
> Now we have cell size limit check based on this parameter 
> *hbase.server.keyvalue.maxsize* 
> One case was missing: appending to a cell only take append op's cell size 
> into account against this limit check. we should check against the potential 
> final cell size after the append.'
> It's easy to reproduce this :
>  
> Apply this diff
>  
> {code:java}
> diff --git 
> a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java
>  
> b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java
>  index 5a285ef6ba..8633177ebe 100644 --- 
> a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java
>  +++ 
> b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java
>  @@ -6455,7 +6455,7 
> - t.append(new Append(ROW).addColumn(FAMILY, QUALIFIER, new byte[10 * 
> 1024])); 
> + t.append(new Append(ROW).addColumn(FAMILY, QUALIFIER, new byte[2 * 1024])); 
> {code}
>  
> Fix is to add this check in #reckonDeltas in HRegion class, where we have 
> already got the appended cell's size. 
> Will throw DoNotRetryIOException if checks is failed.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-22274) Cell size limit check on append should consider cell's previous size.

2019-05-09 Thread Xu Cang (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-22274?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16836777#comment-16836777
 ] 

Xu Cang commented on HBASE-22274:
-

When cell size exceeds the limit, we always throw exception like this

 
{code:java}
throw new DoNotRetryIOException(msg);{code}
And this will be highly likely end up in client's log.

Reading doe in this class "RSRpcServices", I can see the pattern people 
followed before was : we do not log client originated error or informational 
messages by default.  Such as this one

 
{code:java}
if (LOG.isDebugEnabled()) {
 LOG.debug(
 "Server shutting down and client tried to access missing scanner " + 
scannerName);
}{code}
So, I am OK with keeping the current logging strategy for cell limit one. As 
long as server rejects it, client will handle this case from his end. Open to 
discuss more about this if you think otherwise, [~apurtell]

> Cell size limit check on append should consider cell's previous size.
> -
>
> Key: HBASE-22274
> URL: https://issues.apache.org/jira/browse/HBASE-22274
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 3.0.0, 2.0.0, 1.3.5
>Reporter: Xu Cang
>Assignee: Xu Cang
>Priority: Minor
> Attachments: HBASE-22274-branch-1.001.patch, 
> HBASE-22274-branch-1.002.patch, HBASE-22274-branch-1.003.patch, 
> HBASE-22274-branch-1.003.patch, HBASE-22274-branch-1.004.patch, 
> HBASE-22274-master.001.patch, HBASE-22274-master.002.patch, 
> HBASE-22274-master.002.patch, HBASE-22274-master.003.patch
>
>
> Now we have cell size limit check based on this parameter 
> *hbase.server.keyvalue.maxsize* 
> One case was missing: appending to a cell only take append op's cell size 
> into account against this limit check. we should check against the potential 
> final cell size after the append.'
> It's easy to reproduce this :
>  
> Apply this diff
>  
> {code:java}
> diff --git 
> a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java
>  
> b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java
>  index 5a285ef6ba..8633177ebe 100644 --- 
> a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java
>  +++ 
> b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java
>  @@ -6455,7 +6455,7 
> - t.append(new Append(ROW).addColumn(FAMILY, QUALIFIER, new byte[10 * 
> 1024])); 
> + t.append(new Append(ROW).addColumn(FAMILY, QUALIFIER, new byte[2 * 1024])); 
> {code}
>  
> Fix is to add this check in #reckonDeltas in HRegion class, where we have 
> already got the appended cell's size. 
> Will throw DoNotRetryIOException if checks is failed.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HBASE-22330) Backport HBASE-20724 (Sometimes some compacted storefiles are still opened after region failover) to branch-1

2019-05-09 Thread Abhishek Singh Chouhan (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-22330?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Abhishek Singh Chouhan updated HBASE-22330:
---
  Resolution: Fixed
Hadoop Flags: Reviewed
  Status: Resolved  (was: Patch Available)

> Backport HBASE-20724 (Sometimes some compacted storefiles are still opened 
> after region failover) to branch-1
> -
>
> Key: HBASE-22330
> URL: https://issues.apache.org/jira/browse/HBASE-22330
> Project: HBase
>  Issue Type: Sub-task
>  Components: Compaction, regionserver
>Affects Versions: 1.5.0, 1.4.9, 1.3.4
>Reporter: Andrew Purtell
>Assignee: Abhishek Singh Chouhan
>Priority: Major
> Fix For: 1.5.0, 1.3.5, 1.4.11
>
> Attachments: HBASE-22330-addendum.branch-1.patch, 
> HBASE-22330.branch-1.001.patch, HBASE-22330.branch-1.002.patch, 
> HBASE-22330.branch-1.3.001.patch
>
>
> There appears to be a race condition between close and split which when 
> combined with a side effect of HBASE-20704, leads to the parent region store 
> files getting archived and cleared while daughter regions still have 
> references to those parent region store files.
> Here is the timeline of events observed for an affected region:
>  # RS1 faces ZooKeeper connectivity issue for master node and starts shutting 
> itself down. As part of this it starts to close the store and clean up the 
> compacted files (File A)
>  # Master starts bulk assigning regions and assign parent region to RS2
>  # Region opens on RS2 and ends up opening compacted store file(s) (suspect 
> this is due to HBASE-20724)
>  # Now split happens and daughter regions open on RS2 and try to run a 
> compaction as part of post open
>  # Split request at this point is complete. However now archiving proceeds on 
> RS1 and ends up archiving the store file that is referenced by the daughter. 
> Compaction fails due to FileNotFoundException and all subsequent attempts to 
> open the region will fail until manual resolution.
> We think having HBASE-20724 would help in such situations since we won't end 
> up loading compacted store files in the first place. 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (HBASE-22391) Fix flaky tests from TestFromClientSide

2019-05-09 Thread Xu Cang (JIRA)
Xu Cang created HBASE-22391:
---

 Summary: Fix flaky tests from TestFromClientSide
 Key: HBASE-22391
 URL: https://issues.apache.org/jira/browse/HBASE-22391
 Project: HBase
  Issue Type: New Feature
  Components: test
Affects Versions: 2.0.5, 3.0.0, 1.5.1
Reporter: Xu Cang


tests in TestFromClientSide.java in general are flaky due to the reason that 
after createTable, they did not wait for table to be ready before adding data 
into table.

Found this issue when working on HBASE-22274

 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Assigned] (HBASE-22391) Fix flaky tests from TestFromClientSide

2019-05-09 Thread Andrew Purtell (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-22391?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew Purtell reassigned HBASE-22391:
--

Assignee: Xu Cang

> Fix flaky tests from TestFromClientSide
> ---
>
> Key: HBASE-22391
> URL: https://issues.apache.org/jira/browse/HBASE-22391
> Project: HBase
>  Issue Type: New Feature
>  Components: test
>Affects Versions: 3.0.0, 2.0.5, 1.5.1
>Reporter: Xu Cang
>Assignee: Xu Cang
>Priority: Minor
> Fix For: 3.0.0, 1.5.0, 2.3.0, 2.1.5, 2.2.1, 1.3.5, 1.4.11
>
>
> tests in TestFromClientSide.java in general are flaky due to the reason that 
> after createTable, they did not wait for table to be ready before adding data 
> into table.
> Found this issue when working on HBASE-22274
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HBASE-22391) Fix flaky tests from TestFromClientSide

2019-05-09 Thread Andrew Purtell (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-22391?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew Purtell updated HBASE-22391:
---
Fix Version/s: 1.4.11
   1.3.5
   2.2.1
   2.1.5
   2.3.0
   1.5.0
   3.0.0

> Fix flaky tests from TestFromClientSide
> ---
>
> Key: HBASE-22391
> URL: https://issues.apache.org/jira/browse/HBASE-22391
> Project: HBase
>  Issue Type: New Feature
>  Components: test
>Affects Versions: 3.0.0, 2.0.5, 1.5.1
>Reporter: Xu Cang
>Priority: Minor
> Fix For: 3.0.0, 1.5.0, 2.3.0, 2.1.5, 2.2.1, 1.3.5, 1.4.11
>
>
> tests in TestFromClientSide.java in general are flaky due to the reason that 
> after createTable, they did not wait for table to be ready before adding data 
> into table.
> Found this issue when working on HBASE-22274
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-22274) Cell size limit check on append should consider cell's previous size.

2019-05-09 Thread Xu Cang (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-22274?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16836766#comment-16836766
 ] 

Xu Cang commented on HBASE-22274:
-

Moving unit test fix to https://issues.apache.org/jira/browse/HBASE-22391

 

> Cell size limit check on append should consider cell's previous size.
> -
>
> Key: HBASE-22274
> URL: https://issues.apache.org/jira/browse/HBASE-22274
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 3.0.0, 2.0.0, 1.3.5
>Reporter: Xu Cang
>Assignee: Xu Cang
>Priority: Minor
> Attachments: HBASE-22274-branch-1.001.patch, 
> HBASE-22274-branch-1.002.patch, HBASE-22274-branch-1.003.patch, 
> HBASE-22274-branch-1.003.patch, HBASE-22274-branch-1.004.patch, 
> HBASE-22274-master.001.patch, HBASE-22274-master.002.patch, 
> HBASE-22274-master.002.patch, HBASE-22274-master.003.patch
>
>
> Now we have cell size limit check based on this parameter 
> *hbase.server.keyvalue.maxsize* 
> One case was missing: appending to a cell only take append op's cell size 
> into account against this limit check. we should check against the potential 
> final cell size after the append.'
> It's easy to reproduce this :
>  
> Apply this diff
>  
> {code:java}
> diff --git 
> a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java
>  
> b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java
>  index 5a285ef6ba..8633177ebe 100644 --- 
> a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java
>  +++ 
> b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java
>  @@ -6455,7 +6455,7 
> - t.append(new Append(ROW).addColumn(FAMILY, QUALIFIER, new byte[10 * 
> 1024])); 
> + t.append(new Append(ROW).addColumn(FAMILY, QUALIFIER, new byte[2 * 1024])); 
> {code}
>  
> Fix is to add this check in #reckonDeltas in HRegion class, where we have 
> already got the appended cell's size. 
> Will throw DoNotRetryIOException if checks is failed.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-22330) Backport HBASE-20724 (Sometimes some compacted storefiles are still opened after region failover) to branch-1

2019-05-09 Thread Abhishek Singh Chouhan (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-22330?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16836764#comment-16836764
 ] 

Abhishek Singh Chouhan commented on HBASE-22330:


Pushed to relevant branches. Thanks [~xucang] [~apurtell]

> Backport HBASE-20724 (Sometimes some compacted storefiles are still opened 
> after region failover) to branch-1
> -
>
> Key: HBASE-22330
> URL: https://issues.apache.org/jira/browse/HBASE-22330
> Project: HBase
>  Issue Type: Sub-task
>  Components: Compaction, regionserver
>Affects Versions: 1.5.0, 1.4.9, 1.3.4
>Reporter: Andrew Purtell
>Assignee: Abhishek Singh Chouhan
>Priority: Major
> Fix For: 1.5.0, 1.3.5, 1.4.11
>
> Attachments: HBASE-22330-addendum.branch-1.patch, 
> HBASE-22330.branch-1.001.patch, HBASE-22330.branch-1.002.patch, 
> HBASE-22330.branch-1.3.001.patch
>
>
> There appears to be a race condition between close and split which when 
> combined with a side effect of HBASE-20704, leads to the parent region store 
> files getting archived and cleared while daughter regions still have 
> references to those parent region store files.
> Here is the timeline of events observed for an affected region:
>  # RS1 faces ZooKeeper connectivity issue for master node and starts shutting 
> itself down. As part of this it starts to close the store and clean up the 
> compacted files (File A)
>  # Master starts bulk assigning regions and assign parent region to RS2
>  # Region opens on RS2 and ends up opening compacted store file(s) (suspect 
> this is due to HBASE-20724)
>  # Now split happens and daughter regions open on RS2 and try to run a 
> compaction as part of post open
>  # Split request at this point is complete. However now archiving proceeds on 
> RS1 and ends up archiving the store file that is referenced by the daughter. 
> Compaction fails due to FileNotFoundException and all subsequent attempts to 
> open the region will fail until manual resolution.
> We think having HBASE-20724 would help in such situations since we won't end 
> up loading compacted store files in the first place. 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] [hbase] Apache-HBase commented on issue #230: HBASE-22390 Backport HBASE-22190 to branch-1

2019-05-09 Thread GitBox
Apache-HBase commented on issue #230: HBASE-22390 Backport HBASE-22190 to 
branch-1
URL: https://github.com/apache/hbase/pull/230#issuecomment-491093578
 
 
   :broken_heart: **-1 overall**
   
   
   
   
   
   
   | Vote | Subsystem | Runtime | Comment |
   |::|--:|:|:|
   | 0 | reexec | 1278 | Docker mode activated. |
   ||| _ Prechecks _ |
   | 0 | findbugs | 0 | Findbugs executables are not available. |
   | +1 | hbaseanti | 0 |  Patch does not have any anti-patterns. |
   | +1 | @author | 0 | The patch does not contain any @author tags. |
   | -0 | test4tests | 0 | The patch doesn't appear to include any new or 
modified tests.  Please justify why no new tests are needed for this patch. 
Also please list what manual steps were performed to verify this patch. |
   ||| _ branch-1 Compile Tests _ |
   | +1 | mvninstall | 150 | branch-1 passed |
   | +1 | compile | 38 | branch-1 passed with JDK v1.8.0_212 |
   | +1 | compile | 48 | branch-1 passed with JDK v1.7.0_222 |
   | +1 | checkstyle | 80 | branch-1 passed |
   | +1 | shadedjars | 159 | branch has no errors when building our shaded 
downstream artifacts. |
   | +1 | javadoc | 57 | branch-1 passed with JDK v1.8.0_212 |
   | +1 | javadoc | 62 | branch-1 passed with JDK v1.7.0_222 |
   ||| _ Patch Compile Tests _ |
   | +1 | mvninstall | 151 | the patch passed |
   | +1 | compile | 62 | the patch passed with JDK v1.8.0_212 |
   | +1 | javac | 62 | the patch passed |
   | +1 | compile | 50 | the patch passed with JDK v1.7.0_222 |
   | +1 | javac | 50 | the patch passed |
   | -1 | checkstyle | 79 | hbase-server: The patch generated 4 new + 2 
unchanged - 2 fixed = 6 total (was 4) |
   | +1 | whitespace | 0 | The patch has no whitespace issues. |
   | +1 | shadedjars | 155 | patch has no errors when building our shaded 
downstream artifacts. |
   | +1 | hadoopcheck | 109 | Patch does not cause any errors with Hadoop 
2.7.4. |
   | +1 | javadoc | 28 | the patch passed with JDK v1.8.0_212 |
   | +1 | javadoc | 38 | the patch passed with JDK v1.7.0_222 |
   ||| _ Other Tests _ |
   | -1 | unit | 8889 | hbase-server in the patch failed. |
   | +1 | asflicense | 87 | The patch does not generate ASF License warnings. |
   | | | 11603 | |
   
   
   | Reason | Tests |
   |---:|:--|
   | Failed junit tests | hadoop.hbase.regionserver.wal.TestLogRolling |
   |   | hadoop.hbase.regionserver.TestPerColumnFamilyFlush |
   |   | hadoop.hbase.regionserver.TestEncryptionKeyRotation |
   |   | hadoop.hbase.replication.TestReplicationSmallTests |
   
   
   | Subsystem | Report/Notes |
   |--:|:-|
   | Docker | Client=17.05.0-ce Server=17.05.0-ce base: 
https://builds.apache.org/job/HBase-PreCommit-GitHub-PR/job/PR-230/1/artifact/out/Dockerfile
 |
   | GITHUB PR | https://github.com/apache/hbase/pull/230 |
   | Optional Tests |  dupname  asflicense  javac  javadoc  unit  findbugs  
shadedjars  hadoopcheck  hbaseanti  checkstyle  compile  |
   | uname | Linux b54248fbc097 4.4.0-139-generic #165-Ubuntu SMP Wed Oct 24 
10:58:50 UTC 2018 x86_64 x86_64 x86_64 GNU/Linux |
   | Build tool | maven |
   | Personality | /testptch/patchprocess/precommit/personality/provided.sh |
   | git revision | branch-1 / 56548fc |
   | maven | version: Apache Maven 3.0.5 |
   | Default Java | 1.7.0_222 |
   | Multi-JDK versions |  /usr/lib/jvm/java-8-openjdk-amd64:1.8.0_212 
/usr/lib/jvm/java-7-openjdk-amd64:1.7.0_222 |
   | checkstyle | 
https://builds.apache.org/job/HBase-PreCommit-GitHub-PR/job/PR-230/1/artifact/out/diff-checkstyle-hbase-server.txt
 |
   | unit | 
https://builds.apache.org/job/HBase-PreCommit-GitHub-PR/job/PR-230/1/artifact/out/patch-unit-hbase-server.txt
 |
   |  Test Results | 
https://builds.apache.org/job/HBase-PreCommit-GitHub-PR/job/PR-230/1/testReport/
 |
   | Max. process+thread count | 3812 (vs. ulimit of 1) |
   | modules | C: hbase-server U: hbase-server |
   | Console output | 
https://builds.apache.org/job/HBase-PreCommit-GitHub-PR/job/PR-230/1/console |
   | Powered by | Apache Yetus 0.9.0 http://yetus.apache.org |
   
   
   This message was automatically generated.
   
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (HBASE-22274) Cell size limit check on append should consider cell's previous size.

2019-05-09 Thread Andrew Purtell (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-22274?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16836731#comment-16836731
 ] 

Andrew Purtell commented on HBASE-22274:


bq. There are 88 'createTable' calls in TestFromClientSide.java none of them 
calls 'waitTableAvailable'. It's incorrect to do 'Put' before table is 
available in my opinion. That's why it's causing flaky testing result.

Let's break this part of the patch out into a separate JIRA so we have a 
clearer history for the test fix. We can ignore the TestFromClientSide result 
in the context of this issue. 

> Cell size limit check on append should consider cell's previous size.
> -
>
> Key: HBASE-22274
> URL: https://issues.apache.org/jira/browse/HBASE-22274
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 3.0.0, 2.0.0, 1.3.5
>Reporter: Xu Cang
>Assignee: Xu Cang
>Priority: Minor
> Attachments: HBASE-22274-branch-1.001.patch, 
> HBASE-22274-branch-1.002.patch, HBASE-22274-branch-1.003.patch, 
> HBASE-22274-branch-1.003.patch, HBASE-22274-branch-1.004.patch, 
> HBASE-22274-master.001.patch, HBASE-22274-master.002.patch, 
> HBASE-22274-master.002.patch, HBASE-22274-master.003.patch
>
>
> Now we have cell size limit check based on this parameter 
> *hbase.server.keyvalue.maxsize* 
> One case was missing: appending to a cell only take append op's cell size 
> into account against this limit check. we should check against the potential 
> final cell size after the append.'
> It's easy to reproduce this :
>  
> Apply this diff
>  
> {code:java}
> diff --git 
> a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java
>  
> b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java
>  index 5a285ef6ba..8633177ebe 100644 --- 
> a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java
>  +++ 
> b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java
>  @@ -6455,7 +6455,7 
> - t.append(new Append(ROW).addColumn(FAMILY, QUALIFIER, new byte[10 * 
> 1024])); 
> + t.append(new Append(ROW).addColumn(FAMILY, QUALIFIER, new byte[2 * 1024])); 
> {code}
>  
> Fix is to add this check in #reckonDeltas in HRegion class, where we have 
> already got the appended cell's size. 
> Will throw DoNotRetryIOException if checks is failed.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Comment Edited] (HBASE-22274) Cell size limit check on append should consider cell's previous size.

2019-05-09 Thread Andrew Purtell (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-22274?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16836729#comment-16836729
 ] 

Andrew Purtell edited comment on HBASE-22274 at 5/9/19 9:49 PM:


v4 patch lgtm

One minor question, on this change:
{code}
+int newCellSize = CellUtil.estimatedSerializedSizeOf(newCell);
+if (newCellSize > this.maxCellSize) {
+  String msg = "Cell with size " + newCellSize + " exceeds 
limit of " +
+  this.maxCellSize + " bytes";
+  if (LOG.isDebugEnabled()) {
+LOG.debug(msg);
+  }
+  throw new DoNotRetryIOException(msg);
+}
{code}

Should we be logging this at a higher log level, like INFO, or even WARN? 
Whatever we decide here should be consistent with other warnings issued at the 
other sites where we do this size limit test. (I.e. if DEBUG there and we 
decide WARN, update to patch both warnings to WARN, or vice versa.)


was (Author: apurtell):
v4 patch lgtm

The test issue is clearer now.

One minor question, on this change:
{code}
+int newCellSize = CellUtil.estimatedSerializedSizeOf(newCell);
+if (newCellSize > this.maxCellSize) {
+  String msg = "Cell with size " + newCellSize + " exceeds 
limit of " +
+  this.maxCellSize + " bytes";
+  if (LOG.isDebugEnabled()) {
+LOG.debug(msg);
+  }
+  throw new DoNotRetryIOException(msg);
+}
{code}

Should we be logging this at a higher log level, like INFO, or even WARN? 
Whatever we decide here should be consistent with other warnings issued at the 
other sites where we do this size limit test. (I.e. if DEBUG there and we 
decide WARN, update to patch both warnings to WARN, or vice versa.)

> Cell size limit check on append should consider cell's previous size.
> -
>
> Key: HBASE-22274
> URL: https://issues.apache.org/jira/browse/HBASE-22274
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 3.0.0, 2.0.0, 1.3.5
>Reporter: Xu Cang
>Assignee: Xu Cang
>Priority: Minor
> Attachments: HBASE-22274-branch-1.001.patch, 
> HBASE-22274-branch-1.002.patch, HBASE-22274-branch-1.003.patch, 
> HBASE-22274-branch-1.003.patch, HBASE-22274-branch-1.004.patch, 
> HBASE-22274-master.001.patch, HBASE-22274-master.002.patch, 
> HBASE-22274-master.002.patch, HBASE-22274-master.003.patch
>
>
> Now we have cell size limit check based on this parameter 
> *hbase.server.keyvalue.maxsize* 
> One case was missing: appending to a cell only take append op's cell size 
> into account against this limit check. we should check against the potential 
> final cell size after the append.'
> It's easy to reproduce this :
>  
> Apply this diff
>  
> {code:java}
> diff --git 
> a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java
>  
> b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java
>  index 5a285ef6ba..8633177ebe 100644 --- 
> a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java
>  +++ 
> b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java
>  @@ -6455,7 +6455,7 
> - t.append(new Append(ROW).addColumn(FAMILY, QUALIFIER, new byte[10 * 
> 1024])); 
> + t.append(new Append(ROW).addColumn(FAMILY, QUALIFIER, new byte[2 * 1024])); 
> {code}
>  
> Fix is to add this check in #reckonDeltas in HRegion class, where we have 
> already got the appended cell's size. 
> Will throw DoNotRetryIOException if checks is failed.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-22274) Cell size limit check on append should consider cell's previous size.

2019-05-09 Thread Andrew Purtell (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-22274?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16836729#comment-16836729
 ] 

Andrew Purtell commented on HBASE-22274:


v4 patch lgtm

The test issue is clearer now.

One minor question, on this change:
{code}
+int newCellSize = CellUtil.estimatedSerializedSizeOf(newCell);
+if (newCellSize > this.maxCellSize) {
+  String msg = "Cell with size " + newCellSize + " exceeds 
limit of " +
+  this.maxCellSize + " bytes";
+  if (LOG.isDebugEnabled()) {
+LOG.debug(msg);
+  }
+  throw new DoNotRetryIOException(msg);
+}
{code}

Should we be logging this at a higher log level, like INFO, or even WARN? 
Whatever we decide here should be consistent with other warnings issued at the 
other sites where we do this size limit test. (I.e. if DEBUG there and we 
decide WARN, update to patch both warnings to WARN, or vice versa.)

> Cell size limit check on append should consider cell's previous size.
> -
>
> Key: HBASE-22274
> URL: https://issues.apache.org/jira/browse/HBASE-22274
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 3.0.0, 2.0.0, 1.3.5
>Reporter: Xu Cang
>Assignee: Xu Cang
>Priority: Minor
> Attachments: HBASE-22274-branch-1.001.patch, 
> HBASE-22274-branch-1.002.patch, HBASE-22274-branch-1.003.patch, 
> HBASE-22274-branch-1.003.patch, HBASE-22274-branch-1.004.patch, 
> HBASE-22274-master.001.patch, HBASE-22274-master.002.patch, 
> HBASE-22274-master.002.patch, HBASE-22274-master.003.patch
>
>
> Now we have cell size limit check based on this parameter 
> *hbase.server.keyvalue.maxsize* 
> One case was missing: appending to a cell only take append op's cell size 
> into account against this limit check. we should check against the potential 
> final cell size after the append.'
> It's easy to reproduce this :
>  
> Apply this diff
>  
> {code:java}
> diff --git 
> a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java
>  
> b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java
>  index 5a285ef6ba..8633177ebe 100644 --- 
> a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java
>  +++ 
> b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java
>  @@ -6455,7 +6455,7 
> - t.append(new Append(ROW).addColumn(FAMILY, QUALIFIER, new byte[10 * 
> 1024])); 
> + t.append(new Append(ROW).addColumn(FAMILY, QUALIFIER, new byte[2 * 1024])); 
> {code}
>  
> Fix is to add this check in #reckonDeltas in HRegion class, where we have 
> already got the appended cell's size. 
> Will throw DoNotRetryIOException if checks is failed.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-22389) Revert HBASE-19275 TestSnapshotFileCache never worked properly on branch-1

2019-05-09 Thread Andrew Purtell (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-22389?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16836728#comment-16836728
 ] 

Andrew Purtell commented on HBASE-22389:


Approved PRs #229 and #230. 

> Revert HBASE-19275 TestSnapshotFileCache never worked properly on branch-1
> --
>
> Key: HBASE-22389
> URL: https://issues.apache.org/jira/browse/HBASE-22389
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 1.5.0
>Reporter: Zach York
>Assignee: Zach York
>Priority: Major
>
> This commit seems broken on branch-1. 
> https://github.com/apache/hbase/commit/d80d3fa45487fedf203c7144e8267753e4d53529#diff-ae046012dd310fc116a729bf36c7df81
>  seems to have fixed the issue that the other HBASE-19275 patches fixed, but 
> when HBASE-19275 was applied to branch-1 it appears to have reverted this 
> code: 
> https://github.com/apache/hbase/commit/e8e4beacb039299a9e56cd355e3ada08784a7acb#diff-ae046012dd310fc116a729bf36c7df81
>  meaning the tests always pass again.
> [~apurtell] [~xucang]



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HBASE-21991) Fix MetaMetrics issues - [Race condition, Faulty remove logic], few improvements

2019-05-09 Thread Sakthi (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-21991?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sakthi updated HBASE-21991:
---
Attachment: hbase-21991.addendum.patch

> Fix MetaMetrics issues - [Race condition, Faulty remove logic], few 
> improvements
> 
>
> Key: HBASE-21991
> URL: https://issues.apache.org/jira/browse/HBASE-21991
> Project: HBase
>  Issue Type: Bug
>  Components: Coprocessors, metrics
>Reporter: Sakthi
>Assignee: Sakthi
>Priority: Major
> Fix For: 3.0.0, 1.5.0, 2.2.0, 1.4.10, 2.3.0
>
> Attachments: hbase-21991.addendum.patch, 
> hbase-21991.branch-1.001.patch, hbase-21991.branch-1.002.patch, 
> hbase-21991.master.001.patch, hbase-21991.master.002.patch, 
> hbase-21991.master.003.patch, hbase-21991.master.004.patch, 
> hbase-21991.master.005.patch, hbase-21991.master.006.patch
>
>
> Here is a list of the issues related to the MetaMetrics implementation:
> +*Bugs*+:
>  # [_Lossy counting for top-k_] *Faulty remove logic of non-eligible meters*: 
> Under certain conditions, we might end up storing/exposing all the meters 
> rather than top-k-ish
>  # MetaMetrics can throw NPE resulting in aborting of the RS because of a 
> *Race Condition*.
> +*Improvements*+:
>  # With high number of regions in the cluster, exposure of metrics for each 
> region blows up the JMX from ~140 Kbs to 100+ Mbs depending on the number of 
> regions. It's better to use *lossy counting to maintain top-k for region 
> metrics* as well.
>  # As the lossy meters do not represent actual counts, I think, it'll be 
> better to *rename the meters to include "lossy" in the name*. It would be 
> more informative while monitoring the metrics and there would be less 
> confusion regarding actual counts to lossy counts.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-22389) Revert HBASE-19275 TestSnapshotFileCache never worked properly on branch-1

2019-05-09 Thread Andrew Purtell (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-22389?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16836726#comment-16836726
 ] 

Andrew Purtell commented on HBASE-22389:


Do not remember how it happened at this point. Let's fix it.
+1


> Revert HBASE-19275 TestSnapshotFileCache never worked properly on branch-1
> --
>
> Key: HBASE-22389
> URL: https://issues.apache.org/jira/browse/HBASE-22389
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 1.5.0
>Reporter: Zach York
>Assignee: Zach York
>Priority: Major
>
> This commit seems broken on branch-1. 
> https://github.com/apache/hbase/commit/d80d3fa45487fedf203c7144e8267753e4d53529#diff-ae046012dd310fc116a729bf36c7df81
>  seems to have fixed the issue that the other HBASE-19275 patches fixed, but 
> when HBASE-19275 was applied to branch-1 it appears to have reverted this 
> code: 
> https://github.com/apache/hbase/commit/e8e4beacb039299a9e56cd355e3ada08784a7acb#diff-ae046012dd310fc116a729bf36c7df81
>  meaning the tests always pass again.
> [~apurtell] [~xucang]



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Comment Edited] (HBASE-21991) Fix MetaMetrics issues - [Race condition, Faulty remove logic], few improvements

2019-05-09 Thread Andrew Purtell (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21991?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16836722#comment-16836722
 ] 

Andrew Purtell edited comment on HBASE-21991 at 5/9/19 9:35 PM:


When resolving please consider copying the below release note text proposal 
into this issue's release note field. In case you want to raise this on dev@ 
first [~busbey] I think we will defer to you to re-resolve this:
{quote}
The class LossyCounting was unintentionally marked Public but was never 
intended to be part of our public API. This oversight has been corrected and 
LossyCounting is now marked as Private and going forward may be subject to 
additional breaking changes or removal without notice. If you have taken a 
dependency on this class we recommend cloning it locally into your project 
before upgrading to this release.
{quote}


was (Author: apurtell):
When resolving please consider copying the below release note text proposal 
into this issue's release note field. In case you want to raise this on dev@ 
first [~busbey] I think we will defer to you to re-resolve this:
{quote}
The class LossyCounting was unintentionally marked Public but was never 
intended to be part of our public API. This oversight has been corrected and 
LossyCounting is now marked as Private and going forward may be subject to 
breaking changes or removal without notice. If you have taken a dependency on 
this class we recommend cloning it locally into your project before further 
changes are made.
{quote}

> Fix MetaMetrics issues - [Race condition, Faulty remove logic], few 
> improvements
> 
>
> Key: HBASE-21991
> URL: https://issues.apache.org/jira/browse/HBASE-21991
> Project: HBase
>  Issue Type: Bug
>  Components: Coprocessors, metrics
>Reporter: Sakthi
>Assignee: Sakthi
>Priority: Major
> Fix For: 3.0.0, 1.5.0, 2.2.0, 1.4.10, 2.3.0
>
> Attachments: hbase-21991.branch-1.001.patch, 
> hbase-21991.branch-1.002.patch, hbase-21991.master.001.patch, 
> hbase-21991.master.002.patch, hbase-21991.master.003.patch, 
> hbase-21991.master.004.patch, hbase-21991.master.005.patch, 
> hbase-21991.master.006.patch
>
>
> Here is a list of the issues related to the MetaMetrics implementation:
> +*Bugs*+:
>  # [_Lossy counting for top-k_] *Faulty remove logic of non-eligible meters*: 
> Under certain conditions, we might end up storing/exposing all the meters 
> rather than top-k-ish
>  # MetaMetrics can throw NPE resulting in aborting of the RS because of a 
> *Race Condition*.
> +*Improvements*+:
>  # With high number of regions in the cluster, exposure of metrics for each 
> region blows up the JMX from ~140 Kbs to 100+ Mbs depending on the number of 
> regions. It's better to use *lossy counting to maintain top-k for region 
> metrics* as well.
>  # As the lossy meters do not represent actual counts, I think, it'll be 
> better to *rename the meters to include "lossy" in the name*. It would be 
> more informative while monitoring the metrics and there would be less 
> confusion regarding actual counts to lossy counts.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-21991) Fix MetaMetrics issues - [Race condition, Faulty remove logic], few improvements

2019-05-09 Thread Andrew Purtell (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21991?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16836722#comment-16836722
 ] 

Andrew Purtell commented on HBASE-21991:


When resolving please consider copying the below release note text proposal 
into this issue's release note field. In case you want to raise this on dev@ 
first [~busbey] I think we will defer to you to re-resolve this:
{quote}
The class LossyCounting was unintentionally marked Public but was never 
intended to be part of our public API. This oversight has been corrected and 
LossyCounting is now marked as Private and going forward may be subject to 
breaking changes or removal without notice. If you have taken a dependency on 
this class we recommend cloning it locally into your project before further 
changes are made.
{quote}

> Fix MetaMetrics issues - [Race condition, Faulty remove logic], few 
> improvements
> 
>
> Key: HBASE-21991
> URL: https://issues.apache.org/jira/browse/HBASE-21991
> Project: HBase
>  Issue Type: Bug
>  Components: Coprocessors, metrics
>Reporter: Sakthi
>Assignee: Sakthi
>Priority: Major
> Fix For: 3.0.0, 1.5.0, 2.2.0, 1.4.10, 2.3.0
>
> Attachments: hbase-21991.branch-1.001.patch, 
> hbase-21991.branch-1.002.patch, hbase-21991.master.001.patch, 
> hbase-21991.master.002.patch, hbase-21991.master.003.patch, 
> hbase-21991.master.004.patch, hbase-21991.master.005.patch, 
> hbase-21991.master.006.patch
>
>
> Here is a list of the issues related to the MetaMetrics implementation:
> +*Bugs*+:
>  # [_Lossy counting for top-k_] *Faulty remove logic of non-eligible meters*: 
> Under certain conditions, we might end up storing/exposing all the meters 
> rather than top-k-ish
>  # MetaMetrics can throw NPE resulting in aborting of the RS because of a 
> *Race Condition*.
> +*Improvements*+:
>  # With high number of regions in the cluster, exposure of metrics for each 
> region blows up the JMX from ~140 Kbs to 100+ Mbs depending on the number of 
> regions. It's better to use *lossy counting to maintain top-k for region 
> metrics* as well.
>  # As the lossy meters do not represent actual counts, I think, it'll be 
> better to *rename the meters to include "lossy" in the name*. It would be 
> more informative while monitoring the metrics and there would be less 
> confusion regarding actual counts to lossy counts.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] [hbase] Apache-HBase commented on issue #229: HBASE-22389 Revert "HBASE-19275 TestSnapshotFileCache never worked pr…

2019-05-09 Thread GitBox
Apache-HBase commented on issue #229: HBASE-22389 Revert "HBASE-19275 
TestSnapshotFileCache never worked pr…
URL: https://github.com/apache/hbase/pull/229#issuecomment-491076582
 
 
   :broken_heart: **-1 overall**
   
   
   
   
   
   
   | Vote | Subsystem | Runtime | Comment |
   |::|--:|:|:|
   | 0 | reexec | 85 | Docker mode activated. |
   ||| _ Prechecks _ |
   | 0 | findbugs | 5 | Findbugs executables are not available. |
   | +1 | hbaseanti | 0 |  Patch does not have any anti-patterns. |
   | +1 | @author | 0 | The patch does not contain any @author tags. |
   | +1 | test4tests | 0 | The patch appears to include 1 new or modified test 
files. |
   ||| _ branch-1 Compile Tests _ |
   | +1 | mvninstall | 241 | branch-1 passed |
   | +1 | compile | 48 | branch-1 passed with JDK v1.8.0_212 |
   | +1 | compile | 44 | branch-1 passed with JDK v1.7.0_222 |
   | +1 | checkstyle | 88 | branch-1 passed |
   | +1 | shadedjars | 221 | branch has no errors when building our shaded 
downstream artifacts. |
   | +1 | javadoc | 47 | branch-1 passed with JDK v1.8.0_212 |
   | +1 | javadoc | 45 | branch-1 passed with JDK v1.7.0_222 |
   ||| _ Patch Compile Tests _ |
   | +1 | mvninstall | 109 | the patch passed |
   | +1 | compile | 40 | the patch passed with JDK v1.8.0_212 |
   | +1 | javac | 40 | the patch passed |
   | +1 | compile | 44 | the patch passed with JDK v1.7.0_222 |
   | +1 | javac | 44 | the patch passed |
   | -1 | checkstyle | 82 | hbase-server: The patch generated 2 new + 4 
unchanged - 0 fixed = 6 total (was 4) |
   | +1 | whitespace | 0 | The patch has no whitespace issues. |
   | +1 | shadedjars | 167 | patch has no errors when building our shaded 
downstream artifacts. |
   | +1 | hadoopcheck | 98 | Patch does not cause any errors with Hadoop 2.7.4. 
|
   | +1 | javadoc | 29 | the patch passed with JDK v1.8.0_212 |
   | +1 | javadoc | 38 | the patch passed with JDK v1.7.0_222 |
   ||| _ Other Tests _ |
   | -1 | unit | 7184 | hbase-server in the patch failed. |
   | +1 | asflicense | 25 | The patch does not generate ASF License warnings. |
   | | | 8746 | |
   
   
   | Reason | Tests |
   |---:|:--|
   | Failed junit tests | hadoop.hbase.master.snapshot.TestSnapshotFileCache |
   
   
   | Subsystem | Report/Notes |
   |--:|:-|
   | Docker | Client=17.05.0-ce Server=17.05.0-ce base: 
https://builds.apache.org/job/HBase-PreCommit-GitHub-PR/job/PR-229/1/artifact/out/Dockerfile
 |
   | GITHUB PR | https://github.com/apache/hbase/pull/229 |
   | Optional Tests |  dupname  asflicense  javac  javadoc  unit  findbugs  
shadedjars  hadoopcheck  hbaseanti  checkstyle  compile  |
   | uname | Linux 4a8c9b6ca0cd 4.4.0-143-generic #169~14.04.2-Ubuntu SMP Wed 
Feb 13 15:00:41 UTC 2019 x86_64 x86_64 x86_64 GNU/Linux |
   | Build tool | maven |
   | Personality | /testptch/patchprocess/precommit/personality/provided.sh |
   | git revision | branch-1 / 56548fc |
   | maven | version: Apache Maven 3.0.5 |
   | Default Java | 1.7.0_222 |
   | Multi-JDK versions |  /usr/lib/jvm/java-8-openjdk-amd64:1.8.0_212 
/usr/lib/jvm/java-7-openjdk-amd64:1.7.0_222 |
   | checkstyle | 
https://builds.apache.org/job/HBase-PreCommit-GitHub-PR/job/PR-229/1/artifact/out/diff-checkstyle-hbase-server.txt
 |
   | unit | 
https://builds.apache.org/job/HBase-PreCommit-GitHub-PR/job/PR-229/1/artifact/out/patch-unit-hbase-server.txt
 |
   |  Test Results | 
https://builds.apache.org/job/HBase-PreCommit-GitHub-PR/job/PR-229/1/testReport/
 |
   | Max. process+thread count | 3672 (vs. ulimit of 1) |
   | modules | C: hbase-server U: hbase-server |
   | Console output | 
https://builds.apache.org/job/HBase-PreCommit-GitHub-PR/job/PR-229/1/console |
   | Powered by | Apache Yetus 0.9.0 http://yetus.apache.org |
   
   
   This message was automatically generated.
   
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (HBASE-22330) Backport HBASE-20724 (Sometimes some compacted storefiles are still opened after region failover) to branch-1

2019-05-09 Thread HBase QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-22330?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16836714#comment-16836714
 ] 

HBase QA commented on HBASE-22330:
--

| (/) *{color:green}+1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  0m 
24s{color} | {color:blue} Docker mode activated. {color} |
|| || || || {color:brown} Prechecks {color} ||
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m  
1s{color} | {color:blue} Findbugs executables are not available. {color} |
| {color:green}+1{color} | {color:green} hbaseanti {color} | {color:green}  0m  
0s{color} | {color:green} Patch does not have any anti-patterns. {color} |
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:green}+1{color} | {color:green} test4tests {color} | {color:green}  0m 
 0s{color} | {color:green} The patch appears to include 1 new or modified test 
files. {color} |
|| || || || {color:brown} branch-1 Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  1m 
57s{color} | {color:green} branch-1 passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
38s{color} | {color:green} branch-1 passed with JDK v1.8.0_212 {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
40s{color} | {color:green} branch-1 passed with JDK v1.7.0_222 {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  1m 
22s{color} | {color:green} branch-1 passed {color} |
| {color:green}+1{color} | {color:green} shadedjars {color} | {color:green}  2m 
48s{color} | {color:green} branch has no errors when building our shaded 
downstream artifacts. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
27s{color} | {color:green} branch-1 passed with JDK v1.8.0_212 {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
37s{color} | {color:green} branch-1 passed with JDK v1.7.0_222 {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  1m 
41s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
37s{color} | {color:green} the patch passed with JDK v1.8.0_212 {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  0m 
37s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
39s{color} | {color:green} the patch passed with JDK v1.7.0_222 {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  0m 
39s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  1m 
22s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} shadedjars {color} | {color:green}  2m 
47s{color} | {color:green} patch has no errors when building our shaded 
downstream artifacts. {color} |
| {color:green}+1{color} | {color:green} hadoopcheck {color} | {color:green}  
1m 40s{color} | {color:green} Patch does not cause any errors with Hadoop 
2.7.4. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
27s{color} | {color:green} the patch passed with JDK v1.8.0_212 {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
37s{color} | {color:green} the patch passed with JDK v1.7.0_222 {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} unit {color} | {color:green} 96m 
30s{color} | {color:green} hbase-server in the patch passed. {color} |
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
19s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black}116m  4s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Docker | Client=17.05.0-ce Server=17.05.0-ce base: 
https://builds.apache.org/job/PreCommit-HBASE-Build/278/artifact/patchprocess/Dockerfile
 |
| JIRA Issue | HBASE-22330 |
| JIRA Patch URL | 
https://issues.apache.org/jira/secure/attachment/12968320/HBASE-22330-addendum.branch-1.patch
 |
| Optional Tests |  dupname  asflicense  javac  javadoc  unit  findbugs  
shadedjars  hadoopcheck  hbaseanti  checkstyle  compile  |
| uname | Linux 9953d0ee7b05 4.4.0-143-generic #169~14.04.2-Ubuntu SMP Wed Feb 
13 15:00:41 

[jira] [Commented] (HBASE-21991) Fix MetaMetrics issues - [Race condition, Faulty remove logic], few improvements

2019-05-09 Thread Sakthi (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21991?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16836715#comment-16836715
 ] 

Sakthi commented on HBASE-21991:


Will upload an addendum patch with 
{code:java}
+@InterfaceAudience.Private
-@InterfaceAudience.Public
public class LossyCounting {{code}
 

> Fix MetaMetrics issues - [Race condition, Faulty remove logic], few 
> improvements
> 
>
> Key: HBASE-21991
> URL: https://issues.apache.org/jira/browse/HBASE-21991
> Project: HBase
>  Issue Type: Bug
>  Components: Coprocessors, metrics
>Reporter: Sakthi
>Assignee: Sakthi
>Priority: Major
> Fix For: 3.0.0, 1.5.0, 2.2.0, 1.4.10, 2.3.0
>
> Attachments: hbase-21991.branch-1.001.patch, 
> hbase-21991.branch-1.002.patch, hbase-21991.master.001.patch, 
> hbase-21991.master.002.patch, hbase-21991.master.003.patch, 
> hbase-21991.master.004.patch, hbase-21991.master.005.patch, 
> hbase-21991.master.006.patch
>
>
> Here is a list of the issues related to the MetaMetrics implementation:
> +*Bugs*+:
>  # [_Lossy counting for top-k_] *Faulty remove logic of non-eligible meters*: 
> Under certain conditions, we might end up storing/exposing all the meters 
> rather than top-k-ish
>  # MetaMetrics can throw NPE resulting in aborting of the RS because of a 
> *Race Condition*.
> +*Improvements*+:
>  # With high number of regions in the cluster, exposure of metrics for each 
> region blows up the JMX from ~140 Kbs to 100+ Mbs depending on the number of 
> regions. It's better to use *lossy counting to maintain top-k for region 
> metrics* as well.
>  # As the lossy meters do not represent actual counts, I think, it'll be 
> better to *rename the meters to include "lossy" in the name*. It would be 
> more informative while monitoring the metrics and there would be less 
> confusion regarding actual counts to lossy counts.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-21991) Fix MetaMetrics issues - [Race condition, Faulty remove logic], few improvements

2019-05-09 Thread Andrew Purtell (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21991?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16836712#comment-16836712
 ] 

Andrew Purtell commented on HBASE-21991:


Let's mark it Private immediate. We can make a release note. It was never 
intended to be part of our public API, that seems pretty clear, and is trivial 
to clone if someone decided to depend on it. 

> Fix MetaMetrics issues - [Race condition, Faulty remove logic], few 
> improvements
> 
>
> Key: HBASE-21991
> URL: https://issues.apache.org/jira/browse/HBASE-21991
> Project: HBase
>  Issue Type: Bug
>  Components: Coprocessors, metrics
>Reporter: Sakthi
>Assignee: Sakthi
>Priority: Major
> Fix For: 3.0.0, 1.5.0, 2.2.0, 1.4.10, 2.3.0
>
> Attachments: hbase-21991.branch-1.001.patch, 
> hbase-21991.branch-1.002.patch, hbase-21991.master.001.patch, 
> hbase-21991.master.002.patch, hbase-21991.master.003.patch, 
> hbase-21991.master.004.patch, hbase-21991.master.005.patch, 
> hbase-21991.master.006.patch
>
>
> Here is a list of the issues related to the MetaMetrics implementation:
> +*Bugs*+:
>  # [_Lossy counting for top-k_] *Faulty remove logic of non-eligible meters*: 
> Under certain conditions, we might end up storing/exposing all the meters 
> rather than top-k-ish
>  # MetaMetrics can throw NPE resulting in aborting of the RS because of a 
> *Race Condition*.
> +*Improvements*+:
>  # With high number of regions in the cluster, exposure of metrics for each 
> region blows up the JMX from ~140 Kbs to 100+ Mbs depending on the number of 
> regions. It's better to use *lossy counting to maintain top-k for region 
> metrics* as well.
>  # As the lossy meters do not represent actual counts, I think, it'll be 
> better to *rename the meters to include "lossy" in the name*. It would be 
> more informative while monitoring the metrics and there would be less 
> confusion regarding actual counts to lossy counts.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-21800) RegionServer aborted due to NPE from MetaTableMetrics coprocessor

2019-05-09 Thread Andrew Purtell (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21800?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16836713#comment-16836713
 ] 

Andrew Purtell commented on HBASE-21800:


Seems we are going to address the issues with LossyCounting in the context of 
HBASE-21991 so will resolve this one.

> RegionServer aborted due to NPE from MetaTableMetrics coprocessor
> -
>
> Key: HBASE-21800
> URL: https://issues.apache.org/jira/browse/HBASE-21800
> Project: HBase
>  Issue Type: Bug
>  Components: Coprocessors, meta, metrics, Operability
>Reporter: Sakthi
>Assignee: Sakthi
>Priority: Critical
>  Labels: Meta
> Fix For: 3.0.0, 1.5.0, 2.2.0, 1.4.10, 2.3.0
>
> Attachments: hbase-21800.branch-1.001.patch, 
> hbase-21800.branch-1.002.patch, hbase-21800.branch-1.003.patch, 
> hbase-21800.branch-1.004.patch, hbase-21800.master.001.patch, 
> hbase-21800.master.002.patch, hbase-21800.master.003.patch
>
>
> I was just playing around the code, trying to capture "Top k" table metrics 
> from MetaMetrics, when I bumped into this issue. Though currently we are not 
> capturing "Top K" table metrics, but we can encounter this issue because of 
> the "Top k Clients" that is implemented using the LossyAlgo.
>  
> RegionServer gets aborted due to a NPE from MetaTableMetrics coprocessor. The 
> log looks somewhat like this:
> {code:java}
> 2019-01-28 23:31:10,311 ERROR 
> [RpcServer.priority.FPBQ.Fifo.handler=19,queue=1,port=16020] 
> coprocessor.CoprocessorHost: The coprocessor 
> org.apache.hadoop.hbase.coprocessor.MetaTableMetrics threw 
> java.lang.NullPointerException
> java.lang.NullPointerException
>   at 
> org.apache.hadoop.hbase.coprocessor.MetaTableMetrics$ExampleRegionObserverMeta.markMeterIfPresent(MetaTableMetrics.java:123)
>   at 
> org.apache.hadoop.hbase.coprocessor.MetaTableMetrics$ExampleRegionObserverMeta.tableMetricRegisterAndMark2(MetaTableMetrics.java:233)
>   at 
> org.apache.hadoop.hbase.coprocessor.MetaTableMetrics$ExampleRegionObserverMeta.preGetOp(MetaTableMetrics.java:82)
>   at 
> org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost$19.call(RegionCoprocessorHost.java:840)
>   at 
> org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost$19.call(RegionCoprocessorHost.java:837)
>   at 
> org.apache.hadoop.hbase.coprocessor.CoprocessorHost$ObserverOperationWithoutResult.callObserver(CoprocessorHost.java:551)
>   at 
> org.apache.hadoop.hbase.coprocessor.CoprocessorHost.execOperation(CoprocessorHost.java:625)
>   at 
> org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.preGet(RegionCoprocessorHost.java:837)
>   at 
> org.apache.hadoop.hbase.regionserver.RSRpcServices.get(RSRpcServices.java:2608)
>   at 
> org.apache.hadoop.hbase.regionserver.RSRpcServices.get(RSRpcServices.java:2547)
>   at 
> org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:41998)
>   at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:413)
>   at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:130)
>   at 
> org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:324)
>   at 
> org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:304)
> 2019-01-28 23:31:10,314 ERROR 
> [RpcServer.priority.FPBQ.Fifo.handler=19,queue=1,port=16020] 
> regionserver.HRegionServer: * ABORTING region server 
> 10.0.0.24,16020,1548747043814: The coprocessor 
> org.apache.hadoop.hbase.coprocessor.MetaTableMetrics threw 
> java.lang.NullPointerException *
> java.lang.NullPointerException
>   at 
> org.apache.hadoop.hbase.coprocessor.MetaTableMetrics$ExampleRegionObserverMeta.markMeterIfPresent(MetaTableMetrics.java:123)
>   at 
> org.apache.hadoop.hbase.coprocessor.MetaTableMetrics$ExampleRegionObserverMeta.tableMetricRegisterAndMark2(MetaTableMetrics.java:233)
>   at 
> org.apache.hadoop.hbase.coprocessor.MetaTableMetrics$ExampleRegionObserverMeta.preGetOp(MetaTableMetrics.java:82)
>   at 
> org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost$19.call(RegionCoprocessorHost.java:840)
>   at 
> org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost$19.call(RegionCoprocessorHost.java:837)
>   at 
> org.apache.hadoop.hbase.coprocessor.CoprocessorHost$ObserverOperationWithoutResult.callObserver(CoprocessorHost.java:551)
>   at 
> org.apache.hadoop.hbase.coprocessor.CoprocessorHost.execOperation(CoprocessorHost.java:625)
>   at 
> org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.preGet(RegionCoprocessorHost.java:837)
>   at 
> org.apache.hadoop.hbase.regionserver.RSRpcServices.get(RSRpcServices.java:2608)
>   at 
> 

[jira] [Commented] (HBASE-21991) Fix MetaMetrics issues - [Race condition, Faulty remove logic], few improvements

2019-05-09 Thread Sakthi (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21991?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16836711#comment-16836711
 ] 

Sakthi commented on HBASE-21991:


[~busbey] sounds good to me.

> Fix MetaMetrics issues - [Race condition, Faulty remove logic], few 
> improvements
> 
>
> Key: HBASE-21991
> URL: https://issues.apache.org/jira/browse/HBASE-21991
> Project: HBase
>  Issue Type: Bug
>  Components: Coprocessors, metrics
>Reporter: Sakthi
>Assignee: Sakthi
>Priority: Major
> Fix For: 3.0.0, 1.5.0, 2.2.0, 1.4.10, 2.3.0
>
> Attachments: hbase-21991.branch-1.001.patch, 
> hbase-21991.branch-1.002.patch, hbase-21991.master.001.patch, 
> hbase-21991.master.002.patch, hbase-21991.master.003.patch, 
> hbase-21991.master.004.patch, hbase-21991.master.005.patch, 
> hbase-21991.master.006.patch
>
>
> Here is a list of the issues related to the MetaMetrics implementation:
> +*Bugs*+:
>  # [_Lossy counting for top-k_] *Faulty remove logic of non-eligible meters*: 
> Under certain conditions, we might end up storing/exposing all the meters 
> rather than top-k-ish
>  # MetaMetrics can throw NPE resulting in aborting of the RS because of a 
> *Race Condition*.
> +*Improvements*+:
>  # With high number of regions in the cluster, exposure of metrics for each 
> region blows up the JMX from ~140 Kbs to 100+ Mbs depending on the number of 
> regions. It's better to use *lossy counting to maintain top-k for region 
> metrics* as well.
>  # As the lossy meters do not represent actual counts, I think, it'll be 
> better to *rename the meters to include "lossy" in the name*. It would be 
> more informative while monitoring the metrics and there would be less 
> confusion regarding actual counts to lossy counts.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-22330) Backport HBASE-20724 (Sometimes some compacted storefiles are still opened after region failover) to branch-1

2019-05-09 Thread Andrew Purtell (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-22330?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16836718#comment-16836718
 ] 

Andrew Purtell commented on HBASE-22330:


+1
Please commit addendum.

> Backport HBASE-20724 (Sometimes some compacted storefiles are still opened 
> after region failover) to branch-1
> -
>
> Key: HBASE-22330
> URL: https://issues.apache.org/jira/browse/HBASE-22330
> Project: HBase
>  Issue Type: Sub-task
>  Components: Compaction, regionserver
>Affects Versions: 1.5.0, 1.4.9, 1.3.4
>Reporter: Andrew Purtell
>Assignee: Abhishek Singh Chouhan
>Priority: Major
> Fix For: 1.5.0, 1.3.5, 1.4.11
>
> Attachments: HBASE-22330-addendum.branch-1.patch, 
> HBASE-22330.branch-1.001.patch, HBASE-22330.branch-1.002.patch, 
> HBASE-22330.branch-1.3.001.patch
>
>
> There appears to be a race condition between close and split which when 
> combined with a side effect of HBASE-20704, leads to the parent region store 
> files getting archived and cleared while daughter regions still have 
> references to those parent region store files.
> Here is the timeline of events observed for an affected region:
>  # RS1 faces ZooKeeper connectivity issue for master node and starts shutting 
> itself down. As part of this it starts to close the store and clean up the 
> compacted files (File A)
>  # Master starts bulk assigning regions and assign parent region to RS2
>  # Region opens on RS2 and ends up opening compacted store file(s) (suspect 
> this is due to HBASE-20724)
>  # Now split happens and daughter regions open on RS2 and try to run a 
> compaction as part of post open
>  # Split request at this point is complete. However now archiving proceeds on 
> RS1 and ends up archiving the store file that is referenced by the daughter. 
> Compaction fails due to FileNotFoundException and all subsequent attempts to 
> open the region will fail until manual resolution.
> We think having HBASE-20724 would help in such situations since we won't end 
> up loading compacted store files in the first place. 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Resolved] (HBASE-21800) RegionServer aborted due to NPE from MetaTableMetrics coprocessor

2019-05-09 Thread Andrew Purtell (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-21800?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew Purtell resolved HBASE-21800.

Resolution: Fixed

> RegionServer aborted due to NPE from MetaTableMetrics coprocessor
> -
>
> Key: HBASE-21800
> URL: https://issues.apache.org/jira/browse/HBASE-21800
> Project: HBase
>  Issue Type: Bug
>  Components: Coprocessors, meta, metrics, Operability
>Reporter: Sakthi
>Assignee: Sakthi
>Priority: Critical
>  Labels: Meta
> Fix For: 3.0.0, 1.5.0, 2.2.0, 2.3.0, 1.4.10
>
> Attachments: hbase-21800.branch-1.001.patch, 
> hbase-21800.branch-1.002.patch, hbase-21800.branch-1.003.patch, 
> hbase-21800.branch-1.004.patch, hbase-21800.master.001.patch, 
> hbase-21800.master.002.patch, hbase-21800.master.003.patch
>
>
> I was just playing around the code, trying to capture "Top k" table metrics 
> from MetaMetrics, when I bumped into this issue. Though currently we are not 
> capturing "Top K" table metrics, but we can encounter this issue because of 
> the "Top k Clients" that is implemented using the LossyAlgo.
>  
> RegionServer gets aborted due to a NPE from MetaTableMetrics coprocessor. The 
> log looks somewhat like this:
> {code:java}
> 2019-01-28 23:31:10,311 ERROR 
> [RpcServer.priority.FPBQ.Fifo.handler=19,queue=1,port=16020] 
> coprocessor.CoprocessorHost: The coprocessor 
> org.apache.hadoop.hbase.coprocessor.MetaTableMetrics threw 
> java.lang.NullPointerException
> java.lang.NullPointerException
>   at 
> org.apache.hadoop.hbase.coprocessor.MetaTableMetrics$ExampleRegionObserverMeta.markMeterIfPresent(MetaTableMetrics.java:123)
>   at 
> org.apache.hadoop.hbase.coprocessor.MetaTableMetrics$ExampleRegionObserverMeta.tableMetricRegisterAndMark2(MetaTableMetrics.java:233)
>   at 
> org.apache.hadoop.hbase.coprocessor.MetaTableMetrics$ExampleRegionObserverMeta.preGetOp(MetaTableMetrics.java:82)
>   at 
> org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost$19.call(RegionCoprocessorHost.java:840)
>   at 
> org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost$19.call(RegionCoprocessorHost.java:837)
>   at 
> org.apache.hadoop.hbase.coprocessor.CoprocessorHost$ObserverOperationWithoutResult.callObserver(CoprocessorHost.java:551)
>   at 
> org.apache.hadoop.hbase.coprocessor.CoprocessorHost.execOperation(CoprocessorHost.java:625)
>   at 
> org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.preGet(RegionCoprocessorHost.java:837)
>   at 
> org.apache.hadoop.hbase.regionserver.RSRpcServices.get(RSRpcServices.java:2608)
>   at 
> org.apache.hadoop.hbase.regionserver.RSRpcServices.get(RSRpcServices.java:2547)
>   at 
> org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:41998)
>   at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:413)
>   at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:130)
>   at 
> org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:324)
>   at 
> org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:304)
> 2019-01-28 23:31:10,314 ERROR 
> [RpcServer.priority.FPBQ.Fifo.handler=19,queue=1,port=16020] 
> regionserver.HRegionServer: * ABORTING region server 
> 10.0.0.24,16020,1548747043814: The coprocessor 
> org.apache.hadoop.hbase.coprocessor.MetaTableMetrics threw 
> java.lang.NullPointerException *
> java.lang.NullPointerException
>   at 
> org.apache.hadoop.hbase.coprocessor.MetaTableMetrics$ExampleRegionObserverMeta.markMeterIfPresent(MetaTableMetrics.java:123)
>   at 
> org.apache.hadoop.hbase.coprocessor.MetaTableMetrics$ExampleRegionObserverMeta.tableMetricRegisterAndMark2(MetaTableMetrics.java:233)
>   at 
> org.apache.hadoop.hbase.coprocessor.MetaTableMetrics$ExampleRegionObserverMeta.preGetOp(MetaTableMetrics.java:82)
>   at 
> org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost$19.call(RegionCoprocessorHost.java:840)
>   at 
> org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost$19.call(RegionCoprocessorHost.java:837)
>   at 
> org.apache.hadoop.hbase.coprocessor.CoprocessorHost$ObserverOperationWithoutResult.callObserver(CoprocessorHost.java:551)
>   at 
> org.apache.hadoop.hbase.coprocessor.CoprocessorHost.execOperation(CoprocessorHost.java:625)
>   at 
> org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.preGet(RegionCoprocessorHost.java:837)
>   at 
> org.apache.hadoop.hbase.regionserver.RSRpcServices.get(RSRpcServices.java:2608)
>   at 
> org.apache.hadoop.hbase.regionserver.RSRpcServices.get(RSRpcServices.java:2547)
>   at 
> 

[jira] [Commented] (HBASE-21536) Fix completebulkload usage instructions

2019-05-09 Thread Artem Ervits (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21536?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16836691#comment-16836691
 ] 

Artem Ervits commented on HBASE-21536:
--

patch v2 applies to current master, additionally clarified invocation for both 
legacy and hbase 3.0. [~psomogyi]

> Fix completebulkload usage instructions
> ---
>
> Key: HBASE-21536
> URL: https://issues.apache.org/jira/browse/HBASE-21536
> Project: HBase
>  Issue Type: Task
>  Components: documentation, mapreduce
>Reporter: Artem Ervits
>Assignee: Artem Ervits
>Priority: Trivial
> Attachments: HBASE-21536.v01.patch, HBASE-21536.v02.patch
>
>
> Usage information upon invoking LoadIncrementalHFiles is misleading and 
> error-prone.
> {code:java}
> usage: completebulkload /path/to/hfileoutputformat-output tablename -loadTable
> -Dcreate.table=no - can be used to avoid creation of table by this tool
> Note: if you set this to 'no', then the target table must already exist in 
> HBase
> -loadTable implies your baseDirectory to store file has a depth of 3 ,you 
> must have an existing table
> -Dignore.unmatched.families=yes - can be used to ignore unmatched column 
> families{code}
> in case of invoking the class via hbase command, completebulkload argument is 
> unnecessary and only required via hadoop jar invocation. This is also an 
> attempt to clarify where <-loadTable> and <-Dargs> arguments go on the 
> command line.
> Furthermore, since LoadIncrementalHFiles was recently moved out of 
> hbase-server.jar to hbase-mapreduce, updating ref guide to demonstrate as 
> such.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-21991) Fix MetaMetrics issues - [Race condition, Faulty remove logic], few improvements

2019-05-09 Thread Sean Busbey (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21991?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16836686#comment-16836686
 ] 

Sean Busbey commented on HBASE-21991:
-

Agreed LossyCounting shouldn't be public. Maybe a quick DISCUSS on dev@ about 
cutting our losses and just release noting making it Private without any 
deprecation?

> Fix MetaMetrics issues - [Race condition, Faulty remove logic], few 
> improvements
> 
>
> Key: HBASE-21991
> URL: https://issues.apache.org/jira/browse/HBASE-21991
> Project: HBase
>  Issue Type: Bug
>  Components: Coprocessors, metrics
>Reporter: Sakthi
>Assignee: Sakthi
>Priority: Major
> Fix For: 3.0.0, 1.5.0, 2.2.0, 1.4.10, 2.3.0
>
> Attachments: hbase-21991.branch-1.001.patch, 
> hbase-21991.branch-1.002.patch, hbase-21991.master.001.patch, 
> hbase-21991.master.002.patch, hbase-21991.master.003.patch, 
> hbase-21991.master.004.patch, hbase-21991.master.005.patch, 
> hbase-21991.master.006.patch
>
>
> Here is a list of the issues related to the MetaMetrics implementation:
> +*Bugs*+:
>  # [_Lossy counting for top-k_] *Faulty remove logic of non-eligible meters*: 
> Under certain conditions, we might end up storing/exposing all the meters 
> rather than top-k-ish
>  # MetaMetrics can throw NPE resulting in aborting of the RS because of a 
> *Race Condition*.
> +*Improvements*+:
>  # With high number of regions in the cluster, exposure of metrics for each 
> region blows up the JMX from ~140 Kbs to 100+ Mbs depending on the number of 
> regions. It's better to use *lossy counting to maintain top-k for region 
> metrics* as well.
>  # As the lossy meters do not represent actual counts, I think, it'll be 
> better to *rename the meters to include "lossy" in the name*. It would be 
> more informative while monitoring the metrics and there would be less 
> confusion regarding actual counts to lossy counts.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HBASE-21536) Fix completebulkload usage instructions

2019-05-09 Thread Artem Ervits (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-21536?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Artem Ervits updated HBASE-21536:
-
Attachment: HBASE-21536.v02.patch

> Fix completebulkload usage instructions
> ---
>
> Key: HBASE-21536
> URL: https://issues.apache.org/jira/browse/HBASE-21536
> Project: HBase
>  Issue Type: Task
>  Components: documentation, mapreduce
>Reporter: Artem Ervits
>Assignee: Artem Ervits
>Priority: Trivial
> Attachments: HBASE-21536.v01.patch, HBASE-21536.v02.patch
>
>
> Usage information upon invoking LoadIncrementalHFiles is misleading and 
> error-prone.
> {code:java}
> usage: completebulkload /path/to/hfileoutputformat-output tablename -loadTable
> -Dcreate.table=no - can be used to avoid creation of table by this tool
> Note: if you set this to 'no', then the target table must already exist in 
> HBase
> -loadTable implies your baseDirectory to store file has a depth of 3 ,you 
> must have an existing table
> -Dignore.unmatched.families=yes - can be used to ignore unmatched column 
> families{code}
> in case of invoking the class via hbase command, completebulkload argument is 
> unnecessary and only required via hadoop jar invocation. This is also an 
> attempt to clarify where <-loadTable> and <-Dargs> arguments go on the 
> command line.
> Furthermore, since LoadIncrementalHFiles was recently moved out of 
> hbase-server.jar to hbase-mapreduce, updating ref guide to demonstrate as 
> such.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HBASE-21536) Fix completebulkload usage instructions

2019-05-09 Thread Artem Ervits (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-21536?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Artem Ervits updated HBASE-21536:
-
Status: Patch Available  (was: Open)

> Fix completebulkload usage instructions
> ---
>
> Key: HBASE-21536
> URL: https://issues.apache.org/jira/browse/HBASE-21536
> Project: HBase
>  Issue Type: Task
>  Components: documentation, mapreduce
>Reporter: Artem Ervits
>Assignee: Artem Ervits
>Priority: Trivial
> Attachments: HBASE-21536.v01.patch, HBASE-21536.v02.patch
>
>
> Usage information upon invoking LoadIncrementalHFiles is misleading and 
> error-prone.
> {code:java}
> usage: completebulkload /path/to/hfileoutputformat-output tablename -loadTable
> -Dcreate.table=no - can be used to avoid creation of table by this tool
> Note: if you set this to 'no', then the target table must already exist in 
> HBase
> -loadTable implies your baseDirectory to store file has a depth of 3 ,you 
> must have an existing table
> -Dignore.unmatched.families=yes - can be used to ignore unmatched column 
> families{code}
> in case of invoking the class via hbase command, completebulkload argument is 
> unnecessary and only required via hadoop jar invocation. This is also an 
> attempt to clarify where <-loadTable> and <-Dargs> arguments go on the 
> command line.
> Furthermore, since LoadIncrementalHFiles was recently moved out of 
> hbase-server.jar to hbase-mapreduce, updating ref guide to demonstrate as 
> such.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HBASE-21536) Fix completebulkload usage instructions

2019-05-09 Thread Artem Ervits (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-21536?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Artem Ervits updated HBASE-21536:
-
Status: Open  (was: Patch Available)

> Fix completebulkload usage instructions
> ---
>
> Key: HBASE-21536
> URL: https://issues.apache.org/jira/browse/HBASE-21536
> Project: HBase
>  Issue Type: Task
>  Components: documentation, mapreduce
>Reporter: Artem Ervits
>Assignee: Artem Ervits
>Priority: Trivial
> Attachments: HBASE-21536.v01.patch, HBASE-21536.v02.patch
>
>
> Usage information upon invoking LoadIncrementalHFiles is misleading and 
> error-prone.
> {code:java}
> usage: completebulkload /path/to/hfileoutputformat-output tablename -loadTable
> -Dcreate.table=no - can be used to avoid creation of table by this tool
> Note: if you set this to 'no', then the target table must already exist in 
> HBase
> -loadTable implies your baseDirectory to store file has a depth of 3 ,you 
> must have an existing table
> -Dignore.unmatched.families=yes - can be used to ignore unmatched column 
> families{code}
> in case of invoking the class via hbase command, completebulkload argument is 
> unnecessary and only required via hadoop jar invocation. This is also an 
> attempt to clarify where <-loadTable> and <-Dargs> arguments go on the 
> command line.
> Furthermore, since LoadIncrementalHFiles was recently moved out of 
> hbase-server.jar to hbase-mapreduce, updating ref guide to demonstrate as 
> such.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-21991) Fix MetaMetrics issues - [Race condition, Faulty remove logic], few improvements

2019-05-09 Thread Sean Busbey (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-21991?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16836689#comment-16836689
 ] 

Sean Busbey commented on HBASE-21991:
-

sorry for the unclear phrasing. to be clear, I plan to do the above unless 
someone complains here.

> Fix MetaMetrics issues - [Race condition, Faulty remove logic], few 
> improvements
> 
>
> Key: HBASE-21991
> URL: https://issues.apache.org/jira/browse/HBASE-21991
> Project: HBase
>  Issue Type: Bug
>  Components: Coprocessors, metrics
>Reporter: Sakthi
>Assignee: Sakthi
>Priority: Major
> Fix For: 3.0.0, 1.5.0, 2.2.0, 1.4.10, 2.3.0
>
> Attachments: hbase-21991.branch-1.001.patch, 
> hbase-21991.branch-1.002.patch, hbase-21991.master.001.patch, 
> hbase-21991.master.002.patch, hbase-21991.master.003.patch, 
> hbase-21991.master.004.patch, hbase-21991.master.005.patch, 
> hbase-21991.master.006.patch
>
>
> Here is a list of the issues related to the MetaMetrics implementation:
> +*Bugs*+:
>  # [_Lossy counting for top-k_] *Faulty remove logic of non-eligible meters*: 
> Under certain conditions, we might end up storing/exposing all the meters 
> rather than top-k-ish
>  # MetaMetrics can throw NPE resulting in aborting of the RS because of a 
> *Race Condition*.
> +*Improvements*+:
>  # With high number of regions in the cluster, exposure of metrics for each 
> region blows up the JMX from ~140 Kbs to 100+ Mbs depending on the number of 
> regions. It's better to use *lossy counting to maintain top-k for region 
> metrics* as well.
>  # As the lossy meters do not represent actual counts, I think, it'll be 
> better to *rename the meters to include "lossy" in the name*. It would be 
> more informative while monitoring the metrics and there would be less 
> confusion regarding actual counts to lossy counts.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-22358) Change rubocop configuration for method length

2019-05-09 Thread Hudson (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-22358?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16836684#comment-16836684
 ] 

Hudson commented on HBASE-22358:


Results for branch branch-2.2
[build #247 on 
builds.a.o|https://builds.apache.org/job/HBase%20Nightly/job/branch-2.2/247/]: 
(x) *{color:red}-1 overall{color}*

details (if available):

(/) {color:green}+1 general checks{color}
-- For more information [see general 
report|https://builds.apache.org/job/HBase%20Nightly/job/branch-2.2/247//General_Nightly_Build_Report/]




(/) {color:green}+1 jdk8 hadoop2 checks{color}
-- For more information [see jdk8 (hadoop2) 
report|https://builds.apache.org/job/HBase%20Nightly/job/branch-2.2/247//JDK8_Nightly_Build_Report_(Hadoop2)/]


(x) {color:red}-1 jdk8 hadoop3 checks{color}
-- For more information [see jdk8 (hadoop3) 
report|https://builds.apache.org/job/HBase%20Nightly/job/branch-2.2/247//JDK8_Nightly_Build_Report_(Hadoop3)/]


(/) {color:green}+1 source release artifact{color}
-- See build output for details.


(/) {color:green}+1 client integration test{color}


> Change rubocop configuration for method length
> --
>
> Key: HBASE-22358
> URL: https://issues.apache.org/jira/browse/HBASE-22358
> Project: HBase
>  Issue Type: Improvement
>  Components: community, shell
>Reporter: Jan Hentschel
>Assignee: Murtaza Hassan
>Priority: Minor
>  Labels: beginner, beginners
> Fix For: 3.0.0, 1.5.0, 2.3.0, 2.0.6, 2.1.5, 2.2.1, 1.3.5, 1.4.11
>
>
> rubocop currently uses a maximum method length for the Ruby code of 10, which 
> is way too restrictive. In Checkstyle we're using 150 lines per method. Don't 
> know if it needs to be that much, but something between 50 and 75 seems to be 
> more realistic, especially for test cases.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-22324) loss a mass of data when the sequenceId of cells greater than Integer.Max, because MemStoreMergerSegmentsIterator can not merge segments

2019-05-09 Thread Hudson (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-22324?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16836683#comment-16836683
 ] 

Hudson commented on HBASE-22324:


Results for branch branch-2.2
[build #247 on 
builds.a.o|https://builds.apache.org/job/HBase%20Nightly/job/branch-2.2/247/]: 
(x) *{color:red}-1 overall{color}*

details (if available):

(/) {color:green}+1 general checks{color}
-- For more information [see general 
report|https://builds.apache.org/job/HBase%20Nightly/job/branch-2.2/247//General_Nightly_Build_Report/]




(/) {color:green}+1 jdk8 hadoop2 checks{color}
-- For more information [see jdk8 (hadoop2) 
report|https://builds.apache.org/job/HBase%20Nightly/job/branch-2.2/247//JDK8_Nightly_Build_Report_(Hadoop2)/]


(x) {color:red}-1 jdk8 hadoop3 checks{color}
-- For more information [see jdk8 (hadoop3) 
report|https://builds.apache.org/job/HBase%20Nightly/job/branch-2.2/247//JDK8_Nightly_Build_Report_(Hadoop3)/]


(/) {color:green}+1 source release artifact{color}
-- See build output for details.


(/) {color:green}+1 client integration test{color}


>  loss a mass of data when the sequenceId of cells greater than Integer.Max, 
> because MemStoreMergerSegmentsIterator can not merge segments 
> --
>
> Key: HBASE-22324
> URL: https://issues.apache.org/jira/browse/HBASE-22324
> Project: HBase
>  Issue Type: Bug
>  Components: in-memory-compaction
>Affects Versions: 2.1.0, 2.2.0
>Reporter: chenyang
>Assignee: ChenYang
>Priority: Blocker
>  Labels: patch
> Fix For: 3.0.0, 2.1.0, 2.2.0, 2.3.0, 2.0.6
>
> Attachments: HBASE-22324.branch-2.1.0005.patch, 
> HBASE-22324.branch-2.1.0006.patch
>
>
> if your memstore type is CompactingMemStore,MemStoreMergerSegmentsIterator 
> can not merge memstore segments when the seqId of cells greater than 
> Integer.Max, as a result, lossing a mass of data. the reason is that 
> MemStoreMergerSegmentsIterator use Integer.Max as readPt when create Scanner, 
>  but the seqId of cell  may be greater than Integer.MAX_VALUE,  it`s type is 
> long.   code as below:
> {code:java}
> public MemStoreMergerSegmentsIterator(List segments, 
> CellComparator comparator,
> int compactionKVMax) throws IOException {
>   super(compactionKVMax);
>   // create the list of scanners to traverse over all the data
>   // no dirty reads here as these are immutable segments
>   AbstractMemStore.addToScanners(segments, Integer.MAX_VALUE, scanners); 
> //bug, should use Long.MAX_VALUE
>   heap = new KeyValueHeap(scanners, comparator);
> }
> SegmentScanner.java code as below
> protected void updateCurrent() {
>   Cell startKV = current;
>   Cell next = null;
>   try {
> while (iter.hasNext()) {
>   next = iter.next();
>   // here, if seqId>readPoint(Integer.MAX_VALUE), never read cell, as a 
> result, lossing lots of cells
>   if (next.getSequenceId() <= this.readPoint) {
> current = next;
> return;// skip irrelevant versions
>   }
>   if (stopSkippingKVsIfNextRow &&   // for backwardSeek() stay in the
>   startKV != null &&// boundaries of a single row
>   segment.compareRows(next, startKV) > 0) {
> current = null;
> return;
>   }
> } // end of while
> current = null; // nothing found
>   } finally {
> if (next != null) {
>   // in all cases, remember the last KV we iterated to, needed for 
> reseek()
>   last = next;
> }
>   }
> }
> MemStoreCompactorSegmentsIterator has the same bug
> public MemStoreCompactorSegmentsIterator(List segments,
> CellComparator comparator, int compactionKVMax, HStore store) throws 
> IOException {
>   super(compactionKVMax);
>   List scanners = new ArrayList();
>   AbstractMemStore.addToScanners(segments, Integer.MAX_VALUE, scanners);   
> //bug, should use Long.MAX_VALUE
>   // build the scanner based on Query Matcher
>   // reinitialize the compacting scanner for each instance of iterator
>   compactingScanner = createScanner(store, scanners);
>   refillKVS();
> }{code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-22389) Revert HBASE-19275 TestSnapshotFileCache never worked properly on branch-1

2019-05-09 Thread Zach York (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-22389?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16836645#comment-16836645
 ] 

Zach York commented on HBASE-22389:
---

looks to also be an issue with branch 1.4

> Revert HBASE-19275 TestSnapshotFileCache never worked properly on branch-1
> --
>
> Key: HBASE-22389
> URL: https://issues.apache.org/jira/browse/HBASE-22389
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 1.5.0
>Reporter: Zach York
>Assignee: Zach York
>Priority: Major
>
> This commit seems broken on branch-1. 
> https://github.com/apache/hbase/commit/d80d3fa45487fedf203c7144e8267753e4d53529#diff-ae046012dd310fc116a729bf36c7df81
>  seems to have fixed the issue that the other HBASE-19275 patches fixed, but 
> when HBASE-19275 was applied to branch-1 it appears to have reverted this 
> code: 
> https://github.com/apache/hbase/commit/e8e4beacb039299a9e56cd355e3ada08784a7acb#diff-ae046012dd310fc116a729bf36c7df81
>  meaning the tests always pass again.
> [~apurtell] [~xucang]



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HBASE-22390) backport HBASE-22190 to branch-1

2019-05-09 Thread Zach York (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-22390?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zach York updated HBASE-22390:
--
Status: Patch Available  (was: Open)

> backport HBASE-22190 to branch-1
> 
>
> Key: HBASE-22390
> URL: https://issues.apache.org/jira/browse/HBASE-22390
> Project: HBase
>  Issue Type: Bug
>Affects Versions: 1.4.10, 1.5.0
>Reporter: Zach York
>Assignee: Zach York
>Priority: Major
>
> HBASE-22190 was only applied to branch-2+ while the bug exists in branch-1 as 
> well.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] [hbase] z-york commented on issue #229: HBASE-22389 Revert "HBASE-19275 TestSnapshotFileCache never worked pr…

2019-05-09 Thread GitBox
z-york commented on issue #229: HBASE-22389 Revert "HBASE-19275 
TestSnapshotFileCache never worked pr…
URL: https://github.com/apache/hbase/pull/229#issuecomment-491035742
 
 
   https://github.com/apache/hbase/pull/230 should be merged first since it 
will cause test failures otherwise (as mentioned previously)


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [hbase] z-york commented on issue #230: HBASE-22390 Backport HBASE-22190 to branch-1

2019-05-09 Thread GitBox
z-york commented on issue #230: HBASE-22390 Backport HBASE-22190 to branch-1
URL: https://github.com/apache/hbase/pull/230#issuecomment-491035414
 
 
   Also please advise if I attributed this incorrectly, it's a little wonky 
with a PR.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [hbase] z-york commented on issue #230: HBASE-22390 Backport HBASE-22190 to branch-1

2019-05-09 Thread GitBox
z-york commented on issue #230: HBASE-22390 Backport HBASE-22190 to branch-1
URL: https://github.com/apache/hbase/pull/230#issuecomment-491035174
 
 
   @Apache9 @apurtell (for visibility since this fixes HBASE-21070)


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [hbase] z-york opened a new pull request #230: HBASE-22390 Backport HBASE-22190 to branch-1

2019-05-09 Thread GitBox
z-york opened a new pull request #230: HBASE-22390 Backport HBASE-22190 to 
branch-1
URL: https://github.com/apache/hbase/pull/230
 
 
HBASE-22190 SnapshotFileCache may fail to load the correct snapshot file 
list when there is an on-going snapshot operation


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (HBASE-22220) Release hbase-connectors-1.0.0

2019-05-09 Thread Hudson (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-0?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16836643#comment-16836643
 ] 

Hudson commented on HBASE-0:


Results for branch master
[build #995 on 
builds.a.o|https://builds.apache.org/job/HBase%20Nightly/job/master/995/]: (x) 
*{color:red}-1 overall{color}*

details (if available):

(x) {color:red}-1 general checks{color}
-- For more information [see general 
report|https://builds.apache.org/job/HBase%20Nightly/job/master/995//General_Nightly_Build_Report/]




(/) {color:green}+1 jdk8 hadoop2 checks{color}
-- For more information [see jdk8 (hadoop2) 
report|https://builds.apache.org/job/HBase%20Nightly/job/master/995//JDK8_Nightly_Build_Report_(Hadoop2)/]


(x) {color:red}-1 jdk8 hadoop3 checks{color}
-- For more information [see jdk8 (hadoop3) 
report|https://builds.apache.org/job/HBase%20Nightly/job/master/995//JDK8_Nightly_Build_Report_(Hadoop3)/]


(/) {color:green}+1 source release artifact{color}
-- See build output for details.


(/) {color:green}+1 client integration test{color}


> Release hbase-connectors-1.0.0
> --
>
> Key: HBASE-0
> URL: https://issues.apache.org/jira/browse/HBASE-0
> Project: HBase
>  Issue Type: Task
>  Components: hbase-connectors
>Affects Versions: connector-1.0.0
>Reporter: Balazs Meszaros
>Assignee: Balazs Meszaros
>Priority: Major
> Fix For: connector-1.0.0
>
>




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-22358) Change rubocop configuration for method length

2019-05-09 Thread Hudson (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-22358?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16836642#comment-16836642
 ] 

Hudson commented on HBASE-22358:


Results for branch master
[build #995 on 
builds.a.o|https://builds.apache.org/job/HBase%20Nightly/job/master/995/]: (x) 
*{color:red}-1 overall{color}*

details (if available):

(x) {color:red}-1 general checks{color}
-- For more information [see general 
report|https://builds.apache.org/job/HBase%20Nightly/job/master/995//General_Nightly_Build_Report/]




(/) {color:green}+1 jdk8 hadoop2 checks{color}
-- For more information [see jdk8 (hadoop2) 
report|https://builds.apache.org/job/HBase%20Nightly/job/master/995//JDK8_Nightly_Build_Report_(Hadoop2)/]


(x) {color:red}-1 jdk8 hadoop3 checks{color}
-- For more information [see jdk8 (hadoop3) 
report|https://builds.apache.org/job/HBase%20Nightly/job/master/995//JDK8_Nightly_Build_Report_(Hadoop3)/]


(/) {color:green}+1 source release artifact{color}
-- See build output for details.


(/) {color:green}+1 client integration test{color}


> Change rubocop configuration for method length
> --
>
> Key: HBASE-22358
> URL: https://issues.apache.org/jira/browse/HBASE-22358
> Project: HBase
>  Issue Type: Improvement
>  Components: community, shell
>Reporter: Jan Hentschel
>Assignee: Murtaza Hassan
>Priority: Minor
>  Labels: beginner, beginners
> Fix For: 3.0.0, 1.5.0, 2.3.0, 2.0.6, 2.1.5, 2.2.1, 1.3.5, 1.4.11
>
>
> rubocop currently uses a maximum method length for the Ruby code of 10, which 
> is way too restrictive. In Checkstyle we're using 150 lines per method. Don't 
> know if it needs to be that much, but something between 50 and 75 seems to be 
> more realistic, especially for test cases.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (HBASE-22390) backport HBASE-22190 to branch-1

2019-05-09 Thread Zach York (JIRA)
Zach York created HBASE-22390:
-

 Summary: backport HBASE-22190 to branch-1
 Key: HBASE-22390
 URL: https://issues.apache.org/jira/browse/HBASE-22390
 Project: HBase
  Issue Type: Bug
Affects Versions: 1.4.10, 1.5.0
Reporter: Zach York
Assignee: Zach York


HBASE-22190 was only applied to branch-2+ while the bug exists in branch-1 as 
well.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-22330) Backport HBASE-20724 (Sometimes some compacted storefiles are still opened after region failover) to branch-1

2019-05-09 Thread Xu Cang (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-22330?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16836640#comment-16836640
 ] 

Xu Cang commented on HBASE-22330:
-

Thanks! I ran tests after applying the new addendum. 5 out of 5 times passed.

+1

> Backport HBASE-20724 (Sometimes some compacted storefiles are still opened 
> after region failover) to branch-1
> -
>
> Key: HBASE-22330
> URL: https://issues.apache.org/jira/browse/HBASE-22330
> Project: HBase
>  Issue Type: Sub-task
>  Components: Compaction, regionserver
>Affects Versions: 1.5.0, 1.4.9, 1.3.4
>Reporter: Andrew Purtell
>Assignee: Abhishek Singh Chouhan
>Priority: Major
> Fix For: 1.5.0, 1.3.5, 1.4.11
>
> Attachments: HBASE-22330-addendum.branch-1.patch, 
> HBASE-22330.branch-1.001.patch, HBASE-22330.branch-1.002.patch, 
> HBASE-22330.branch-1.3.001.patch
>
>
> There appears to be a race condition between close and split which when 
> combined with a side effect of HBASE-20704, leads to the parent region store 
> files getting archived and cleared while daughter regions still have 
> references to those parent region store files.
> Here is the timeline of events observed for an affected region:
>  # RS1 faces ZooKeeper connectivity issue for master node and starts shutting 
> itself down. As part of this it starts to close the store and clean up the 
> compacted files (File A)
>  # Master starts bulk assigning regions and assign parent region to RS2
>  # Region opens on RS2 and ends up opening compacted store file(s) (suspect 
> this is due to HBASE-20724)
>  # Now split happens and daughter regions open on RS2 and try to run a 
> compaction as part of post open
>  # Split request at this point is complete. However now archiving proceeds on 
> RS1 and ends up archiving the store file that is referenced by the daughter. 
> Compaction fails due to FileNotFoundException and all subsequent attempts to 
> open the region will fail until manual resolution.
> We think having HBASE-20724 would help in such situations since we won't end 
> up loading compacted store files in the first place. 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HBASE-22330) Backport HBASE-20724 (Sometimes some compacted storefiles are still opened after region failover) to branch-1

2019-05-09 Thread Abhishek Singh Chouhan (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-22330?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Abhishek Singh Chouhan updated HBASE-22330:
---
Attachment: HBASE-22330-addendum.branch-1.patch

> Backport HBASE-20724 (Sometimes some compacted storefiles are still opened 
> after region failover) to branch-1
> -
>
> Key: HBASE-22330
> URL: https://issues.apache.org/jira/browse/HBASE-22330
> Project: HBase
>  Issue Type: Sub-task
>  Components: Compaction, regionserver
>Affects Versions: 1.5.0, 1.4.9, 1.3.4
>Reporter: Andrew Purtell
>Assignee: Abhishek Singh Chouhan
>Priority: Major
> Fix For: 1.5.0, 1.3.5, 1.4.11
>
> Attachments: HBASE-22330-addendum.branch-1.patch, 
> HBASE-22330.branch-1.001.patch, HBASE-22330.branch-1.002.patch, 
> HBASE-22330.branch-1.3.001.patch
>
>
> There appears to be a race condition between close and split which when 
> combined with a side effect of HBASE-20704, leads to the parent region store 
> files getting archived and cleared while daughter regions still have 
> references to those parent region store files.
> Here is the timeline of events observed for an affected region:
>  # RS1 faces ZooKeeper connectivity issue for master node and starts shutting 
> itself down. As part of this it starts to close the store and clean up the 
> compacted files (File A)
>  # Master starts bulk assigning regions and assign parent region to RS2
>  # Region opens on RS2 and ends up opening compacted store file(s) (suspect 
> this is due to HBASE-20724)
>  # Now split happens and daughter regions open on RS2 and try to run a 
> compaction as part of post open
>  # Split request at this point is complete. However now archiving proceeds on 
> RS1 and ends up archiving the store file that is referenced by the daughter. 
> Compaction fails due to FileNotFoundException and all subsequent attempts to 
> open the region will fail until manual resolution.
> We think having HBASE-20724 would help in such situations since we won't end 
> up loading compacted store files in the first place. 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HBASE-22330) Backport HBASE-20724 (Sometimes some compacted storefiles are still opened after region failover) to branch-1

2019-05-09 Thread Abhishek Singh Chouhan (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-22330?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Abhishek Singh Chouhan updated HBASE-22330:
---
Hadoop Flags:   (was: Reviewed)
  Status: Patch Available  (was: Reopened)

> Backport HBASE-20724 (Sometimes some compacted storefiles are still opened 
> after region failover) to branch-1
> -
>
> Key: HBASE-22330
> URL: https://issues.apache.org/jira/browse/HBASE-22330
> Project: HBase
>  Issue Type: Sub-task
>  Components: Compaction, regionserver
>Affects Versions: 1.3.4, 1.4.9, 1.5.0
>Reporter: Andrew Purtell
>Assignee: Abhishek Singh Chouhan
>Priority: Major
> Fix For: 1.5.0, 1.3.5, 1.4.11
>
> Attachments: HBASE-22330-addendum.branch-1.patch, 
> HBASE-22330.branch-1.001.patch, HBASE-22330.branch-1.002.patch, 
> HBASE-22330.branch-1.3.001.patch
>
>
> There appears to be a race condition between close and split which when 
> combined with a side effect of HBASE-20704, leads to the parent region store 
> files getting archived and cleared while daughter regions still have 
> references to those parent region store files.
> Here is the timeline of events observed for an affected region:
>  # RS1 faces ZooKeeper connectivity issue for master node and starts shutting 
> itself down. As part of this it starts to close the store and clean up the 
> compacted files (File A)
>  # Master starts bulk assigning regions and assign parent region to RS2
>  # Region opens on RS2 and ends up opening compacted store file(s) (suspect 
> this is due to HBASE-20724)
>  # Now split happens and daughter regions open on RS2 and try to run a 
> compaction as part of post open
>  # Split request at this point is complete. However now archiving proceeds on 
> RS1 and ends up archiving the store file that is referenced by the daughter. 
> Compaction fails due to FileNotFoundException and all subsequent attempts to 
> open the region will fail until manual resolution.
> We think having HBASE-20724 would help in such situations since we won't end 
> up loading compacted store files in the first place. 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Reopened] (HBASE-22330) Backport HBASE-20724 (Sometimes some compacted storefiles are still opened after region failover) to branch-1

2019-05-09 Thread Abhishek Singh Chouhan (JIRA)


 [ 
https://issues.apache.org/jira/browse/HBASE-22330?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Abhishek Singh Chouhan reopened HBASE-22330:


Hardening the test

> Backport HBASE-20724 (Sometimes some compacted storefiles are still opened 
> after region failover) to branch-1
> -
>
> Key: HBASE-22330
> URL: https://issues.apache.org/jira/browse/HBASE-22330
> Project: HBase
>  Issue Type: Sub-task
>  Components: Compaction, regionserver
>Affects Versions: 1.5.0, 1.4.9, 1.3.4
>Reporter: Andrew Purtell
>Assignee: Abhishek Singh Chouhan
>Priority: Major
> Fix For: 1.5.0, 1.3.5, 1.4.11
>
> Attachments: HBASE-22330.branch-1.001.patch, 
> HBASE-22330.branch-1.002.patch, HBASE-22330.branch-1.3.001.patch
>
>
> There appears to be a race condition between close and split which when 
> combined with a side effect of HBASE-20704, leads to the parent region store 
> files getting archived and cleared while daughter regions still have 
> references to those parent region store files.
> Here is the timeline of events observed for an affected region:
>  # RS1 faces ZooKeeper connectivity issue for master node and starts shutting 
> itself down. As part of this it starts to close the store and clean up the 
> compacted files (File A)
>  # Master starts bulk assigning regions and assign parent region to RS2
>  # Region opens on RS2 and ends up opening compacted store file(s) (suspect 
> this is due to HBASE-20724)
>  # Now split happens and daughter regions open on RS2 and try to run a 
> compaction as part of post open
>  # Split request at this point is complete. However now archiving proceeds on 
> RS1 and ends up archiving the store file that is referenced by the daughter. 
> Compaction fails due to FileNotFoundException and all subsequent attempts to 
> open the region will fail until manual resolution.
> We think having HBASE-20724 would help in such situations since we won't end 
> up loading compacted store files in the first place. 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] [hbase] z-york commented on issue #229: HBASE-22389 Revert "HBASE-19275 TestSnapshotFileCache never worked pr…

2019-05-09 Thread GitBox
z-york commented on issue #229: HBASE-22389 Revert "HBASE-19275 
TestSnapshotFileCache never worked pr…
URL: https://github.com/apache/hbase/pull/229#issuecomment-491029923
 
 
   Note this will actually cause the new test added in HBASE-21070 in 
TestSnapshotFileCache to fail until HBASE-22190 is backported (I'm opening a 
different PR for that). I originally missed that it wasn't backported to 
branch-1, then when I realized it wasn't, I was confused why the test I added 
was passing which led me to this.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (HBASE-22330) Backport HBASE-20724 (Sometimes some compacted storefiles are still opened after region failover) to branch-1

2019-05-09 Thread Abhishek Singh Chouhan (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-22330?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16836633#comment-16836633
 ] 

Abhishek Singh Chouhan commented on HBASE-22330:


1 in 10 runs fail for me locally. Looks to be a test only issue due to 
differences in what testUtil.waitTableAvailable(..) does between master and 
branch-1. Let me put up an addendum for the test which removes flakiness. 

> Backport HBASE-20724 (Sometimes some compacted storefiles are still opened 
> after region failover) to branch-1
> -
>
> Key: HBASE-22330
> URL: https://issues.apache.org/jira/browse/HBASE-22330
> Project: HBase
>  Issue Type: Sub-task
>  Components: Compaction, regionserver
>Affects Versions: 1.5.0, 1.4.9, 1.3.4
>Reporter: Andrew Purtell
>Assignee: Abhishek Singh Chouhan
>Priority: Major
> Fix For: 1.5.0, 1.3.5, 1.4.11
>
> Attachments: HBASE-22330.branch-1.001.patch, 
> HBASE-22330.branch-1.002.patch, HBASE-22330.branch-1.3.001.patch
>
>
> There appears to be a race condition between close and split which when 
> combined with a side effect of HBASE-20704, leads to the parent region store 
> files getting archived and cleared while daughter regions still have 
> references to those parent region store files.
> Here is the timeline of events observed for an affected region:
>  # RS1 faces ZooKeeper connectivity issue for master node and starts shutting 
> itself down. As part of this it starts to close the store and clean up the 
> compacted files (File A)
>  # Master starts bulk assigning regions and assign parent region to RS2
>  # Region opens on RS2 and ends up opening compacted store file(s) (suspect 
> this is due to HBASE-20724)
>  # Now split happens and daughter regions open on RS2 and try to run a 
> compaction as part of post open
>  # Split request at this point is complete. However now archiving proceeds on 
> RS1 and ends up archiving the store file that is referenced by the daughter. 
> Compaction fails due to FileNotFoundException and all subsequent attempts to 
> open the region will fail until manual resolution.
> We think having HBASE-20724 would help in such situations since we won't end 
> up loading compacted store files in the first place. 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (HBASE-22389) Revert HBASE-19275 TestSnapshotFileCache never worked properly on branch-1

2019-05-09 Thread Zach York (JIRA)
Zach York created HBASE-22389:
-

 Summary: Revert HBASE-19275 TestSnapshotFileCache never worked 
properly on branch-1
 Key: HBASE-22389
 URL: https://issues.apache.org/jira/browse/HBASE-22389
 Project: HBase
  Issue Type: Bug
Affects Versions: 1.5.0
Reporter: Zach York
Assignee: Zach York


This commit seems broken on branch-1. 
https://github.com/apache/hbase/commit/d80d3fa45487fedf203c7144e8267753e4d53529#diff-ae046012dd310fc116a729bf36c7df81
 seems to have fixed the issue that the other HBASE-19275 patches fixed, but 
when HBASE-19275 was applied to branch-1 it appears to have reverted this code: 
https://github.com/apache/hbase/commit/e8e4beacb039299a9e56cd355e3ada08784a7acb#diff-ae046012dd310fc116a729bf36c7df81
 meaning the tests always pass again.

[~apurtell] [~xucang]



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] [hbase] z-york opened a new pull request #229: HBASE-22389 Revert "HBASE-19275 TestSnapshotFileCache never worked pr…

2019-05-09 Thread GitBox
z-york opened a new pull request #229: HBASE-22389 Revert "HBASE-19275 
TestSnapshotFileCache never worked pr…
URL: https://github.com/apache/hbase/pull/229
 
 
   …operly"
   
   This reverts commit e8e4beacb039299a9e56cd355e3ada08784a7acb.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (HBASE-22358) Change rubocop configuration for method length

2019-05-09 Thread Hudson (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-22358?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16836632#comment-16836632
 ] 

Hudson commented on HBASE-22358:


Results for branch branch-2
[build #1879 on 
builds.a.o|https://builds.apache.org/job/HBase%20Nightly/job/branch-2/1879/]: 
(x) *{color:red}-1 overall{color}*

details (if available):

(x) {color:red}-1 general checks{color}
-- For more information [see general 
report|https://builds.apache.org/job/HBase%20Nightly/job/branch-2/1879//General_Nightly_Build_Report/]




(x) {color:red}-1 jdk8 hadoop2 checks{color}
-- For more information [see jdk8 (hadoop2) 
report|https://builds.apache.org/job/HBase%20Nightly/job/branch-2/1879//JDK8_Nightly_Build_Report_(Hadoop2)/]


(/) {color:green}+1 jdk8 hadoop3 checks{color}
-- For more information [see jdk8 (hadoop3) 
report|https://builds.apache.org/job/HBase%20Nightly/job/branch-2/1879//JDK8_Nightly_Build_Report_(Hadoop3)/]


(/) {color:green}+1 source release artifact{color}
-- See build output for details.


(/) {color:green}+1 client integration test{color}


> Change rubocop configuration for method length
> --
>
> Key: HBASE-22358
> URL: https://issues.apache.org/jira/browse/HBASE-22358
> Project: HBase
>  Issue Type: Improvement
>  Components: community, shell
>Reporter: Jan Hentschel
>Assignee: Murtaza Hassan
>Priority: Minor
>  Labels: beginner, beginners
> Fix For: 3.0.0, 1.5.0, 2.3.0, 2.0.6, 2.1.5, 2.2.1, 1.3.5, 1.4.11
>
>
> rubocop currently uses a maximum method length for the Ruby code of 10, which 
> is way too restrictive. In Checkstyle we're using 150 lines per method. Don't 
> know if it needs to be that much, but something between 50 and 75 seems to be 
> more realistic, especially for test cases.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HBASE-22324) loss a mass of data when the sequenceId of cells greater than Integer.Max, because MemStoreMergerSegmentsIterator can not merge segments

2019-05-09 Thread Hudson (JIRA)


[ 
https://issues.apache.org/jira/browse/HBASE-22324?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16836631#comment-16836631
 ] 

Hudson commented on HBASE-22324:


Results for branch branch-2
[build #1879 on 
builds.a.o|https://builds.apache.org/job/HBase%20Nightly/job/branch-2/1879/]: 
(x) *{color:red}-1 overall{color}*

details (if available):

(x) {color:red}-1 general checks{color}
-- For more information [see general 
report|https://builds.apache.org/job/HBase%20Nightly/job/branch-2/1879//General_Nightly_Build_Report/]




(x) {color:red}-1 jdk8 hadoop2 checks{color}
-- For more information [see jdk8 (hadoop2) 
report|https://builds.apache.org/job/HBase%20Nightly/job/branch-2/1879//JDK8_Nightly_Build_Report_(Hadoop2)/]


(/) {color:green}+1 jdk8 hadoop3 checks{color}
-- For more information [see jdk8 (hadoop3) 
report|https://builds.apache.org/job/HBase%20Nightly/job/branch-2/1879//JDK8_Nightly_Build_Report_(Hadoop3)/]


(/) {color:green}+1 source release artifact{color}
-- See build output for details.


(/) {color:green}+1 client integration test{color}


>  loss a mass of data when the sequenceId of cells greater than Integer.Max, 
> because MemStoreMergerSegmentsIterator can not merge segments 
> --
>
> Key: HBASE-22324
> URL: https://issues.apache.org/jira/browse/HBASE-22324
> Project: HBase
>  Issue Type: Bug
>  Components: in-memory-compaction
>Affects Versions: 2.1.0, 2.2.0
>Reporter: chenyang
>Assignee: ChenYang
>Priority: Blocker
>  Labels: patch
> Fix For: 3.0.0, 2.1.0, 2.2.0, 2.3.0, 2.0.6
>
> Attachments: HBASE-22324.branch-2.1.0005.patch, 
> HBASE-22324.branch-2.1.0006.patch
>
>
> if your memstore type is CompactingMemStore,MemStoreMergerSegmentsIterator 
> can not merge memstore segments when the seqId of cells greater than 
> Integer.Max, as a result, lossing a mass of data. the reason is that 
> MemStoreMergerSegmentsIterator use Integer.Max as readPt when create Scanner, 
>  but the seqId of cell  may be greater than Integer.MAX_VALUE,  it`s type is 
> long.   code as below:
> {code:java}
> public MemStoreMergerSegmentsIterator(List segments, 
> CellComparator comparator,
> int compactionKVMax) throws IOException {
>   super(compactionKVMax);
>   // create the list of scanners to traverse over all the data
>   // no dirty reads here as these are immutable segments
>   AbstractMemStore.addToScanners(segments, Integer.MAX_VALUE, scanners); 
> //bug, should use Long.MAX_VALUE
>   heap = new KeyValueHeap(scanners, comparator);
> }
> SegmentScanner.java code as below
> protected void updateCurrent() {
>   Cell startKV = current;
>   Cell next = null;
>   try {
> while (iter.hasNext()) {
>   next = iter.next();
>   // here, if seqId>readPoint(Integer.MAX_VALUE), never read cell, as a 
> result, lossing lots of cells
>   if (next.getSequenceId() <= this.readPoint) {
> current = next;
> return;// skip irrelevant versions
>   }
>   if (stopSkippingKVsIfNextRow &&   // for backwardSeek() stay in the
>   startKV != null &&// boundaries of a single row
>   segment.compareRows(next, startKV) > 0) {
> current = null;
> return;
>   }
> } // end of while
> current = null; // nothing found
>   } finally {
> if (next != null) {
>   // in all cases, remember the last KV we iterated to, needed for 
> reseek()
>   last = next;
> }
>   }
> }
> MemStoreCompactorSegmentsIterator has the same bug
> public MemStoreCompactorSegmentsIterator(List segments,
> CellComparator comparator, int compactionKVMax, HStore store) throws 
> IOException {
>   super(compactionKVMax);
>   List scanners = new ArrayList();
>   AbstractMemStore.addToScanners(segments, Integer.MAX_VALUE, scanners);   
> //bug, should use Long.MAX_VALUE
>   // build the scanner based on Query Matcher
>   // reinitialize the compacting scanner for each instance of iterator
>   compactingScanner = createScanner(store, scanners);
>   refillKVS();
> }{code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


  1   2   >