It is not you. There are a bunch of flies we need to fix. This latter is for sure flakey. Let me take a look. Thanks, JMS.
S On Jan 6, 2018 5:57 PM, "Jean-Marc Spaggiari" <jean-m...@spaggiari.org> wrote: I might not doing the right magic to get that run.... If someone is able to get all the tests pass, can you please share the command you run? Thanks, JMS [INFO] Results: [INFO] [ERROR] Failures: [ERROR] TestFromClientSide.testCheckAndDeleteWithCompareOp:4982 expected:<false> but was:<true> [ERROR] org.apache.hadoop.hbase.master.assignment.TestMergeTableRegionsProcedure .testMergeRegionsConcurrently(org.apache.hadoop.hbase.master.assig nment.TestMergeTableRegionsProcedure) [ERROR] Run 1: TestMergeTableRegionsProcedure.setup:111->resetProcExecutorTestingKillFl ag:138 expected executor to be running [ERROR] Run 2: TestMergeTableRegionsProcedure.tearDown:128->resetProcExecutorTestingKillFl ag:138 expected executor to be running [INFO] [ERROR] org.apache.hadoop.hbase.master.assignment.TestMergeTableRegionsProcedure .testMergeTwoRegions(org.apache.hadoop.hbase.master.assignment.Tes tMergeTableRegionsProcedure) [ERROR] Run 1: TestMergeTableRegionsProcedure.setup:111->resetProcExecutorTestingKillFl ag:138 expected executor to be running [ERROR] Run 2: TestMergeTableRegionsProcedure.tearDown:128->resetProcExecutorTestingKillFl ag:138 expected executor to be running [INFO] [ERROR] org.apache.hadoop.hbase.master.assignment.TestMergeTableRegionsProcedure. testRecoveryAndDoubleExecution(org.apache.hadoop.hbase.master.ass ignment.TestMergeTableRegionsProcedure) [ERROR] Run 1: TestMergeTableRegionsProcedure.setup:111->resetProcExecutorTestingKillFl ag:138 expected executor to be running [ERROR] Run 2: TestMergeTableRegionsProcedure.tearDown:128->resetProcExecutorTestingKillFl ag:138 expected executor to be running [INFO] [ERROR] org.apache.hadoop.hbase.master.assignment.TestMergeTableRegionsProcedure. testRollbackAndDoubleExecution(org.apache.hadoop.hbase.master.ass ignment.TestMergeTableRegionsProcedure) [ERROR] Run 1: TestMergeTableRegionsProcedure.testRollbackAndDoubleExecution:272 expected:<true> but was:<false> [ERROR] Run 2: TestMergeTableRegionsProcedure.tearDown:128->resetProcExecutorTestingKillFl ag:138 expected executor to be running [INFO] [ERROR] TestSnapshotQuotaObserverChore.testSnapshotSize:276 Waiting timed out after [30 000] msec [ERROR] TestHRegionWithInMemoryFlush>TestHRegion.testWritesWhileScanning:3813 expected null, but was:<org.apache.hadoop.hbase.NotServingRegionException: testWritesWhileScanning,,1515277468063.468265483817cb6da632026ba5b306f6. is closing> [ERROR] Errors: [ERROR] TestDLSAsyncFSWAL>AbstractTestDLS.testThreeRSAbort:401 » TableNotFound testThr... [ERROR] org.apache.hadoop.hbase.master.balancer.TestRegionsOnMasterOptions. testRegionsOnAllServers(org.apache.hadoop.hbase.master.balancer. TestRegionsOnMasterOptions) [ERROR] Run 1: TestRegionsOnMasterOptions.testRegionsOnAllServers:94-> checkBalance:207->Object.wait:-2 » TestTimedOut [ERROR] Run 2: TestRegionsOnMasterOptions.testRegionsOnAllServers » Appears to be stuck in t... [INFO] [INFO] [ERROR] Tests run: 3604, Failures: 7, Errors: 2, Skipped: 44 [INFO] 2018-01-06 15:52 GMT-05:00 Jean-Marc Spaggiari <jean-m...@spaggiari.org>: > Deleted the class to get all the tests running. Was running on the RC1 > from the tar. > > I know get those one failing. > > [ERROR] Failures: > [ERROR] TestFavoredStochasticLoadBalancer.test2FavoredNodesDead:352 > Balancer did not run > [ERROR] TestRegionMergeTransactionOnCluster.testCleanMergeReference:284 > hdfs://localhost:45311/user/jmspaggi/test-data/7c269e83- > 5982-449e-8cf8-6babaaaa4c7c/data/default/testCleanMergeReference/ > f1bdc6441b090dbacb391c74eaf0d1d0 > [ERROR] Errors: > [ERROR] TestDLSAsyncFSWAL>AbstractTestDLS.testThreeRSAbort:401 » > TableNotFound Region ... > [INFO] > [ERROR] Tests run: 3604, Failures: 2, Errors: 1, Skipped: 44 > > > I have not been able to get all the tests passed locally for a while :( > > JM > > 2018-01-06 15:05 GMT-05:00 Ted Yu <yuzhih...@gmail.com>: > >> Looks like you didn't include HBASE-19666 which would be in the next RC. >> >> On Sat, Jan 6, 2018 at 10:52 AM, Jean-Marc Spaggiari < >> jean-m...@spaggiari.org> wrote: >> >> > Trying with a different command line (mvn test -P runAllTests >> > -Dsurefire.secondPartThreadCount=12 -Dtest.build.data.basedirector >> y=/ram4g >> > ) I get all those one failing. How are you able to get everything >> passed??? >> > >> > [INFO] Results: >> > [INFO] >> > [ERROR] Failures: >> > [ERROR] TestDefaultCompactSelection.testCompactionRatio:74->TestCom >> > pactionPolicy.compactEquals:182->TestCompactionPolicy.compactEquals:201 >> > expected:<[[4, 2, 1]]> but was:<[[]]> >> > [ERROR] TestDefaultCompactSelection.testStuckStoreCompaction:145->T >> > estCompactionPolicy.compactEquals:182->TestCompactionPolicy. >> compactEquals:201 >> > expected:<[[]30, 30, 30]> but was:<[[99, 30, ]30, 30, 30]> >> > [INFO] >> > [ERROR] Tests run: 1235, Failures: 2, Errors: 0, Skipped: 4 >> > >> > Second run: >> > [INFO] Results: >> > [INFO] >> > [ERROR] Failures: >> > [ERROR] TestDefaultCompactSelection.testCompactionRatio:74-> >> > TestCompactionPolicy.compactEquals:182->TestCompactionPolicy >> .compactEquals:201 >> > expected:<[[4, 2, 1]]> but was:<[[]]> >> > [ERROR] TestDefaultCompactSelection.testStuckStoreCompaction:145-> >> > TestCompactionPolicy.compactEquals:182->TestCompactionPolicy >> .compactEquals:201 >> > expected:<[[]30, 30, 30]> but was:<[[99, 30, ]30, 30, 30]> >> > [INFO] >> > [ERROR] Tests run: 1235, Failures: 2, Errors: 0, Skipped: 4 >> > >> > Then again: >> > >> > [INFO] Results: >> > [INFO] >> > [ERROR] Failures: >> > [ERROR] TestDefaultCompactSelection.testCompactionRatio:74-> >> > TestCompactionPolicy.compactEquals:182->TestCompactionPolicy >> .compactEquals:201 >> > expected:<[[4, 2, 1]]> but was:<[[]]> >> > [ERROR] TestDefaultCompactSelection.testStuckStoreCompaction:145-> >> > TestCompactionPolicy.compactEquals:182->TestCompactionPolicy >> .compactEquals:201 >> > expected:<[[]30, 30, 30]> but was:<[[99, 30, ]30, 30, 30]> >> > [INFO] >> > [ERROR] Tests run: 1235, Failures: 2, Errors: 0, Skipped: 4 >> > [INFO] >> > [INFO] ------------------------------------------------------------ >> > ------------ >> > [INFO] Reactor Summary: >> > >> > >> > Sound like it's always the exact same result. Do I have a way to exclude >> > this TestCompactionPolicy test from the run? >> > >> > Here are more details from the last failure: >> > ------------------------------------------------------------ >> > ------------------- >> > Test set: org.apache.hadoop.hbase.regionserver.TestDefaultCompactSelec >> tion >> > ------------------------------------------------------------ >> > ------------------- >> > Tests run: 4, Failures: 2, Errors: 0, Skipped: 0, Time elapsed: 1.323 s >> > <<< FAILURE! - in org.apache.hadoop.hbase.regionserver. >> > TestDefaultCompactSelection >> > testStuckStoreCompaction(org.apache.hadoop.hbase.regionserve >> r.TestDefaultCompactSelection) >> > Time elapsed: 1.047 s <<< FAILURE! >> > org.junit.ComparisonFailure: expected:<[[]30, 30, 30]> but was:<[[99, >> 30, >> > ]30, 30, 30]> >> > at org.apache.hadoop.hbase.regionserver. >> > TestDefaultCompactSelection.testStuckStoreCompaction( >> > TestDefaultCompactSelection.java:145) >> > >> > testCompactionRatio(org.apache.hadoop.hbase.regionserver.Tes >> tDefaultCompactSelection) >> > Time elapsed: 0.096 s <<< FAILURE! >> > org.junit.ComparisonFailure: expected:<[[4, 2, 1]]> but was:<[[]]> >> > at org.apache.hadoop.hbase.regionserver. >> > TestDefaultCompactSelection.testCompactionRatio( >> > TestDefaultCompactSelection.java:74) >> > >> > >> > 2018-01-06 12:53:53,240 WARN [StoreOpener-22ce1d683ba4b6b93 >> 73a3c541ebab2a2-1] >> > util.CommonFSUtils(536): FileSystem doesn't support setStoragePolicy; >> > HDFS-6584, HDFS-9345 not available. This is normal and expected on >> earlier >> > Hadoop versions. >> > java.lang.NoSuchMethodException: org.apache.hadoop.fs.LocalFileSystem. >> > setStoragePolicy(org.apache.hadoop.fs.Path, java.lang.String) >> > at java.lang.Class.getDeclaredMethod(Class.java:2130) >> > at org.apache.hadoop.hbase.util.CommonFSUtils. >> > invokeSetStoragePolicy(CommonFSUtils.java:528) >> > at org.apache.hadoop.hbase.util.CommonFSUtils.setStoragePolicy( >> > CommonFSUtils.java:518) >> > at org.apache.hadoop.hbase.regionserver.HRegionFileSystem. >> > setStoragePolicy(HRegionFileSystem.java:193) >> > at org.apache.hadoop.hbase.regionserver.HStore.<init>( >> > HStore.java:250) >> > at org.apache.hadoop.hbase.regionserver.HRegion. >> > instantiateHStore(HRegion.java:5497) >> > at org.apache.hadoop.hbase.regionserver.HRegion$1.call( >> > HRegion.java:1002) >> > at org.apache.hadoop.hbase.regionserver.HRegion$1.call( >> > HRegion.java:999) >> > at java.util.concurrent.FutureTask.run(FutureTask.java:266) >> > at java.util.concurrent.Executors$RunnableAdapter. >> > call(Executors.java:511) >> > at java.util.concurrent.FutureTask.run(FutureTask.java:266) >> > at java.util.concurrent.ThreadPoolExecutor.runWorker( >> > ThreadPoolExecutor.java:1149) >> > at java.util.concurrent.ThreadPoolExecutor$Worker.run( >> > ThreadPoolExecutor.java:624) >> > at java.lang.Thread.run(Thread.java:748) >> > >> > >> > >> > 2018-01-06 12:53:53,322 DEBUG [main] util.CommonFSUtils(538): FileSystem >> > doesn't support setStoragePolicy; HDFS-6584, HDFS-9345 not available. >> This >> > is normal and expected on earlier Hadoop versions. >> > java.lang.NoSuchMethodException: org.apache.hadoop.fs.LocalFileSystem. >> > setStoragePolicy(org.apache.hadoop.fs.Path, java.lang.String) >> > at java.lang.Class.getDeclaredMethod(Class.java:2130) >> > at org.apache.hadoop.hbase.util.CommonFSUtils. >> > invokeSetStoragePolicy(CommonFSUtils.java:528) >> > at org.apache.hadoop.hbase.util.CommonFSUtils.setStoragePolicy( >> > CommonFSUtils.java:518) >> > at org.apache.hadoop.hbase.regionserver.HRegionFileSystem. >> > setStoragePolicy(HRegionFileSystem.java:193) >> > at org.apache.hadoop.hbase.regionserver.HStore.<init>( >> > HStore.java:250) >> > at org.apache.hadoop.hbase.regionserver.TestCompactionPolicy. >> > initialize(TestCompactionPolicy.java:109) >> > at org.apache.hadoop.hbase.regionserver. >> > TestCompactionPolicy.setUp(TestCompactionPolicy.java:69) >> > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >> > at sun.reflect.NativeMethodAccessorImpl.invoke( >> > NativeMethodAccessorImpl.java:62) >> > at sun.reflect.DelegatingMethodAccessorImpl.invoke( >> > DelegatingMethodAccessorImpl.java:43) >> > at java.lang.reflect.Method.invoke(Method.java:498) >> > at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall( >> > FrameworkMethod.java:50) >> > at org.junit.internal.runners.model.ReflectiveCallable.run( >> > ReflectiveCallable.java:12) >> > at org.junit.runners.model.FrameworkMethod.invokeExplosively( >> > FrameworkMethod.java:47) >> > at org.junit.internal.runners.statements.RunBefores. >> > evaluate(RunBefores.java:24) >> > at org.junit.internal.runners.statements.RunAfters.evaluate( >> > RunAfters.java:27) >> > at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:325 >> ) >> > at org.junit.runners.BlockJUnit4ClassRunner.runChild( >> > BlockJUnit4ClassRunner.java:78) >> > at org.junit.runners.BlockJUnit4ClassRunner.runChild( >> > BlockJUnit4ClassRunner.java:57) >> > at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290) >> > at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java: >> 71) >> > at org.junit.runners.ParentRunner.runChildren( >> > ParentRunner.java:288) >> > at org.junit.runners.ParentRunner.access$000(ParentRunner.java: >> 58) >> > at org.junit.runners.ParentRunner$2.evaluate( >> > ParentRunner.java:268) >> > at org.junit.runners.ParentRunner.run(ParentRunner.java:363) >> > at org.junit.runners.Suite.runChild(Suite.java:128) >> > at org.junit.runners.Suite.runChild(Suite.java:27) >> > at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290) >> > at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java: >> 71) >> > at org.junit.runners.ParentRunner.runChildren( >> > ParentRunner.java:288) >> > at org.junit.runners.ParentRunner.access$000(ParentRunner.java: >> 58) >> > at org.junit.runners.ParentRunner$2.evaluate( >> > ParentRunner.java:268) >> > at org.junit.runners.ParentRunner.run(ParentRunner.java:363) >> > at org.apache.maven.surefire.junitcore.JUnitCore.run( >> > JUnitCore.java:55) >> > at org.apache.maven.surefire.junitcore.JUnitCoreWrapper. >> > createRequestAndRun(JUnitCoreWrapper.java:137) >> > at org.apache.maven.surefire.junitcore.JUnitCoreWrapper. >> > executeEager(JUnitCoreWrapper.java:107) >> > at org.apache.maven.surefire.junitcore.JUnitCoreWrapper. >> > execute(JUnitCoreWrapper.java:83) >> > at org.apache.maven.surefire.junitcore.JUnitCoreWrapper. >> > execute(JUnitCoreWrapper.java:75) >> > at org.apache.maven.surefire.junitcore.JUnitCoreProvider. >> > invoke(JUnitCoreProvider.java:159) >> > at org.apache.maven.surefire.booter.ForkedBooter. >> > invokeProviderInSameClassLoader(ForkedBooter.java:373) >> > at org.apache.maven.surefire.booter.ForkedBooter. >> > runSuitesInProcess(ForkedBooter.java:334) >> > at org.apache.maven.surefire.booter.ForkedBooter.execute( >> > ForkedBooter.java:119) >> > at org.apache.maven.surefire.booter.ForkedBooter.main( >> > ForkedBooter.java:407) >> > >> > >> > 2018-01-06 12:53:53,398 INFO [main] hbase.ResourceChecker(172): after: >> > regionserver.TestDefaultCompactSelection#testStuckStoreCompaction >> > Thread=11 (was 7) >> > Potentially hanging thread: Monitor thread for TaskMonitor >> > java.lang.Thread.sleep(Native Method) >> > org.apache.hadoop.hbase.monitoring.TaskMonitor$ >> > MonitorRunnable.run(TaskMonitor.java:302) >> > java.lang.Thread.run(Thread.java:748) >> > >> > Potentially hanging thread: org.apache.hadoop.fs.FileSystem$Statistics$ >> > StatisticsDataReferenceCleaner >> > java.lang.Object.wait(Native Method) >> > java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:143) >> > java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:164) >> > org.apache.hadoop.fs.FileSystem$Statistics$ >> > StatisticsDataReferenceCleaner.run(FileSystem.java:3063) >> > java.lang.Thread.run(Thread.java:748) >> > >> > Potentially hanging thread: LruBlockCacheStatsExecutor >> > sun.misc.Unsafe.park(Native Method) >> > java.util.concurrent.locks.LockSupport.parkNanos( >> > LockSupport.java:215) >> > java.util.concurrent.locks.AbstractQueuedSynchronizer$ >> > ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078) >> > java.util.concurrent.ScheduledThreadPoolExecutor$ >> > DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1093) >> > java.util.concurrent.ScheduledThreadPoolExecutor$ >> > DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:809) >> > java.util.concurrent.ThreadPoolExecutor.getTask( >> > ThreadPoolExecutor.java:1074) >> > java.util.concurrent.ThreadPoolExecutor.runWorker( >> > ThreadPoolExecutor.java:1134) >> > java.util.concurrent.ThreadPoolExecutor$Worker.run( >> > ThreadPoolExecutor.java:624) >> > java.lang.Thread.run(Thread.java:748) >> > >> > Potentially hanging thread: StoreOpener-22ce1d683ba4b6b9373a3c541ebab2 >> > a2-1.LruBlockCache.EvictionThread >> > java.lang.Object.wait(Native Method) >> > org.apache.hadoop.hbase.io.hfile.LruBlockCache$EvictionThre >> ad.run( >> > LruBlockCache.java:894) >> > java.lang.Thread.run(Thread.java:748) >> > - Thread LEAK? -, OpenFileDescriptor=232 (was 232), >> > MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=204 (was >> 204), >> > ProcessCount=273 (was 273), AvailableMemoryMB=4049 (was 4132) >> > >> > >> > Full log attached >> > >> > Thanks, >> > >> > JMS >> > >> > >> > 2018-01-06 9:34 GMT-05:00 Mike Drob <md...@apache.org>: >> > >> >> I can reproduce the issue locally. I think it has to do with the java >> >> version being used - IIRC this is related to the version of java used, >> but >> >> we can discuss in more detail on the JIRA. >> >> >> >> https://issues.apache.org/jira/browse/HBASE-19721 >> >> >> >> Thanks, JMS! >> >> >> >> On Sat, Jan 6, 2018 at 6:42 AM, Jean-Marc Spaggiari < >> >> jean-m...@spaggiari.org >> >> > wrote: >> >> >> >> > How you guys are able to get the tests running? >> >> > >> >> > For me it keeps failing on TestReversedScannerCallable. >> >> > >> >> > I tried many times, always fails in the same place. I'm running on a >> 4GB >> >> > tmpfs. Details are below. Am I doing something wrong? >> >> > >> >> > JM >> >> > >> >> > >> >> > >> >> > ./dev-support/hbasetests.sh runAllTests >> >> > >> >> > >> >> > >> >> > [INFO] Running org.apache.hadoop.hbase.client.TestOperation >> >> > [INFO] >> >> > [INFO] Results: >> >> > [INFO] >> >> > [ERROR] Errors: >> >> > [ERROR] TestReversedScannerCallable.unnecessary Mockito stubbings >> » >> >> > UnnecessaryStubbing >> >> > [INFO] >> >> > [ERROR] Tests run: 245, Failures: 0, Errors: 1, Skipped: 8 >> >> > [INFO] >> >> > [INFO] >> >> > ------------------------------------------------------------ >> >> ------------ >> >> > [INFO] Reactor Summary: >> >> > [INFO] >> >> > [INFO] Apache HBase ....................................... SUCCESS >> [ >> >> > 1.409 s] >> >> > [INFO] Apache HBase - Checkstyle .......................... SUCCESS [ >> >> > 1.295 s] >> >> > [INFO] Apache HBase - Build Support ....................... SUCCESS [ >> >> > 0.038 s] >> >> > [INFO] Apache HBase - Error Prone Rules ................... SUCCESS [ >> >> > 1.069 s] >> >> > [INFO] Apache HBase - Annotations ......................... SUCCESS [ >> >> > 1.450 s] >> >> > [INFO] Apache HBase - Build Configuration ................. SUCCESS [ >> >> > 0.073 s] >> >> > [INFO] Apache HBase - Shaded Protocol ..................... SUCCESS [ >> >> > 14.292 s] >> >> > [INFO] Apache HBase - Common .............................. SUCCESS >> >> [01:51 >> >> > min] >> >> > [INFO] Apache HBase - Metrics API ......................... SUCCESS [ >> >> > 2.878 s] >> >> > [INFO] Apache HBase - Hadoop Compatibility ................ SUCCESS [ >> >> > 12.216 s] >> >> > [INFO] Apache HBase - Metrics Implementation .............. SUCCESS [ >> >> > 7.206 s] >> >> > [INFO] Apache HBase - Hadoop Two Compatibility ............ SUCCESS [ >> >> > 12.440 s] >> >> > [INFO] Apache HBase - Protocol ............................ SUCCESS [ >> >> > 0.074 s] >> >> > [INFO] Apache HBase - Client .............................. FAILURE >> >> [02:10 >> >> > min] >> >> > [INFO] Apache HBase - Zookeeper ........................... SKIPPED >> >> > [INFO] Apache HBase - Replication ......................... SKIPPED >> >> > >> >> > >> >> > >> >> > >> >> > >> >> > ------------------------------------------------------------ >> >> > ------------------- >> >> > Test set: org.apache.hadoop.hbase.client.TestReversedScannerCallable >> >> > ------------------------------------------------------------ >> >> > ------------------- >> >> > Tests run: 3, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: >> 1.515 s >> >> <<< >> >> > FAILURE! - in org.apache.hadoop.hbase.client >> >> .TestReversedScannerCallable >> >> > unnecessary Mockito >> >> > stubbings(org.apache.hadoop.hbase.client.TestReversedScanner >> Callable) >> >> > Time >> >> > elapsed: 0.014 s <<< ERROR! >> >> > org.mockito.exceptions.misusing.UnnecessaryStubbingException: >> >> > >> >> > Unnecessary stubbings detected in test class: >> >> TestReversedScannerCallable >> >> > Clean & maintainable test code requires zero unnecessary code. >> >> > Following stubbings are unnecessary (click to navigate to relevant >> line >> >> of >> >> > code): >> >> > 1. -> at >> >> > org.apache.hadoop.hbase.client.TestReversedScannerCallable.setUp( >> >> > TestReversedScannerCallable.java:66) >> >> > 2. -> at >> >> > org.apache.hadoop.hbase.client.TestReversedScannerCallable.setUp( >> >> > TestReversedScannerCallable.java:68) >> >> > Please remove unnecessary stubbings. More info: javadoc for >> >> > UnnecessaryStubbingException class. >> >> > >> >> > >> >> > 2018-01-06 0:44 GMT-05:00 stack <saint....@gmail.com>: >> >> > >> >> > > On Jan 5, 2018 4:44 PM, "Apekshit Sharma" <a...@cloudera.com> >> wrote: >> >> > > >> >> > > bq. Care needs to be exercised backporting. Bug fixes only please. >> If >> >> in >> >> > > doubt, ping me, the RM, please. Thanks. >> >> > > In that case, shouldn't we branch out branch-2.0? We can then do >> >> normal >> >> > > backports to branch-2 and only bug fixes to branch-2.0. >> >> > > >> >> > > >> >> > > >> >> > > Don't you think we have enough branches already mighty Appy? >> >> > > >> >> > > No new features on branch-2? New features are in master/3.0.0 only? >> >> > > >> >> > > S >> >> > > >> >> > > >> >> > > >> >> > > >> >> > > >> >> > > >> >> > > On Fri, Jan 5, 2018 at 9:48 AM, Andrew Purtell < >> apurt...@apache.org> >> >> > > wrote: >> >> > > >> >> > > > TestMemstoreLABWithoutPool is a flake, not a consistent fail. >> >> > > > >> >> > > > >> >> > > > On Fri, Jan 5, 2018 at 7:18 AM, Stack <st...@duboce.net> wrote: >> >> > > > >> >> > > > > On Thu, Jan 4, 2018 at 2:24 PM, Andrew Purtell < >> >> apurt...@apache.org> >> >> > > > > wrote: >> >> > > > > >> >> > > > > > This one is probably my fault: >> >> > > > > > >> >> > > > > > TestDefaultCompactSelection >> >> > > > > > >> >> > > > > > HBASE-19406 >> >> > > > > > >> >> > > > > > >> >> > > > > Balazs fixed it above, HBASE-19666 >> >> > > > > >> >> > > > > >> >> > > > > >> >> > > > > > It can easily be reverted. The failure of interest >> >> > > > > > is TestMemstoreLABWithoutPool.tes >> tLABChunkQueueWithMultipleMSLA >> >> Bs. >> >> > > > > > >> >> > > > > > >> >> > > > > This seems fine. Passes in nightly >> >> > > > > https://builds.apache.org/view/H-L/view/HBase/job/HBase% >> >> > > > > 20Nightly/job/branch-2/171/testReport/org.apache.hadoop. >> >> > > > > hbase.regionserver/TestMemstoreLABWithoutPool/ >> >> > > > > and locally against the tag. It fails consistently for you >> Andrew? >> >> > > > > >> >> > > > > >> >> > > > > > > Should all unit tests pass on a beta? I think so, at least >> if >> >> the >> >> > > > > > failures >> >> > > > > > > are 100% repeatable. >> >> > > > > > > >> >> > > > > > >> >> > > > > >> >> > > > > This is fair. Let me squash this RC and roll another. >> >> > > > > >> >> > > > > Will put it up in a few hours. >> >> > > > > >> >> > > > > Thanks, >> >> > > > > S >> >> > > > > >> >> > > > > >> >> > > > > >> >> > > > > > > -0 >> >> > > > > > > >> >> > > > > > > Checked sums and signatures: ok >> >> > > > > > > RAT check: ok >> >> > > > > > > Built from source: ok (8u144) >> >> > > > > > > Ran unit tests: some failures (8u144) >> >> > > > > > > >> >> > > > > > > [ERROR] TestDefaultCompactSelection.t >> >> estCompactionRatio:74-> >> >> > > > > > > TestCompactionPolicy.compactEq >> uals:182->TestCompactionPolicy. >> >> > > > > > compactEquals:201 >> >> > > > > > > expected:<[[4, 2, 1]]> but was:<[[]]> >> >> > > > > > > >> >> > > > > > > [ERROR] TestDefaultCompactSelection. >> >> > > testStuckStoreCompaction:145-> >> >> > > > > > > TestCompactionPolicy.compactEq >> uals:182->TestCompactionPolicy. >> >> > > > > > compactEquals:201 >> >> > > > > > > expected:<[[]30, 30, 30]> but was:<[[99, 30, ]30, 30, 30]> >> >> > > > > > > >> >> > > > > > > [ERROR] TestMemstoreLABWithoutPool. >> >> > > testLABChunkQueueWithMultipleM >> >> > > > > > SLABs:143 >> >> > > > > > > All the chunks must have been cleared >> >> > > > > > > >> >> > > > > > > >> >> > > > > > > >> >> > > > > > > On Fri, Dec 29, 2017 at 10:15 AM, Stack <st...@duboce.net> >> >> > wrote: >> >> > > > > > > >> >> > > > > > >> The first release candidate for HBase 2.0.0-beta-1 is up >> at: >> >> > > > > > >> >> >> > > > > > >> https://dist.apache.org/repos/ >> dist/dev/hbase/hbase-2.0.0- >> >> > > > beta-1-RC0/ >> >> > > > > > >> >> >> > > > > > >> Maven artifacts are available from a staging directory >> here: >> >> > > > > > >> >> >> > > > > > >> https://repository.apache.org/content/repositories/ >> >> > > > > orgapachehbase-1188 >> >> > > > > > >> >> >> > > > > > >> All was signed with my key at 8ACC93D2 [1] >> >> > > > > > >> >> >> > > > > > >> I tagged the RC as 2.0.0-beta-1-RC0 >> >> > > > > > >> (0907563eb72697b394b8b960fe54887d6ff304fd) >> >> > > > > > >> >> >> > > > > > >> hbase-2.0.0-beta-1 is our first beta release. It includes >> all >> >> > that >> >> > > > was >> >> > > > > > in >> >> > > > > > >> previous alphas (new assignment manager, offheap >> read/write >> >> > path, >> >> > > > > > >> in-memory >> >> > > > > > >> compactions, etc.). The APIs and feature-set are sealed. >> >> > > > > > >> >> >> > > > > > >> hbase-2.0.0-beta-1 is a not-for-production preview of >> >> > hbase-2.0.0. >> >> > > > It >> >> > > > > is >> >> > > > > > >> meant for devs and downstreamers to test drive and flag us >> >> if we >> >> > > > > messed >> >> > > > > > up >> >> > > > > > >> on anything ahead of our rolling GAs. We are particular >> >> > interested >> >> > > > in >> >> > > > > > >> hearing from Coprocessor developers. >> >> > > > > > >> >> >> > > > > > >> The list of features addressed in 2.0.0 so far can be >> found >> >> here >> >> > > > [3]. >> >> > > > > > >> There >> >> > > > > > >> are thousands. The list of ~2k+ fixes in 2.0.0 exclusively >> >> can >> >> > be >> >> > > > > found >> >> > > > > > >> here [4] (My JIRA JQL foo is a bit dodgy -- forgive me if >> >> > > mistakes). >> >> > > > > > >> >> >> > > > > > >> I've updated our overview doc. on the state of 2.0.0 [6]. >> >> We'll >> >> > do >> >> > > > one >> >> > > > > > >> more >> >> > > > > > >> beta before we put up our first 2.0.0 Release Candidate by >> >> the >> >> > end >> >> > > > of >> >> > > > > > >> January, 2.0.0-beta-2. Its focus will be making it so >> users >> >> can >> >> > do >> >> > > a >> >> > > > > > >> rolling upgrade on to hbase-2.x from hbase-1.x (and any >> bug >> >> > fixes >> >> > > > > found >> >> > > > > > >> running beta-1). Here is the list of what we have >> targeted so >> >> > far >> >> > > > for >> >> > > > > > >> beta-2 [5]. Check it out. >> >> > > > > > >> >> >> > > > > > >> One knownissue is that the User API has not been properly >> >> > filtered >> >> > > > so >> >> > > > > it >> >> > > > > > >> shows more than just InterfaceAudience Public content >> >> > > (HBASE-19663, >> >> > > > to >> >> > > > > > be >> >> > > > > > >> fixed by beta-2). >> >> > > > > > >> >> >> > > > > > >> Please take this beta for a spin. Please vote on whether >> it >> >> ok >> >> > to >> >> > > > put >> >> > > > > > out >> >> > > > > > >> this RC as our first beta (Note CHANGES has not yet been >> >> > updated). >> >> > > > Let >> >> > > > > > the >> >> > > > > > >> VOTE be open for 72 hours (Monday) >> >> > > > > > >> >> >> > > > > > >> Thanks, >> >> > > > > > >> Your 2.0.0 Release Manager >> >> > > > > > >> >> >> > > > > > >> 1. http://pgp.mit.edu/pks/lookup?op=get&search= >> >> > 0x9816C7FC8ACC93D2 >> >> > > > > > >> 3. https://goo.gl/scYjJr >> >> > > > > > >> 4. https://goo.gl/dFFT8b >> >> > > > > > >> 5. https://issues.apache.org/jira >> /projects/HBASE/versions/ >> >> > > 12340862 >> >> > > > > > >> 6. https://docs.google.com/document/d/ >> >> > > 1WCsVlnHjJeKUcl7wHwqb4z9iEu_ >> >> > > > > > >> ktczrlKHK8N4SZzs/ >> >> > > > > > >> <https://docs.google.com/document/d/ >> >> > 1WCsVlnHjJeKUcl7wHwqb4z9iEu_ >> >> > > > > > ktczrlKHK8N4SZzs/> >> >> > > > > > >> >> >> > > > > > > >> >> > > > > > > >> >> > > > > > > >> >> > > > > > > -- >> >> > > > > > > Best regards, >> >> > > > > > > Andrew >> >> > > > > > > >> >> > > > > > > Words like orphans lost among the crosstalk, meaning torn >> from >> >> > > > truth's >> >> > > > > > > decrepit hands >> >> > > > > > > - A23, Crosstalk >> >> > > > > > > >> >> > > > > > >> >> > > > > > >> >> > > > > > >> >> > > > > > -- >> >> > > > > > Best regards, >> >> > > > > > Andrew >> >> > > > > > >> >> > > > > > Words like orphans lost among the crosstalk, meaning torn >> from >> >> > > truth's >> >> > > > > > decrepit hands >> >> > > > > > - A23, Crosstalk >> >> > > > > > >> >> > > > > >> >> > > > >> >> > > > >> >> > > > >> >> > > > -- >> >> > > > Best regards, >> >> > > > Andrew >> >> > > > >> >> > > > Words like orphans lost among the crosstalk, meaning torn from >> >> truth's >> >> > > > decrepit hands >> >> > > > - A23, Crosstalk >> >> > > > >> >> > > >> >> > > >> >> > > >> >> > > -- >> >> > > >> >> > > -- Appy >> >> > > >> >> > >> >> >> > >> > >> > >