[jira] [Commented] (HDFS-13560) Insufficient system resources exist to complete the requested service for some tests on Windows
[ https://issues.apache.org/jira/browse/HDFS-13560?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16479950#comment-16479950 ] Hudson commented on HDFS-13560: --- SUCCESS: Integrated in Jenkins build Hadoop-trunk-Commit #14230 (See [https://builds.apache.org/job/Hadoop-trunk-Commit/14230/]) HDFS-13560. Insufficient system resources exist to complete the (inigoiri: rev 53b807a6a8486cefe0b036f7893de9f619bd44a1) * (edit) hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestNameNodeMXBean.java * (edit) hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/Shell.java * (edit) hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestDirectoryScanner.java * (edit) hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/LazyPersistTestCase.java > Insufficient system resources exist to complete the requested service for > some tests on Windows > --- > > Key: HDFS-13560 > URL: https://issues.apache.org/jira/browse/HDFS-13560 > Project: Hadoop HDFS > Issue Type: Bug >Reporter: Anbang Hu >Assignee: Anbang Hu >Priority: Major > Labels: Windows > Fix For: 2.10.0, 3.2.0, 3.1.1, 2.9.2, 3.0.3 > > Attachments: HDFS-13560.000.patch, HDFS-13560.001.patch, > HDFS-13560.002.patch, HDFS-13560.003.patch, HDFS-13560.004.patch > > > On Windows, there are 30 tests in HDFS component giving error like the > following: > {color:#d04437}[ERROR] Tests run: 7, Failures: 0, Errors: 7, Skipped: 0, > Time elapsed: 50.149 s <<< FAILURE! - in > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestLazyPersistFiles{color} > {color:#d04437} [ERROR] > testDisableLazyPersistFileScrubber(org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestLazyPersistFiles) > Time elapsed: 16.513 s <<< ERROR!{color} > {color:#d04437} 1450: Insufficient system resources exist to complete the > requested service.{color} > {color:#d04437}at > org.apache.hadoop.io.nativeio.NativeIO$Windows.extendWorkingSetSize(Native > Method){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:1339){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.(DataNode.java:495){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:2695){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:2598){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.java:1554){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:904){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster.(MiniDFSCluster.java:514){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster$Builder.build(MiniDFSCluster.java:473){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.LazyPersistTestCase.startUpCluster(LazyPersistTestCase.java:316){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.LazyPersistTestCase$ClusterWithRamDiskBuilder.build(LazyPersistTestCase.java:415){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestLazyPersistFiles.testDisableLazyPersistFileScrubber(TestLazyPersistFiles.java:128){color} > {color:#d04437} at sun.reflect.NativeMethodAccessorImpl.invoke0(Native > Method){color} > {color:#d04437} at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62){color} > {color:#d04437} at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43){color} > {color:#d04437} at java.lang.reflect.Method.invoke(Method.java:498){color} > {color:#d04437} at > org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47){color} > {color:#d04437} at > org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12){color} > {color:#d04437} at > org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44){color} > {color:#d04437} at > org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17){color} > {color:#d04437} at > org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27){color} > {color:#d04437} at > org.junit.internal.runners.statements.FailOnTimeout$StatementThread.run(FailOnTimeout.java:74){color} > {color:#33}The involved tests are{color} > {code:java} >
[jira] [Commented] (HDFS-13560) Insufficient system resources exist to complete the requested service for some tests on Windows
[ https://issues.apache.org/jira/browse/HDFS-13560?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16479915#comment-16479915 ] Íñigo Goiri commented on HDFS-13560: Thank you [~huanbang1993] for double checking. +1 on [^HDFS-13560.004.patch]. Committing. > Insufficient system resources exist to complete the requested service for > some tests on Windows > --- > > Key: HDFS-13560 > URL: https://issues.apache.org/jira/browse/HDFS-13560 > Project: Hadoop HDFS > Issue Type: Bug >Reporter: Anbang Hu >Assignee: Anbang Hu >Priority: Major > Labels: Windows > Attachments: HDFS-13560.000.patch, HDFS-13560.001.patch, > HDFS-13560.002.patch, HDFS-13560.003.patch, HDFS-13560.004.patch > > > On Windows, there are 30 tests in HDFS component giving error like the > following: > {color:#d04437}[ERROR] Tests run: 7, Failures: 0, Errors: 7, Skipped: 0, > Time elapsed: 50.149 s <<< FAILURE! - in > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestLazyPersistFiles{color} > {color:#d04437} [ERROR] > testDisableLazyPersistFileScrubber(org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestLazyPersistFiles) > Time elapsed: 16.513 s <<< ERROR!{color} > {color:#d04437} 1450: Insufficient system resources exist to complete the > requested service.{color} > {color:#d04437}at > org.apache.hadoop.io.nativeio.NativeIO$Windows.extendWorkingSetSize(Native > Method){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:1339){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.(DataNode.java:495){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:2695){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:2598){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.java:1554){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:904){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster.(MiniDFSCluster.java:514){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster$Builder.build(MiniDFSCluster.java:473){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.LazyPersistTestCase.startUpCluster(LazyPersistTestCase.java:316){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.LazyPersistTestCase$ClusterWithRamDiskBuilder.build(LazyPersistTestCase.java:415){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestLazyPersistFiles.testDisableLazyPersistFileScrubber(TestLazyPersistFiles.java:128){color} > {color:#d04437} at sun.reflect.NativeMethodAccessorImpl.invoke0(Native > Method){color} > {color:#d04437} at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62){color} > {color:#d04437} at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43){color} > {color:#d04437} at java.lang.reflect.Method.invoke(Method.java:498){color} > {color:#d04437} at > org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47){color} > {color:#d04437} at > org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12){color} > {color:#d04437} at > org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44){color} > {color:#d04437} at > org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17){color} > {color:#d04437} at > org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27){color} > {color:#d04437} at > org.junit.internal.runners.statements.FailOnTimeout$StatementThread.run(FailOnTimeout.java:74){color} > {color:#33}The involved tests are{color} > {code:java} >
[jira] [Commented] (HDFS-13560) Insufficient system resources exist to complete the requested service for some tests on Windows
[ https://issues.apache.org/jira/browse/HDFS-13560?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16479909#comment-16479909 ] Anbang Hu commented on HDFS-13560: -- [~elgoiri] I ran the 30 tests again on local Windows machine and got the same result in my [first post|https://issues.apache.org/jira/browse/HDFS-13560?focusedCommentId=16475420=com.atlassian.jira.plugin.system.issuetabpanels%3Acomment-tabpanel#comment-16475420]. The root cause to "All datanodes are bad" is being dealt with in HDFS-13586. > Insufficient system resources exist to complete the requested service for > some tests on Windows > --- > > Key: HDFS-13560 > URL: https://issues.apache.org/jira/browse/HDFS-13560 > Project: Hadoop HDFS > Issue Type: Bug >Reporter: Anbang Hu >Assignee: Anbang Hu >Priority: Major > Labels: Windows > Attachments: HDFS-13560.000.patch, HDFS-13560.001.patch, > HDFS-13560.002.patch, HDFS-13560.003.patch, HDFS-13560.004.patch > > > On Windows, there are 30 tests in HDFS component giving error like the > following: > {color:#d04437}[ERROR] Tests run: 7, Failures: 0, Errors: 7, Skipped: 0, > Time elapsed: 50.149 s <<< FAILURE! - in > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestLazyPersistFiles{color} > {color:#d04437} [ERROR] > testDisableLazyPersistFileScrubber(org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestLazyPersistFiles) > Time elapsed: 16.513 s <<< ERROR!{color} > {color:#d04437} 1450: Insufficient system resources exist to complete the > requested service.{color} > {color:#d04437}at > org.apache.hadoop.io.nativeio.NativeIO$Windows.extendWorkingSetSize(Native > Method){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:1339){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.(DataNode.java:495){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:2695){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:2598){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.java:1554){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:904){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster.(MiniDFSCluster.java:514){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster$Builder.build(MiniDFSCluster.java:473){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.LazyPersistTestCase.startUpCluster(LazyPersistTestCase.java:316){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.LazyPersistTestCase$ClusterWithRamDiskBuilder.build(LazyPersistTestCase.java:415){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestLazyPersistFiles.testDisableLazyPersistFileScrubber(TestLazyPersistFiles.java:128){color} > {color:#d04437} at sun.reflect.NativeMethodAccessorImpl.invoke0(Native > Method){color} > {color:#d04437} at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62){color} > {color:#d04437} at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43){color} > {color:#d04437} at java.lang.reflect.Method.invoke(Method.java:498){color} > {color:#d04437} at > org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47){color} > {color:#d04437} at > org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12){color} > {color:#d04437} at > org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44){color} > {color:#d04437} at > org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17){color} > {color:#d04437} at > org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27){color} > {color:#d04437} at > org.junit.internal.runners.statements.FailOnTimeout$StatementThread.run(FailOnTimeout.java:74){color} > {color:#33}The involved tests are{color} > {code:java} >
[jira] [Commented] (HDFS-13560) Insufficient system resources exist to complete the requested service for some tests on Windows
[ https://issues.apache.org/jira/browse/HDFS-13560?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16479790#comment-16479790 ] Íñigo Goiri commented on HDFS-13560: In the last Yetus run, the [unit tests|https://builds.apache.org/job/PreCommit-HDFS-Build/24245/testReport/org.apache.hadoop.hdfs.server.datanode.fsdataset.impl/] now pass. [~huanbang1993] does this still work for Windows? > Insufficient system resources exist to complete the requested service for > some tests on Windows > --- > > Key: HDFS-13560 > URL: https://issues.apache.org/jira/browse/HDFS-13560 > Project: Hadoop HDFS > Issue Type: Bug >Reporter: Anbang Hu >Assignee: Anbang Hu >Priority: Major > Labels: Windows > Attachments: HDFS-13560.000.patch, HDFS-13560.001.patch, > HDFS-13560.002.patch, HDFS-13560.003.patch, HDFS-13560.004.patch > > > On Windows, there are 30 tests in HDFS component giving error like the > following: > {color:#d04437}[ERROR] Tests run: 7, Failures: 0, Errors: 7, Skipped: 0, > Time elapsed: 50.149 s <<< FAILURE! - in > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestLazyPersistFiles{color} > {color:#d04437} [ERROR] > testDisableLazyPersistFileScrubber(org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestLazyPersistFiles) > Time elapsed: 16.513 s <<< ERROR!{color} > {color:#d04437} 1450: Insufficient system resources exist to complete the > requested service.{color} > {color:#d04437}at > org.apache.hadoop.io.nativeio.NativeIO$Windows.extendWorkingSetSize(Native > Method){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:1339){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.(DataNode.java:495){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:2695){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:2598){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.java:1554){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:904){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster.(MiniDFSCluster.java:514){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster$Builder.build(MiniDFSCluster.java:473){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.LazyPersistTestCase.startUpCluster(LazyPersistTestCase.java:316){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.LazyPersistTestCase$ClusterWithRamDiskBuilder.build(LazyPersistTestCase.java:415){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestLazyPersistFiles.testDisableLazyPersistFileScrubber(TestLazyPersistFiles.java:128){color} > {color:#d04437} at sun.reflect.NativeMethodAccessorImpl.invoke0(Native > Method){color} > {color:#d04437} at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62){color} > {color:#d04437} at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43){color} > {color:#d04437} at java.lang.reflect.Method.invoke(Method.java:498){color} > {color:#d04437} at > org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47){color} > {color:#d04437} at > org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12){color} > {color:#d04437} at > org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44){color} > {color:#d04437} at > org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17){color} > {color:#d04437} at > org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27){color} > {color:#d04437} at > org.junit.internal.runners.statements.FailOnTimeout$StatementThread.run(FailOnTimeout.java:74){color} > {color:#33}The involved tests are{color} > {code:java} >
[jira] [Commented] (HDFS-13560) Insufficient system resources exist to complete the requested service for some tests on Windows
[ https://issues.apache.org/jira/browse/HDFS-13560?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16479735#comment-16479735 ] genericqa commented on HDFS-13560: -- | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 25s{color} | {color:blue} Docker mode activated. {color} | || || || || {color:brown} Prechecks {color} || | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s{color} | {color:green} The patch appears to include 3 new or modified test files. {color} | || || || || {color:brown} trunk Compile Tests {color} || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 1m 27s{color} | {color:blue} Maven dependency ordering for branch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 23m 50s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 26m 51s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 3m 35s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 2m 14s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 15m 56s{color} | {color:green} branch has no errors when building and testing our client artifacts. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 3m 16s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 53s{color} | {color:green} trunk passed {color} | || || || || {color:brown} Patch Compile Tests {color} || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 18s{color} | {color:blue} Maven dependency ordering for patch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 1m 36s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 25m 34s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 25m 34s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 3m 25s{color} | {color:green} root: The patch generated 0 new + 120 unchanged - 1 fixed = 120 total (was 121) {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 2m 0s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 9m 32s{color} | {color:green} patch has no errors when building and testing our client artifacts. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 3m 33s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 45s{color} | {color:green} the patch passed {color} | || || || || {color:brown} Other Tests {color} || | {color:green}+1{color} | {color:green} unit {color} | {color:green} 8m 13s{color} | {color:green} hadoop-common in the patch passed. {color} | | {color:red}-1{color} | {color:red} unit {color} | {color:red}111m 18s{color} | {color:red} hadoop-hdfs in the patch failed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 33s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black}244m 11s{color} | {color:black} {color} | \\ \\ || Reason || Tests || | Failed junit tests | hadoop.hdfs.server.namenode.metrics.TestNameNodeMetrics | | | hadoop.hdfs.server.namenode.ha.TestRetryCacheWithHA | | | hadoop.hdfs.web.TestWebHdfsTimeouts | | | hadoop.hdfs.client.impl.TestBlockReaderLocal | | | hadoop.hdfs.TestSafeModeWithStripedFile | | | hadoop.hdfs.server.datanode.TestDataNodeVolumeFailureReporting | | | hadoop.hdfs.web.TestWebHDFS | | | hadoop.hdfs.TestPread | \\ \\ || Subsystem || Report/Notes || | Docker | Client=17.05.0-ce Server=17.05.0-ce Image:yetus/hadoop:abb62dd | | JIRA Issue | HDFS-13560 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12923960/HDFS-13560.004.patch | | Optional Tests | asflicense compile javac javadoc mvninstall mvnsite unit shadedclient findbugs checkstyle | | uname | Linux c4cee548f8ca
[jira] [Commented] (HDFS-13560) Insufficient system resources exist to complete the requested service for some tests on Windows
[ https://issues.apache.org/jira/browse/HDFS-13560?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16479648#comment-16479648 ] Giovanni Matteo Fumarola commented on HDFS-13560: - Thanks [~huanbang1993] for the patch. LGTM +1 > Insufficient system resources exist to complete the requested service for > some tests on Windows > --- > > Key: HDFS-13560 > URL: https://issues.apache.org/jira/browse/HDFS-13560 > Project: Hadoop HDFS > Issue Type: Bug >Reporter: Anbang Hu >Assignee: Anbang Hu >Priority: Major > Labels: Windows > Attachments: HDFS-13560.000.patch, HDFS-13560.001.patch, > HDFS-13560.002.patch, HDFS-13560.003.patch, HDFS-13560.004.patch > > > On Windows, there are 30 tests in HDFS component giving error like the > following: > {color:#d04437}[ERROR] Tests run: 7, Failures: 0, Errors: 7, Skipped: 0, > Time elapsed: 50.149 s <<< FAILURE! - in > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestLazyPersistFiles{color} > {color:#d04437} [ERROR] > testDisableLazyPersistFileScrubber(org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestLazyPersistFiles) > Time elapsed: 16.513 s <<< ERROR!{color} > {color:#d04437} 1450: Insufficient system resources exist to complete the > requested service.{color} > {color:#d04437}at > org.apache.hadoop.io.nativeio.NativeIO$Windows.extendWorkingSetSize(Native > Method){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:1339){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.(DataNode.java:495){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:2695){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:2598){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.java:1554){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:904){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster.(MiniDFSCluster.java:514){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster$Builder.build(MiniDFSCluster.java:473){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.LazyPersistTestCase.startUpCluster(LazyPersistTestCase.java:316){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.LazyPersistTestCase$ClusterWithRamDiskBuilder.build(LazyPersistTestCase.java:415){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestLazyPersistFiles.testDisableLazyPersistFileScrubber(TestLazyPersistFiles.java:128){color} > {color:#d04437} at sun.reflect.NativeMethodAccessorImpl.invoke0(Native > Method){color} > {color:#d04437} at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62){color} > {color:#d04437} at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43){color} > {color:#d04437} at java.lang.reflect.Method.invoke(Method.java:498){color} > {color:#d04437} at > org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47){color} > {color:#d04437} at > org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12){color} > {color:#d04437} at > org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44){color} > {color:#d04437} at > org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17){color} > {color:#d04437} at > org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27){color} > {color:#d04437} at > org.junit.internal.runners.statements.FailOnTimeout$StatementThread.run(FailOnTimeout.java:74){color} > {color:#33}The involved tests are{color} > {code:java} >
[jira] [Commented] (HDFS-13560) Insufficient system resources exist to complete the requested service for some tests on Windows
[ https://issues.apache.org/jira/browse/HDFS-13560?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16479410#comment-16479410 ] Anbang Hu commented on HDFS-13560: -- I think maintaining one function in which Windows limit is capped at Integer.MAX_VALUE is better than having two functions targeting Windows and POSIX separately. Please kindly review [^HDFS-13560.004.patch] > Insufficient system resources exist to complete the requested service for > some tests on Windows > --- > > Key: HDFS-13560 > URL: https://issues.apache.org/jira/browse/HDFS-13560 > Project: Hadoop HDFS > Issue Type: Bug >Reporter: Anbang Hu >Assignee: Anbang Hu >Priority: Major > Labels: Windows > Attachments: HDFS-13560.000.patch, HDFS-13560.001.patch, > HDFS-13560.002.patch, HDFS-13560.003.patch, HDFS-13560.004.patch > > > On Windows, there are 30 tests in HDFS component giving error like the > following: > {color:#d04437}[ERROR] Tests run: 7, Failures: 0, Errors: 7, Skipped: 0, > Time elapsed: 50.149 s <<< FAILURE! - in > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestLazyPersistFiles{color} > {color:#d04437} [ERROR] > testDisableLazyPersistFileScrubber(org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestLazyPersistFiles) > Time elapsed: 16.513 s <<< ERROR!{color} > {color:#d04437} 1450: Insufficient system resources exist to complete the > requested service.{color} > {color:#d04437}at > org.apache.hadoop.io.nativeio.NativeIO$Windows.extendWorkingSetSize(Native > Method){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:1339){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.(DataNode.java:495){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:2695){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:2598){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.java:1554){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:904){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster.(MiniDFSCluster.java:514){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster$Builder.build(MiniDFSCluster.java:473){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.LazyPersistTestCase.startUpCluster(LazyPersistTestCase.java:316){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.LazyPersistTestCase$ClusterWithRamDiskBuilder.build(LazyPersistTestCase.java:415){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestLazyPersistFiles.testDisableLazyPersistFileScrubber(TestLazyPersistFiles.java:128){color} > {color:#d04437} at sun.reflect.NativeMethodAccessorImpl.invoke0(Native > Method){color} > {color:#d04437} at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62){color} > {color:#d04437} at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43){color} > {color:#d04437} at java.lang.reflect.Method.invoke(Method.java:498){color} > {color:#d04437} at > org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47){color} > {color:#d04437} at > org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12){color} > {color:#d04437} at > org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44){color} > {color:#d04437} at > org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17){color} > {color:#d04437} at > org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27){color} > {color:#d04437} at > org.junit.internal.runners.statements.FailOnTimeout$StatementThread.run(FailOnTimeout.java:74){color} > {color:#33}The involved tests are{color} > {code:java} >
[jira] [Commented] (HDFS-13560) Insufficient system resources exist to complete the requested service for some tests on Windows
[ https://issues.apache.org/jira/browse/HDFS-13560?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16479300#comment-16479300 ] Íñigo Goiri commented on HDFS-13560: Let's split this into two functions and have one of them do the POSIX call. > Insufficient system resources exist to complete the requested service for > some tests on Windows > --- > > Key: HDFS-13560 > URL: https://issues.apache.org/jira/browse/HDFS-13560 > Project: Hadoop HDFS > Issue Type: Bug >Reporter: Anbang Hu >Assignee: Anbang Hu >Priority: Major > Labels: Windows > Attachments: HDFS-13560.000.patch, HDFS-13560.001.patch, > HDFS-13560.002.patch, HDFS-13560.003.patch > > > On Windows, there are 30 tests in HDFS component giving error like the > following: > {color:#d04437}[ERROR] Tests run: 7, Failures: 0, Errors: 7, Skipped: 0, > Time elapsed: 50.149 s <<< FAILURE! - in > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestLazyPersistFiles{color} > {color:#d04437} [ERROR] > testDisableLazyPersistFileScrubber(org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestLazyPersistFiles) > Time elapsed: 16.513 s <<< ERROR!{color} > {color:#d04437} 1450: Insufficient system resources exist to complete the > requested service.{color} > {color:#d04437}at > org.apache.hadoop.io.nativeio.NativeIO$Windows.extendWorkingSetSize(Native > Method){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:1339){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.(DataNode.java:495){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:2695){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:2598){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.java:1554){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:904){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster.(MiniDFSCluster.java:514){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster$Builder.build(MiniDFSCluster.java:473){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.LazyPersistTestCase.startUpCluster(LazyPersistTestCase.java:316){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.LazyPersistTestCase$ClusterWithRamDiskBuilder.build(LazyPersistTestCase.java:415){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestLazyPersistFiles.testDisableLazyPersistFileScrubber(TestLazyPersistFiles.java:128){color} > {color:#d04437} at sun.reflect.NativeMethodAccessorImpl.invoke0(Native > Method){color} > {color:#d04437} at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62){color} > {color:#d04437} at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43){color} > {color:#d04437} at java.lang.reflect.Method.invoke(Method.java:498){color} > {color:#d04437} at > org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47){color} > {color:#d04437} at > org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12){color} > {color:#d04437} at > org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44){color} > {color:#d04437} at > org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17){color} > {color:#d04437} at > org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27){color} > {color:#d04437} at > org.junit.internal.runners.statements.FailOnTimeout$StatementThread.run(FailOnTimeout.java:74){color} > {color:#33}The involved tests are{color} > {code:java} >
[jira] [Commented] (HDFS-13560) Insufficient system resources exist to complete the requested service for some tests on Windows
[ https://issues.apache.org/jira/browse/HDFS-13560?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16478595#comment-16478595 ] genericqa commented on HDFS-13560: -- | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 32s{color} | {color:blue} Docker mode activated. {color} | || || || || {color:brown} Prechecks {color} || | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s{color} | {color:green} The patch appears to include 3 new or modified test files. {color} | || || || || {color:brown} trunk Compile Tests {color} || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 1m 38s{color} | {color:blue} Maven dependency ordering for branch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 26m 6s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 30m 49s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 3m 20s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 2m 24s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 16m 32s{color} | {color:green} branch has no errors when building and testing our client artifacts. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 3m 44s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 51s{color} | {color:green} trunk passed {color} | || || || || {color:brown} Patch Compile Tests {color} || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 19s{color} | {color:blue} Maven dependency ordering for patch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 2m 6s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 34m 42s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 34m 42s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 3m 40s{color} | {color:green} root: The patch generated 0 new + 120 unchanged - 1 fixed = 120 total (was 121) {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 2m 37s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 11m 43s{color} | {color:green} patch has no errors when building and testing our client artifacts. {color} | | {color:red}-1{color} | {color:red} findbugs {color} | {color:red} 2m 8s{color} | {color:red} hadoop-common-project/hadoop-common generated 1 new + 0 unchanged - 0 fixed = 1 total (was 0) {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 2m 30s{color} | {color:green} the patch passed {color} | || || || || {color:brown} Other Tests {color} || | {color:green}+1{color} | {color:green} unit {color} | {color:green} 10m 39s{color} | {color:green} hadoop-common in the patch passed. {color} | | {color:red}-1{color} | {color:red} unit {color} | {color:red}109m 9s{color} | {color:red} hadoop-hdfs in the patch failed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 44s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black}267m 7s{color} | {color:black} {color} | \\ \\ || Reason || Tests || | FindBugs | module:hadoop-common-project/hadoop-common | | | Boxed value is unboxed and then immediately reboxed in org.apache.hadoop.util.Shell.getMemlockLimit(Long) At Shell.java:then immediately reboxed in org.apache.hadoop.util.Shell.getMemlockLimit(Long) At Shell.java:[line 1408] | | Failed junit tests | hadoop.hdfs.TestDFSStripedOutputStreamWithFailureWithRandomECPolicy | \\ \\ || Subsystem || Report/Notes || | Docker | Client=17.05.0-ce Server=17.05.0-ce Image:yetus/hadoop:abb62dd | | JIRA Issue | HDFS-13560 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12923829/HDFS-13560.003.patch | | Optional Tests | asflicense compile javac javadoc mvninstall mvnsite unit shadedclient findbugs
[jira] [Commented] (HDFS-13560) Insufficient system resources exist to complete the requested service for some tests on Windows
[ https://issues.apache.org/jira/browse/HDFS-13560?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16478363#comment-16478363 ] Anbang Hu commented on HDFS-13560: -- These failures might be related to missing Long.MAX_VALUE. Updated patch with [^HDFS-13560.003.patch]. > Insufficient system resources exist to complete the requested service for > some tests on Windows > --- > > Key: HDFS-13560 > URL: https://issues.apache.org/jira/browse/HDFS-13560 > Project: Hadoop HDFS > Issue Type: Bug >Reporter: Anbang Hu >Assignee: Anbang Hu >Priority: Major > Labels: Windows > Attachments: HDFS-13560.000.patch, HDFS-13560.001.patch, > HDFS-13560.002.patch, HDFS-13560.003.patch > > > On Windows, there are 30 tests in HDFS component giving error like the > following: > {color:#d04437}[ERROR] Tests run: 7, Failures: 0, Errors: 7, Skipped: 0, > Time elapsed: 50.149 s <<< FAILURE! - in > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestLazyPersistFiles{color} > {color:#d04437} [ERROR] > testDisableLazyPersistFileScrubber(org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestLazyPersistFiles) > Time elapsed: 16.513 s <<< ERROR!{color} > {color:#d04437} 1450: Insufficient system resources exist to complete the > requested service.{color} > {color:#d04437}at > org.apache.hadoop.io.nativeio.NativeIO$Windows.extendWorkingSetSize(Native > Method){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:1339){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.(DataNode.java:495){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:2695){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:2598){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.java:1554){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:904){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster.(MiniDFSCluster.java:514){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster$Builder.build(MiniDFSCluster.java:473){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.LazyPersistTestCase.startUpCluster(LazyPersistTestCase.java:316){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.LazyPersistTestCase$ClusterWithRamDiskBuilder.build(LazyPersistTestCase.java:415){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestLazyPersistFiles.testDisableLazyPersistFileScrubber(TestLazyPersistFiles.java:128){color} > {color:#d04437} at sun.reflect.NativeMethodAccessorImpl.invoke0(Native > Method){color} > {color:#d04437} at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62){color} > {color:#d04437} at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43){color} > {color:#d04437} at java.lang.reflect.Method.invoke(Method.java:498){color} > {color:#d04437} at > org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47){color} > {color:#d04437} at > org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12){color} > {color:#d04437} at > org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44){color} > {color:#d04437} at > org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17){color} > {color:#d04437} at > org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27){color} > {color:#d04437} at > org.junit.internal.runners.statements.FailOnTimeout$StatementThread.run(FailOnTimeout.java:74){color} > {color:#33}The involved tests are{color} > {code:java} >
[jira] [Commented] (HDFS-13560) Insufficient system resources exist to complete the requested service for some tests on Windows
[ https://issues.apache.org/jira/browse/HDFS-13560?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16478350#comment-16478350 ] Anbang Hu commented on HDFS-13560: -- hadoop.hdfs.server.datanode.TestDataNodeVolumeFailureReporting doesn't seem to be related. However, hadoop.hdfs.server.datanode.TestDirectoryScanner might be related. > Insufficient system resources exist to complete the requested service for > some tests on Windows > --- > > Key: HDFS-13560 > URL: https://issues.apache.org/jira/browse/HDFS-13560 > Project: Hadoop HDFS > Issue Type: Bug >Reporter: Anbang Hu >Assignee: Anbang Hu >Priority: Major > Labels: Windows > Attachments: HDFS-13560.000.patch, HDFS-13560.001.patch, > HDFS-13560.002.patch > > > On Windows, there are 30 tests in HDFS component giving error like the > following: > {color:#d04437}[ERROR] Tests run: 7, Failures: 0, Errors: 7, Skipped: 0, > Time elapsed: 50.149 s <<< FAILURE! - in > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestLazyPersistFiles{color} > {color:#d04437} [ERROR] > testDisableLazyPersistFileScrubber(org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestLazyPersistFiles) > Time elapsed: 16.513 s <<< ERROR!{color} > {color:#d04437} 1450: Insufficient system resources exist to complete the > requested service.{color} > {color:#d04437}at > org.apache.hadoop.io.nativeio.NativeIO$Windows.extendWorkingSetSize(Native > Method){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:1339){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.(DataNode.java:495){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:2695){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:2598){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.java:1554){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:904){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster.(MiniDFSCluster.java:514){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster$Builder.build(MiniDFSCluster.java:473){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.LazyPersistTestCase.startUpCluster(LazyPersistTestCase.java:316){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.LazyPersistTestCase$ClusterWithRamDiskBuilder.build(LazyPersistTestCase.java:415){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestLazyPersistFiles.testDisableLazyPersistFileScrubber(TestLazyPersistFiles.java:128){color} > {color:#d04437} at sun.reflect.NativeMethodAccessorImpl.invoke0(Native > Method){color} > {color:#d04437} at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62){color} > {color:#d04437} at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43){color} > {color:#d04437} at java.lang.reflect.Method.invoke(Method.java:498){color} > {color:#d04437} at > org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47){color} > {color:#d04437} at > org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12){color} > {color:#d04437} at > org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44){color} > {color:#d04437} at > org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17){color} > {color:#d04437} at > org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27){color} > {color:#d04437} at > org.junit.internal.runners.statements.FailOnTimeout$StatementThread.run(FailOnTimeout.java:74){color} > {color:#33}The involved tests are{color} > {code:java} >
[jira] [Commented] (HDFS-13560) Insufficient system resources exist to complete the requested service for some tests on Windows
[ https://issues.apache.org/jira/browse/HDFS-13560?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16478347#comment-16478347 ] Íñigo Goiri commented on HDFS-13560: The following failed unit tests seem suspicious: * hadoop.hdfs.server.datanode.fsdataset.impl.TestScrLazyPersistFiles * hadoop.hdfs.server.datanode.fsdataset.impl.TestLazyPersistReplicaRecovery * hadoop.hdfs.server.datanode.TestDataNodeVolumeFailureReporting Not sure what the issue is but it's definitely related to this patch. > Insufficient system resources exist to complete the requested service for > some tests on Windows > --- > > Key: HDFS-13560 > URL: https://issues.apache.org/jira/browse/HDFS-13560 > Project: Hadoop HDFS > Issue Type: Bug >Reporter: Anbang Hu >Assignee: Anbang Hu >Priority: Major > Labels: Windows > Attachments: HDFS-13560.000.patch, HDFS-13560.001.patch, > HDFS-13560.002.patch > > > On Windows, there are 30 tests in HDFS component giving error like the > following: > {color:#d04437}[ERROR] Tests run: 7, Failures: 0, Errors: 7, Skipped: 0, > Time elapsed: 50.149 s <<< FAILURE! - in > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestLazyPersistFiles{color} > {color:#d04437} [ERROR] > testDisableLazyPersistFileScrubber(org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestLazyPersistFiles) > Time elapsed: 16.513 s <<< ERROR!{color} > {color:#d04437} 1450: Insufficient system resources exist to complete the > requested service.{color} > {color:#d04437}at > org.apache.hadoop.io.nativeio.NativeIO$Windows.extendWorkingSetSize(Native > Method){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:1339){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.(DataNode.java:495){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:2695){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:2598){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.java:1554){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:904){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster.(MiniDFSCluster.java:514){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster$Builder.build(MiniDFSCluster.java:473){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.LazyPersistTestCase.startUpCluster(LazyPersistTestCase.java:316){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.LazyPersistTestCase$ClusterWithRamDiskBuilder.build(LazyPersistTestCase.java:415){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestLazyPersistFiles.testDisableLazyPersistFileScrubber(TestLazyPersistFiles.java:128){color} > {color:#d04437} at sun.reflect.NativeMethodAccessorImpl.invoke0(Native > Method){color} > {color:#d04437} at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62){color} > {color:#d04437} at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43){color} > {color:#d04437} at java.lang.reflect.Method.invoke(Method.java:498){color} > {color:#d04437} at > org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47){color} > {color:#d04437} at > org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12){color} > {color:#d04437} at > org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44){color} > {color:#d04437} at > org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17){color} > {color:#d04437} at > org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27){color} > {color:#d04437} at > org.junit.internal.runners.statements.FailOnTimeout$StatementThread.run(FailOnTimeout.java:74){color} > {color:#33}The involved tests are{color} > {code:java} >
[jira] [Commented] (HDFS-13560) Insufficient system resources exist to complete the requested service for some tests on Windows
[ https://issues.apache.org/jira/browse/HDFS-13560?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16478334#comment-16478334 ] genericqa commented on HDFS-13560: -- | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 49s{color} | {color:blue} Docker mode activated. {color} | || || || || {color:brown} Prechecks {color} || | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s{color} | {color:green} The patch appears to include 3 new or modified test files. {color} | || || || || {color:brown} trunk Compile Tests {color} || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 24s{color} | {color:blue} Maven dependency ordering for branch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 27m 50s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 32m 0s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 3m 22s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 2m 49s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 18m 43s{color} | {color:green} branch has no errors when building and testing our client artifacts. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 4m 13s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 2m 11s{color} | {color:green} trunk passed {color} | || || || || {color:brown} Patch Compile Tests {color} || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 30s{color} | {color:blue} Maven dependency ordering for patch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 2m 18s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 38m 42s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 38m 42s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 3m 43s{color} | {color:green} root: The patch generated 0 new + 120 unchanged - 1 fixed = 120 total (was 121) {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 2m 58s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 11m 48s{color} | {color:green} patch has no errors when building and testing our client artifacts. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 4m 46s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 2m 25s{color} | {color:green} the patch passed {color} | || || || || {color:brown} Other Tests {color} || | {color:green}+1{color} | {color:green} unit {color} | {color:green} 11m 26s{color} | {color:green} hadoop-common in the patch passed. {color} | | {color:red}-1{color} | {color:red} unit {color} | {color:red}104m 44s{color} | {color:red} hadoop-hdfs in the patch failed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 44s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black}274m 19s{color} | {color:black} {color} | \\ \\ || Reason || Tests || | Failed junit tests | hadoop.hdfs.server.datanode.fsdataset.impl.TestLazyPersistFiles | | | hadoop.hdfs.TestDFSStripedOutputStreamWithFailureWithRandomECPolicy | | | hadoop.hdfs.server.namenode.ha.TestRetryCacheWithHA | | | hadoop.hdfs.TestSafeMode | | | hadoop.hdfs.server.blockmanagement.TestRBWBlockInvalidation | | | hadoop.hdfs.server.datanode.fsdataset.impl.TestScrLazyPersistFiles | | | hadoop.hdfs.server.datanode.TestDataNodeVolumeFailureReporting | | | hadoop.hdfs.server.datanode.fsdataset.impl.TestLazyPersistReplicaRecovery | | | hadoop.hdfs.server.datanode.TestDirectoryScanner | \\ \\ || Subsystem || Report/Notes || | Docker | Client=17.05.0-ce Server=17.05.0-ce Image:yetus/hadoop:abb62dd | | JIRA Issue | HDFS-13560 | | JIRA Patch URL |
[jira] [Commented] (HDFS-13560) Insufficient system resources exist to complete the requested service for some tests on Windows
[ https://issues.apache.org/jira/browse/HDFS-13560?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16478102#comment-16478102 ] Íñigo Goiri commented on HDFS-13560: I think the approach in [^HDFS-13560.002.patch] is correct. Let's see what Yetus says. > Insufficient system resources exist to complete the requested service for > some tests on Windows > --- > > Key: HDFS-13560 > URL: https://issues.apache.org/jira/browse/HDFS-13560 > Project: Hadoop HDFS > Issue Type: Bug >Reporter: Anbang Hu >Assignee: Anbang Hu >Priority: Major > Labels: Windows > Attachments: HDFS-13560.000.patch, HDFS-13560.001.patch, > HDFS-13560.002.patch > > > On Windows, there are 30 tests in HDFS component giving error like the > following: > {color:#d04437}[ERROR] Tests run: 7, Failures: 0, Errors: 7, Skipped: 0, > Time elapsed: 50.149 s <<< FAILURE! - in > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestLazyPersistFiles{color} > {color:#d04437} [ERROR] > testDisableLazyPersistFileScrubber(org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestLazyPersistFiles) > Time elapsed: 16.513 s <<< ERROR!{color} > {color:#d04437} 1450: Insufficient system resources exist to complete the > requested service.{color} > {color:#d04437}at > org.apache.hadoop.io.nativeio.NativeIO$Windows.extendWorkingSetSize(Native > Method){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:1339){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.(DataNode.java:495){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:2695){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:2598){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.java:1554){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:904){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster.(MiniDFSCluster.java:514){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster$Builder.build(MiniDFSCluster.java:473){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.LazyPersistTestCase.startUpCluster(LazyPersistTestCase.java:316){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.LazyPersistTestCase$ClusterWithRamDiskBuilder.build(LazyPersistTestCase.java:415){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestLazyPersistFiles.testDisableLazyPersistFileScrubber(TestLazyPersistFiles.java:128){color} > {color:#d04437} at sun.reflect.NativeMethodAccessorImpl.invoke0(Native > Method){color} > {color:#d04437} at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62){color} > {color:#d04437} at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43){color} > {color:#d04437} at java.lang.reflect.Method.invoke(Method.java:498){color} > {color:#d04437} at > org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47){color} > {color:#d04437} at > org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12){color} > {color:#d04437} at > org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44){color} > {color:#d04437} at > org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17){color} > {color:#d04437} at > org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27){color} > {color:#d04437} at > org.junit.internal.runners.statements.FailOnTimeout$StatementThread.run(FailOnTimeout.java:74){color} > {color:#33}The involved tests are{color} > {code:java} >
[jira] [Commented] (HDFS-13560) Insufficient system resources exist to complete the requested service for some tests on Windows
[ https://issues.apache.org/jira/browse/HDFS-13560?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16478060#comment-16478060 ] Anbang Hu commented on HDFS-13560: -- Thanks [~elgoiri] for the suggestion. I removed timeout change in [^HDFS-13560.002.patch]. > Insufficient system resources exist to complete the requested service for > some tests on Windows > --- > > Key: HDFS-13560 > URL: https://issues.apache.org/jira/browse/HDFS-13560 > Project: Hadoop HDFS > Issue Type: Bug >Reporter: Anbang Hu >Assignee: Anbang Hu >Priority: Major > Labels: Windows > Attachments: HDFS-13560.000.patch, HDFS-13560.001.patch, > HDFS-13560.002.patch > > > On Windows, there are 30 tests in HDFS component giving error like the > following: > {color:#d04437}[ERROR] Tests run: 7, Failures: 0, Errors: 7, Skipped: 0, > Time elapsed: 50.149 s <<< FAILURE! - in > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestLazyPersistFiles{color} > {color:#d04437} [ERROR] > testDisableLazyPersistFileScrubber(org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestLazyPersistFiles) > Time elapsed: 16.513 s <<< ERROR!{color} > {color:#d04437} 1450: Insufficient system resources exist to complete the > requested service.{color} > {color:#d04437}at > org.apache.hadoop.io.nativeio.NativeIO$Windows.extendWorkingSetSize(Native > Method){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:1339){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.(DataNode.java:495){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:2695){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:2598){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.java:1554){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:904){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster.(MiniDFSCluster.java:514){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster$Builder.build(MiniDFSCluster.java:473){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.LazyPersistTestCase.startUpCluster(LazyPersistTestCase.java:316){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.LazyPersistTestCase$ClusterWithRamDiskBuilder.build(LazyPersistTestCase.java:415){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestLazyPersistFiles.testDisableLazyPersistFileScrubber(TestLazyPersistFiles.java:128){color} > {color:#d04437} at sun.reflect.NativeMethodAccessorImpl.invoke0(Native > Method){color} > {color:#d04437} at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62){color} > {color:#d04437} at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43){color} > {color:#d04437} at java.lang.reflect.Method.invoke(Method.java:498){color} > {color:#d04437} at > org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47){color} > {color:#d04437} at > org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12){color} > {color:#d04437} at > org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44){color} > {color:#d04437} at > org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17){color} > {color:#d04437} at > org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27){color} > {color:#d04437} at > org.junit.internal.runners.statements.FailOnTimeout$StatementThread.run(FailOnTimeout.java:74){color} > {color:#33}The involved tests are{color} > {code:java} >
[jira] [Commented] (HDFS-13560) Insufficient system resources exist to complete the requested service for some tests on Windows
[ https://issues.apache.org/jira/browse/HDFS-13560?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16477675#comment-16477675 ] Íñigo Goiri commented on HDFS-13560: I think adding the mem check depending on the platform makes sense. * I would avoid the increase of the timeout in this JIRA, we can open a separate one later on. * I'm not sure if Shell is the right place for the utility function. There are similar methods already there, so if there is no other proposals we can leave it there. [~cnauroth] you did some work here, could you help verify this is correct? > Insufficient system resources exist to complete the requested service for > some tests on Windows > --- > > Key: HDFS-13560 > URL: https://issues.apache.org/jira/browse/HDFS-13560 > Project: Hadoop HDFS > Issue Type: Bug >Reporter: Anbang Hu >Assignee: Anbang Hu >Priority: Major > Labels: Windows > Attachments: HDFS-13560.000.patch, HDFS-13560.001.patch > > > On Windows, there are 30 tests in HDFS component giving error like the > following: > {color:#d04437}[ERROR] Tests run: 7, Failures: 0, Errors: 7, Skipped: 0, > Time elapsed: 50.149 s <<< FAILURE! - in > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestLazyPersistFiles{color} > {color:#d04437} [ERROR] > testDisableLazyPersistFileScrubber(org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestLazyPersistFiles) > Time elapsed: 16.513 s <<< ERROR!{color} > {color:#d04437} 1450: Insufficient system resources exist to complete the > requested service.{color} > {color:#d04437}at > org.apache.hadoop.io.nativeio.NativeIO$Windows.extendWorkingSetSize(Native > Method){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:1339){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.(DataNode.java:495){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:2695){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:2598){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.java:1554){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:904){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster.(MiniDFSCluster.java:514){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster$Builder.build(MiniDFSCluster.java:473){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.LazyPersistTestCase.startUpCluster(LazyPersistTestCase.java:316){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.LazyPersistTestCase$ClusterWithRamDiskBuilder.build(LazyPersistTestCase.java:415){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestLazyPersistFiles.testDisableLazyPersistFileScrubber(TestLazyPersistFiles.java:128){color} > {color:#d04437} at sun.reflect.NativeMethodAccessorImpl.invoke0(Native > Method){color} > {color:#d04437} at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62){color} > {color:#d04437} at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43){color} > {color:#d04437} at java.lang.reflect.Method.invoke(Method.java:498){color} > {color:#d04437} at > org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47){color} > {color:#d04437} at > org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12){color} > {color:#d04437} at > org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44){color} > {color:#d04437} at > org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17){color} > {color:#d04437} at > org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27){color} > {color:#d04437} at > org.junit.internal.runners.statements.FailOnTimeout$StatementThread.run(FailOnTimeout.java:74){color} > {color:#33}The involved tests are{color} > {code:java} >
[jira] [Commented] (HDFS-13560) Insufficient system resources exist to complete the requested service for some tests on Windows
[ https://issues.apache.org/jira/browse/HDFS-13560?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16476957#comment-16476957 ] genericqa commented on HDFS-13560: -- | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 24s{color} | {color:blue} Docker mode activated. {color} | || || || || {color:brown} Prechecks {color} || | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s{color} | {color:green} The patch appears to include 3 new or modified test files. {color} | || || || || {color:brown} trunk Compile Tests {color} || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 1m 38s{color} | {color:blue} Maven dependency ordering for branch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 25m 45s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 29m 29s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 3m 15s{color} | {color:green} trunk passed {color} | | {color:red}-1{color} | {color:red} mvnsite {color} | {color:red} 2m 31s{color} | {color:red} hadoop-common in trunk failed. {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 18m 7s{color} | {color:green} branch has no errors when building and testing our client artifacts. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 3m 38s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 53s{color} | {color:green} trunk passed {color} | || || || || {color:brown} Patch Compile Tests {color} || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 18s{color} | {color:blue} Maven dependency ordering for patch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 1m 51s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 28m 12s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 28m 12s{color} | {color:green} the patch passed {color} | | {color:orange}-0{color} | {color:orange} checkstyle {color} | {color:orange} 3m 13s{color} | {color:orange} root: The patch generated 1 new + 119 unchanged - 2 fixed = 120 total (was 121) {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 2m 26s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 10m 34s{color} | {color:green} patch has no errors when building and testing our client artifacts. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 4m 7s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 53s{color} | {color:green} the patch passed {color} | || || || || {color:brown} Other Tests {color} || | {color:green}+1{color} | {color:green} unit {color} | {color:green} 9m 17s{color} | {color:green} hadoop-common in the patch passed. {color} | | {color:red}-1{color} | {color:red} unit {color} | {color:red}106m 38s{color} | {color:red} hadoop-hdfs in the patch failed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 50s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black}253m 2s{color} | {color:black} {color} | \\ \\ || Reason || Tests || | Failed junit tests | hadoop.hdfs.server.datanode.fsdataset.impl.TestLazyPersistFiles | | | hadoop.hdfs.TestReconstructStripedFileWithRandomECPolicy | | | hadoop.hdfs.server.datanode.fsdataset.impl.TestScrLazyPersistFiles | | | hadoop.hdfs.server.datanode.TestDataNodeVolumeFailureReporting | | | hadoop.hdfs.server.datanode.fsdataset.impl.TestLazyPersistReplicaRecovery | \\ \\ || Subsystem || Report/Notes || | Docker | Client=17.05.0-ce Server=17.05.0-ce Image:yetus/hadoop:abb62dd | | JIRA Issue | HDFS-13560 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12923582/HDFS-13560.001.patch | | Optional Tests | asflicense compile javac javadoc mvninstall mvnsite unit shadedclient findbugs checkstyle | | uname | Linux 683e410bc02f 3.13.0-143-generic
[jira] [Commented] (HDFS-13560) Insufficient system resources exist to complete the requested service for some tests on Windows
[ https://issues.apache.org/jira/browse/HDFS-13560?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16476777#comment-16476777 ] Anbang Hu commented on HDFS-13560: -- [^HDFS-13560.001.patch] is refactored version according to [~elgoiri]'s suggestion. > Insufficient system resources exist to complete the requested service for > some tests on Windows > --- > > Key: HDFS-13560 > URL: https://issues.apache.org/jira/browse/HDFS-13560 > Project: Hadoop HDFS > Issue Type: Bug >Reporter: Anbang Hu >Assignee: Anbang Hu >Priority: Major > Labels: Windows > Attachments: HDFS-13560.000.patch, HDFS-13560.001.patch > > > On Windows, there are 30 tests in HDFS component giving error like the > following: > {color:#d04437}[ERROR] Tests run: 7, Failures: 0, Errors: 7, Skipped: 0, > Time elapsed: 50.149 s <<< FAILURE! - in > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestLazyPersistFiles{color} > {color:#d04437} [ERROR] > testDisableLazyPersistFileScrubber(org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestLazyPersistFiles) > Time elapsed: 16.513 s <<< ERROR!{color} > {color:#d04437} 1450: Insufficient system resources exist to complete the > requested service.{color} > {color:#d04437}at > org.apache.hadoop.io.nativeio.NativeIO$Windows.extendWorkingSetSize(Native > Method){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:1339){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.(DataNode.java:495){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:2695){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:2598){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.java:1554){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:904){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster.(MiniDFSCluster.java:514){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster$Builder.build(MiniDFSCluster.java:473){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.LazyPersistTestCase.startUpCluster(LazyPersistTestCase.java:316){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.LazyPersistTestCase$ClusterWithRamDiskBuilder.build(LazyPersistTestCase.java:415){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestLazyPersistFiles.testDisableLazyPersistFileScrubber(TestLazyPersistFiles.java:128){color} > {color:#d04437} at sun.reflect.NativeMethodAccessorImpl.invoke0(Native > Method){color} > {color:#d04437} at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62){color} > {color:#d04437} at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43){color} > {color:#d04437} at java.lang.reflect.Method.invoke(Method.java:498){color} > {color:#d04437} at > org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47){color} > {color:#d04437} at > org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12){color} > {color:#d04437} at > org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44){color} > {color:#d04437} at > org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17){color} > {color:#d04437} at > org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27){color} > {color:#d04437} at > org.junit.internal.runners.statements.FailOnTimeout$StatementThread.run(FailOnTimeout.java:74){color} > {color:#33}The involved tests are{color} > {code:java} >
[jira] [Commented] (HDFS-13560) Insufficient system resources exist to complete the requested service for some tests on Windows
[ https://issues.apache.org/jira/browse/HDFS-13560?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16476749#comment-16476749 ] Anbang Hu commented on HDFS-13560: -- In startDataNode in DataNode.java: {code:java} if (dnConf.maxLockedMemory > 0) { if (!NativeIO.POSIX.getCacheManipulator().verifyCanMlock()) { throw new RuntimeException(String.format( "Cannot start datanode because the configured max locked memory" + " size (%s) is greater than zero and native code is not available.", DFS_DATANODE_MAX_LOCKED_MEMORY_KEY)); } if (Path.WINDOWS) { NativeIO.Windows.extendWorkingSetSize(dnConf.maxLockedMemory); } else { long ulimit = NativeIO.POSIX.getCacheManipulator().getMemlockLimit(); if (dnConf.maxLockedMemory > ulimit) { throw new RuntimeException(String.format( "Cannot start datanode because the configured max locked memory" + " size (%s) of %d bytes is more than the datanode's available" + " RLIMIT_MEMLOCK ulimit of %d bytes.", DFS_DATANODE_MAX_LOCKED_MEMORY_KEY, dnConf.maxLockedMemory, ulimit)); } } {code} [~cnauroth] do you think we should add before extendWorkingSetSize for Windows here? Not sure what ulimit to get for Windows though. > Insufficient system resources exist to complete the requested service for > some tests on Windows > --- > > Key: HDFS-13560 > URL: https://issues.apache.org/jira/browse/HDFS-13560 > Project: Hadoop HDFS > Issue Type: Bug >Reporter: Anbang Hu >Assignee: Anbang Hu >Priority: Major > Labels: Windows > Attachments: HDFS-13560.000.patch > > > On Windows, there are 30 tests in HDFS component giving error like the > following: > {color:#d04437}[ERROR] Tests run: 7, Failures: 0, Errors: 7, Skipped: 0, > Time elapsed: 50.149 s <<< FAILURE! - in > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestLazyPersistFiles{color} > {color:#d04437} [ERROR] > testDisableLazyPersistFileScrubber(org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestLazyPersistFiles) > Time elapsed: 16.513 s <<< ERROR!{color} > {color:#d04437} 1450: Insufficient system resources exist to complete the > requested service.{color} > {color:#d04437}at > org.apache.hadoop.io.nativeio.NativeIO$Windows.extendWorkingSetSize(Native > Method){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:1339){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.(DataNode.java:495){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:2695){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:2598){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.java:1554){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:904){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster.(MiniDFSCluster.java:514){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster$Builder.build(MiniDFSCluster.java:473){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.LazyPersistTestCase.startUpCluster(LazyPersistTestCase.java:316){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.LazyPersistTestCase$ClusterWithRamDiskBuilder.build(LazyPersistTestCase.java:415){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestLazyPersistFiles.testDisableLazyPersistFileScrubber(TestLazyPersistFiles.java:128){color} > {color:#d04437} at sun.reflect.NativeMethodAccessorImpl.invoke0(Native > Method){color} > {color:#d04437} at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62){color} > {color:#d04437} at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43){color} > {color:#d04437} at java.lang.reflect.Method.invoke(Method.java:498){color} > {color:#d04437} at > org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47){color} > {color:#d04437} at > org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12){color} > {color:#d04437} at > org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44){color} > {color:#d04437} at > org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17){color} > {color:#d04437} at > org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27){color} > {color:#d04437} at >
[jira] [Commented] (HDFS-13560) Insufficient system resources exist to complete the requested service for some tests on Windows
[ https://issues.apache.org/jira/browse/HDFS-13560?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16476149#comment-16476149 ] Íñigo Goiri commented on HDFS-13560: This issue seems like a new difference between Linux and Windows. So far we had: * Windows locks files and not cleaning up some components (e.g., MiniDFSCluster) locks other tests. * Differences in paths. Now we have this memory value; it might make sense to make that if a utility somewhere and use it for what's done in [^HDFS-13560.000.patch]. > Insufficient system resources exist to complete the requested service for > some tests on Windows > --- > > Key: HDFS-13560 > URL: https://issues.apache.org/jira/browse/HDFS-13560 > Project: Hadoop HDFS > Issue Type: Bug >Reporter: Anbang Hu >Assignee: Anbang Hu >Priority: Major > Labels: Windows > Attachments: HDFS-13560.000.patch > > > On Windows, there are 30 tests in HDFS component giving error like the > following: > {color:#d04437}[ERROR] Tests run: 7, Failures: 0, Errors: 7, Skipped: 0, > Time elapsed: 50.149 s <<< FAILURE! - in > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestLazyPersistFiles{color} > {color:#d04437} [ERROR] > testDisableLazyPersistFileScrubber(org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestLazyPersistFiles) > Time elapsed: 16.513 s <<< ERROR!{color} > {color:#d04437} 1450: Insufficient system resources exist to complete the > requested service.{color} > {color:#d04437}at > org.apache.hadoop.io.nativeio.NativeIO$Windows.extendWorkingSetSize(Native > Method){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:1339){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.(DataNode.java:495){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:2695){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:2598){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.java:1554){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:904){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster.(MiniDFSCluster.java:514){color} > {color:#d04437} at > org.apache.hadoop.hdfs.MiniDFSCluster$Builder.build(MiniDFSCluster.java:473){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.LazyPersistTestCase.startUpCluster(LazyPersistTestCase.java:316){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.LazyPersistTestCase$ClusterWithRamDiskBuilder.build(LazyPersistTestCase.java:415){color} > {color:#d04437} at > org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestLazyPersistFiles.testDisableLazyPersistFileScrubber(TestLazyPersistFiles.java:128){color} > {color:#d04437} at sun.reflect.NativeMethodAccessorImpl.invoke0(Native > Method){color} > {color:#d04437} at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62){color} > {color:#d04437} at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43){color} > {color:#d04437} at java.lang.reflect.Method.invoke(Method.java:498){color} > {color:#d04437} at > org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47){color} > {color:#d04437} at > org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12){color} > {color:#d04437} at > org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44){color} > {color:#d04437} at > org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17){color} > {color:#d04437} at > org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27){color} > {color:#d04437} at > org.junit.internal.runners.statements.FailOnTimeout$StatementThread.run(FailOnTimeout.java:74){color} > {color:#33}The involved tests are{color} > {code:java} >
[jira] [Commented] (HDFS-13560) Insufficient system resources exist to complete the requested service for some tests on Windows
[ https://issues.apache.org/jira/browse/HDFS-13560?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16475646#comment-16475646 ] genericqa commented on HDFS-13560: -- | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 34s{color} | {color:blue} Docker mode activated. {color} | || || || || {color:brown} Prechecks {color} || | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s{color} | {color:green} The patch appears to include 3 new or modified test files. {color} | || || || || {color:brown} trunk Compile Tests {color} || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 31m 17s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 1m 13s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 59s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 1m 21s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 13m 28s{color} | {color:green} branch has no errors when building and testing our client artifacts. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 2m 12s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 52s{color} | {color:green} trunk passed {color} | || || || || {color:brown} Patch Compile Tests {color} || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 1m 8s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 1m 9s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 1m 9s{color} | {color:green} the patch passed {color} | | {color:orange}-0{color} | {color:orange} checkstyle {color} | {color:orange} 0m 54s{color} | {color:orange} hadoop-hdfs-project/hadoop-hdfs: The patch generated 1 new + 76 unchanged - 2 fixed = 77 total (was 78) {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 1m 21s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 12m 46s{color} | {color:green} patch has no errors when building and testing our client artifacts. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 2m 20s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 52s{color} | {color:green} the patch passed {color} | || || || || {color:brown} Other Tests {color} || | {color:red}-1{color} | {color:red} unit {color} | {color:red}101m 58s{color} | {color:red} hadoop-hdfs in the patch failed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 26s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black}174m 51s{color} | {color:black} {color} | \\ \\ || Reason || Tests || | Failed junit tests | hadoop.hdfs.TestDFSStripedOutputStreamWithFailureWithRandomECPolicy | | | hadoop.hdfs.server.datanode.TestDataNodeVolumeFailureReporting | \\ \\ || Subsystem || Report/Notes || | Docker | Client=17.05.0-ce Server=17.05.0-ce Image:yetus/hadoop:abb62dd | | JIRA Issue | HDFS-13560 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12923427/HDFS-13560.000.patch | | Optional Tests | asflicense compile javac javadoc mvninstall mvnsite unit shadedclient findbugs checkstyle | | uname | Linux 37a87343e58b 3.13.0-143-generic #192-Ubuntu SMP Tue Feb 27 10:45:36 UTC 2018 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality | /testptch/patchprocess/precommit/personality/provided.sh | | git revision | trunk / 58b97c7 | | maven | version: Apache Maven 3.3.9 | | Default Java | 1.8.0_162 | | findbugs | v3.1.0-RC1 | | checkstyle | https://builds.apache.org/job/PreCommit-HDFS-Build/24210/artifact/out/diff-checkstyle-hadoop-hdfs-project_hadoop-hdfs.txt | | unit | https://builds.apache.org/job/PreCommit-HDFS-Build/24210/artifact/out/patch-unit-hadoop-hdfs-project_hadoop-hdfs.txt | | Test Results |
[jira] [Commented] (HDFS-13560) Insufficient system resources exist to complete the requested service for some tests on Windows
[ https://issues.apache.org/jira/browse/HDFS-13560?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16475432#comment-16475432 ] Anbang Hu commented on HDFS-13560: -- [^HDFS-13560.000.patch] increases timeout for testThrottling because it times out on Windows, and causes another test to fail: {color:#d04437}[ERROR] Tests run: 6, Failures: 0, Errors: 2, Skipped: 0, Time elapsed: 672.924 s <<< FAILURE! - in org.apache.hadoop.hdfs.server.datanode.TestDirectoryScanner{color} {color:#d04437}[ERROR] testThrottling(org.apache.hadoop.hdfs.server.datanode.TestDirectoryScanner) Time elapsed: 600.01 s <<< ERROR!{color} {color:#d04437}java.lang.Exception: test timed out after 60 milliseconds{color} {color:#d04437} at java.lang.Object.wait(Native Method){color} {color:#d04437} at java.lang.Object.wait(Object.java:502){color} {color:#d04437} at org.apache.hadoop.hdfs.DataStreamer.waitAndQueuePacket(DataStreamer.java:925){color} {color:#d04437} at org.apache.hadoop.hdfs.DFSOutputStream.enqueueCurrentPacket(DFSOutputStream.java:440){color} {color:#d04437} at org.apache.hadoop.hdfs.DFSOutputStream.enqueueCurrentPacketFull(DFSOutputStream.java:449){color} {color:#d04437} at org.apache.hadoop.hdfs.DFSOutputStream.writeChunk(DFSOutputStream.java:435){color} {color:#d04437} at org.apache.hadoop.fs.FSOutputSummer.writeChecksumChunks(FSOutputSummer.java:213){color} {color:#d04437} at org.apache.hadoop.fs.FSOutputSummer.write1(FSOutputSummer.java:125){color} {color:#d04437} at org.apache.hadoop.fs.FSOutputSummer.write(FSOutputSummer.java:111){color} {color:#d04437} at org.apache.hadoop.fs.FSDataOutputStream$PositionCache.write(FSDataOutputStream.java:58){color} {color:#d04437} at java.io.DataOutputStream.write(DataOutputStream.java:107){color} {color:#d04437} at org.apache.hadoop.hdfs.DFSTestUtil.createFile(DFSTestUtil.java:421){color} {color:#d04437} at org.apache.hadoop.hdfs.DFSTestUtil.createFile(DFSTestUtil.java:386){color} {color:#d04437} at org.apache.hadoop.hdfs.server.datanode.TestDirectoryScanner.createFile(TestDirectoryScanner.java:113){color} {color:#d04437} at org.apache.hadoop.hdfs.server.datanode.TestDirectoryScanner.testThrottling(TestDirectoryScanner.java:596){color} {color:#d04437} at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method){color} {color:#d04437} at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62){color} {color:#d04437} at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43){color} {color:#d04437} at java.lang.reflect.Method.invoke(Method.java:498){color} {color:#d04437} at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47){color} {color:#d04437} at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12){color} {color:#d04437} at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44){color} {color:#d04437} at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17){color} {color:#d04437} at org.junit.internal.runners.statements.FailOnTimeout$StatementThread.run(FailOnTimeout.java:74){color} {color:#d04437}[ERROR] testDirectoryScannerInFederatedCluster(org.apache.hadoop.hdfs.server.datanode.TestDirectoryScanner) Time elapsed: 1.033 s <<< ERROR!{color} {color:#d04437}java.io.IOException: Could not fully delete D:\OSS\hadoop-branch-2\hadoop-hdfs-project\hadoop-hdfs\target\test\data\dfs\name1{color} {color:#d04437} at org.apache.hadoop.hdfs.MiniDFSCluster.createNameNodesAndSetConf(MiniDFSCluster.java:1047){color} {color:#d04437} at org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:883){color} {color:#d04437} at org.apache.hadoop.hdfs.MiniDFSCluster.(MiniDFSCluster.java:514){color} {color:#d04437} at org.apache.hadoop.hdfs.MiniDFSCluster$Builder.build(MiniDFSCluster.java:473){color} {color:#d04437} at org.apache.hadoop.hdfs.server.datanode.TestDirectoryScanner.testDirectoryScannerInFederatedCluster(TestDirectoryScanner.java:1027){color} {color:#d04437} at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method){color} {color:#d04437} at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62){color} {color:#d04437} at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43){color} {color:#d04437} at java.lang.reflect.Method.invoke(Method.java:498){color} {color:#d04437} at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47){color} {color:#d04437} at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12){color} {color:#d04437} at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44){color} {color:#d04437} at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17){color} {color:#d04437} at