[ https://issues.apache.org/jira/browse/HADOOP-11957?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14988836#comment-14988836 ]
Hadoop QA commented on HADOOP-11957: ------------------------------------ | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 6s {color} | {color:blue} docker + precommit patch detected. {color} | | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s {color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s {color} | {color:green} The patch appears to include 1 new or modified test files. {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 2m 52s {color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 4m 16s {color} | {color:green} trunk passed with JDK v1.8.0_60 {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 4m 11s {color} | {color:green} trunk passed with JDK v1.7.0_79 {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 13s {color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 14s {color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 34s {color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 49s {color} | {color:green} trunk passed with JDK v1.8.0_60 {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 2s {color} | {color:green} trunk passed with JDK v1.7.0_79 {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 1m 29s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 4m 19s {color} | {color:green} the patch passed with JDK v1.8.0_60 {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 4m 19s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 4m 10s {color} | {color:green} the patch passed with JDK v1.7.0_79 {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 4m 10s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 14s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 14s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s {color} | {color:green} Patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 46s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 53s {color} | {color:green} the patch passed with JDK v1.8.0_60 {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 1s {color} | {color:green} the patch passed with JDK v1.7.0_79 {color} | | {color:red}-1{color} | {color:red} unit {color} | {color:red} 11m 51s {color} | {color:red} hadoop-common in the patch failed with JDK v1.8.0_60. {color} | | {color:red}-1{color} | {color:red} unit {color} | {color:red} 12m 17s {color} | {color:red} hadoop-common in the patch failed with JDK v1.7.0_79. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 22s {color} | {color:green} Patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black} 54m 51s {color} | {color:black} {color} | \\ \\ || Reason || Tests || | JDK v1.8.0_60 Failed junit tests | hadoop.ipc.TestDecayRpcScheduler | | | hadoop.net.unix.TestDomainSocket | | | hadoop.metrics2.sink.TestFileSink | | JDK v1.7.0_79 Failed junit tests | hadoop.net.unix.TestDomainSocket | | | hadoop.metrics2.sink.TestFileSink | \\ \\ || Subsystem || Report/Notes || | Docker | Client=1.7.1 Server=1.7.1 Image:test-patch-base-hadoop-date2015-11-04 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12732085/HADOOP-11957.001.patch | | JIRA Issue | HADOOP-11957 | | Optional Tests | asflicense javac javadoc mvninstall unit findbugs checkstyle compile | | uname | Linux 1ecb046fa009 3.13.0-36-lowlatency #63-Ubuntu SMP PREEMPT Wed Sep 3 21:56:12 UTC 2014 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality | /home/jenkins/jenkins-slave/workspace/PreCommit-HADOOP-Build/patchprocess/apache-yetus-d0f6847/precommit/personality/hadoop.sh | | git revision | trunk / 194251c | | Default Java | 1.7.0_79 | | Multi-JDK versions | /usr/lib/jvm/java-8-oracle:1.8.0_60 /usr/lib/jvm/java-7-openjdk-amd64:1.7.0_79 | | findbugs | v3.0.0 | | unit | https://builds.apache.org/job/PreCommit-HADOOP-Build/8022/artifact/patchprocess/patch-unit-hadoop-common-project_hadoop-common-jdk1.8.0_60.txt | | unit | https://builds.apache.org/job/PreCommit-HADOOP-Build/8022/artifact/patchprocess/patch-unit-hadoop-common-project_hadoop-common-jdk1.7.0_79.txt | | unit test logs | https://builds.apache.org/job/PreCommit-HADOOP-Build/8022/artifact/patchprocess/patch-unit-hadoop-common-project_hadoop-common-jdk1.8.0_60.txt https://builds.apache.org/job/PreCommit-HADOOP-Build/8022/artifact/patchprocess/patch-unit-hadoop-common-project_hadoop-common-jdk1.7.0_79.txt | | JDK v1.7.0_79 Test Results | https://builds.apache.org/job/PreCommit-HADOOP-Build/8022/testReport/ | | modules | C: hadoop-common-project/hadoop-common U: hadoop-common-project/hadoop-common | | Max memory used | 226MB | | Powered by | Apache Yetus http://yetus.apache.org | | Console output | https://builds.apache.org/job/PreCommit-HADOOP-Build/8022/console | This message was automatically generated. > if an IOException error is thrown in DomainSocket.close we go into infinite > loop. > --------------------------------------------------------------------------------- > > Key: HADOOP-11957 > URL: https://issues.apache.org/jira/browse/HADOOP-11957 > Project: Hadoop Common > Issue Type: Bug > Components: net > Affects Versions: 2.7.1 > Reporter: Anu Engineer > Assignee: Anu Engineer > Attachments: HADOOP-11957.001.patch > > > if an IOException error is thrown in DomainSocket.close we go into infinite > loop. > Issue : If the shutdown0(fd) call throws an IOException we break out of the > if shutdown call but will continue to loop in the while loop infinitely since > we have no way of decrementing the counter. Please scroll down and see the > comment marked with Bug Bug to see where the issue is. > {code:title=DomainSocket.java} > @Override > public void close() throws IOException { > // Set the closed bit on this DomainSocket > int count = 0; > try { > count = refCount.setClosed(); > } catch (ClosedChannelException e) { > // Someone else already closed the DomainSocket. > return; > } > // Wait for all references to go away > boolean didShutdown = false; > boolean interrupted = false; > while (count > 0) { > if (!didShutdown) { > try { > // Calling shutdown on the socket will interrupt blocking system > // calls like accept, write, and read that are going on in a > // different thread. > shutdown0(fd); > } catch (IOException e) { > LOG.error("shutdown error: ", e); > } > didShutdown = true; > // *BUG BUG* <-- Here the code will never exit the loop > // if the count is greater then 0. we need to break out > // of the while loop in case of IOException Error > } > try { > Thread.sleep(10); > } catch (InterruptedException e) { > interrupted = true; > } > count = refCount.getReferenceCount(); > } > // At this point, nobody has a reference to the file descriptor, > // and nobody will be able to get one in the future either. > // We now call close(2) on the file descriptor. > // After this point, the file descriptor number will be reused by > // something else. Although this DomainSocket object continues to hold > // the old file descriptor number (it's a final field), we never use it > // again because this DomainSocket is closed. > close0(fd); > if (interrupted) { > Thread.currentThread().interrupt(); > } > } > {code} -- This message was sent by Atlassian JIRA (v6.3.4#6332)