[jira] [Commented] (HADOOP-11746) rewrite test-patch.sh
[ https://issues.apache.org/jira/browse/HADOOP-11746?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14482649#comment-14482649 ] Allen Wittenauer commented on HADOOP-11746: --- The rat error is real, the rest are obviously bogus. Will fix and post an update tomorrow. > rewrite test-patch.sh > - > > Key: HADOOP-11746 > URL: https://issues.apache.org/jira/browse/HADOOP-11746 > Project: Hadoop Common > Issue Type: Test > Components: build, test >Affects Versions: 3.0.0 >Reporter: Allen Wittenauer >Assignee: Allen Wittenauer > Attachments: HADOOP-11746-00.patch, HADOOP-11746-01.patch, > HADOOP-11746-02.patch, HADOOP-11746-03.patch, HADOOP-11746-04.patch, > HADOOP-11746-05.patch, HADOOP-11746-06.patch, HADOOP-11746-07.patch, > HADOOP-11746-09.patch > > > This code is bad and you should feel bad. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HADOOP-11717) Add Redirecting WebSSO behavior with JWT Token in Hadoop Auth
[ https://issues.apache.org/jira/browse/HADOOP-11717?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14482617#comment-14482617 ] Kai Zheng commented on HADOOP-11717: Thanks for the update. bq.there is no reason to encrypt and decrypt here - the assumption is that they are to be protected with transport security. Sorry for my late response. I don't quite agree with this, as transport security is used to protect token from being leaked, and the encryption layer in token itself can be used to protect sensitive identity privacy, which means if someone loses his/her token, he/she won't suffer from exposing any sensitive credential identity attributes. I thought that's why JWE is defined and required. I understand in your case/requirement/scenario you may not use encryption feature, but it can be useful for others. I do think we can leave this aspect for future consideration, like tasks in HADOOP-11766. I missed to mention another point. Why we couple this with {{AltKerberosAuthenticationHandler}}? I thought a dedicated authentication handler like {{AuthTokenAuthenticationHandler}} may make more sense. Still, this may be handled separately if sounds good. Another point made by [~wheat9] here, which was also widely discussed before, it would be good to have a generic token abstract from JWT stuff. Again, I would follow up on this in HADOOP-11766 tasks. We still need to think about how to apply this new mechanism across all the web interfaces for HDFS and YARN. Will fire another task for this. Related to above points, to make the work more general, I suggest we change the following configuration items. authentication.provider.url => token.authentication.provider.url public.key.pem => token.signature.publickey; expected.jwt.audiences => expected.token.audiences; Maybe it's a little late to mention these. What I'm worrying about is once we have this in the trunk, we're then not able to enhance or follow up easily in the following. Any good idea for the concern? Thanks! > Add Redirecting WebSSO behavior with JWT Token in Hadoop Auth > - > > Key: HADOOP-11717 > URL: https://issues.apache.org/jira/browse/HADOOP-11717 > Project: Hadoop Common > Issue Type: Improvement > Components: security >Reporter: Larry McCay >Assignee: Larry McCay > Attachments: HADOOP-11717-1.patch, HADOOP-11717-2.patch, > HADOOP-11717-3.patch, HADOOP-11717-4.patch, HADOOP-11717-5.patch, > HADOOP-11717-6.patch, HADOOP-11717-7.patch, HADOOP-11717-8.patch, > RedirectingWebSSOwithJWTforHadoopWebUIs.pdf > > > Extend AltKerberosAuthenticationHandler to provide WebSSO flow for UIs. > The actual authentication is done by some external service that the handler > will redirect to when there is no hadoop.auth cookie and no JWT token found > in the incoming request. > Using JWT provides a number of benefits: > * It is not tied to any specific authentication mechanism - so buys us many > SSO integrations > * It is cryptographically verifiable for determining whether it can be trusted > * Checking for expiration allows for a limited lifetime and window for > compromised use > This will introduce the use of nimbus-jose-jwt library for processing, > validating and parsing JWT tokens. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HADOOP-11805) Better to rename some raw erasure coders
[ https://issues.apache.org/jira/browse/HADOOP-11805?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14482601#comment-14482601 ] Zhe Zhang commented on HADOOP-11805: Thanks Kai. It looks a clean refactor. +1 and thanks for the work! > Better to rename some raw erasure coders > > > Key: HADOOP-11805 > URL: https://issues.apache.org/jira/browse/HADOOP-11805 > Project: Hadoop Common > Issue Type: Sub-task > Components: io >Reporter: Kai Zheng >Assignee: Kai Zheng > Attachments: HADOOP-11805-v1.patch > > > When work on more coders, it was found better to rename some existing raw > coders for consistency, and more meaningful. As a result, we may have: > XORRawErasureCoder, in Java > NativeXORRawErasureCoder, in native > RSRawErasureCoder, in Java > NativeRSRawErasureCoder, in native and using ISA-L -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HADOOP-11801) Update BUILDING.txt
[ https://issues.apache.org/jira/browse/HADOOP-11801?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14482489#comment-14482489 ] Arpit Agarwal commented on HADOOP-11801: Okay can we just add _required_ after protobuf 2.5.0 to emphasize the version? I don't think we support 2.6.x and newer, in case Ubuntu upgrades the package version. {code} -* ProtocolBuffer 2.5.0 +* ProtocolBuffer 2.5.0 (required) $ sudo apt-get -y install libprotobuf-dev protobuf-compiler {code} Thanks. > Update BUILDING.txt > --- > > Key: HADOOP-11801 > URL: https://issues.apache.org/jira/browse/HADOOP-11801 > Project: Hadoop Common > Issue Type: Improvement > Components: documentation >Reporter: Gabor Liptak >Assignee: Gabor Liptak >Priority: Minor > Attachments: HADOOP-11801.patch > > > ProtocolBuffer is packaged in Ubuntu -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HADOOP-11801) Update BUILDING.txt
[ https://issues.apache.org/jira/browse/HADOOP-11801?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Arpit Agarwal updated HADOOP-11801: --- Assignee: Gabor Liptak > Update BUILDING.txt > --- > > Key: HADOOP-11801 > URL: https://issues.apache.org/jira/browse/HADOOP-11801 > Project: Hadoop Common > Issue Type: Improvement > Components: documentation >Reporter: Gabor Liptak >Assignee: Gabor Liptak >Priority: Minor > Attachments: HADOOP-11801.patch > > > ProtocolBuffer is packaged in Ubuntu -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HADOOP-11746) rewrite test-patch.sh
[ https://issues.apache.org/jira/browse/HADOOP-11746?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14482373#comment-14482373 ] Hadoop QA commented on HADOOP-11746: {color:red}-1 overall{color}. Here are the results of testing the latest attachment http://issues.apache.org/jira/secure/attachment/12723460/HADOOP-11746-09.patch against trunk revision 3fb5abf. {color:red}-1 @author{color}. The patch appears to contain 13 @author tags which the Hadoop community has agreed to not allow in code contributions. {color:green}+1 tests included{color}. The patch appears to include 4 new or modified test files. {color:red}-1 javac{color}. The applied patch generated 1148 javac compiler warnings (more than the trunk's current 208 warnings). {color:red}-1 javadoc{color}. The javadoc tool appears to have generated 43 warning messages. See https://builds.apache.org/job/PreCommit-HADOOP-Build/6069//artifact/patchprocess/diffJavadocWarnings.txt for details. {color:green}+1 eclipse:eclipse{color}. The patch built with eclipse:eclipse. {color:green}+1 findbugs{color}. The patch does not introduce any new Findbugs (version 2.0.3) warnings. {color:red}-1 release audit{color}. The applied patch generated 1 release audit warnings. {color:red}-1 core tests{color}. The patch failed these unit tests in hadoop-hdfs-project/hadoop-hdfs: org.apache.hadoop.hdfs.TestDFSOutputStream The following test timeouts occurred in hadoop-hdfs-project/hadoop-hdfs: org.apache.hadoop.hdfs.server.balancer.TestBalancerWithMultipleNameNodes Test results: https://builds.apache.org/job/PreCommit-HADOOP-Build/6069//testReport/ Release audit warnings: https://builds.apache.org/job/PreCommit-HADOOP-Build/6069//artifact/patchprocess/patchReleaseAuditProblems.txt Javac warnings: https://builds.apache.org/job/PreCommit-HADOOP-Build/6069//artifact/patchprocess/diffJavacWarnings.txt Console output: https://builds.apache.org/job/PreCommit-HADOOP-Build/6069//console This message is automatically generated. > rewrite test-patch.sh > - > > Key: HADOOP-11746 > URL: https://issues.apache.org/jira/browse/HADOOP-11746 > Project: Hadoop Common > Issue Type: Test > Components: build, test >Affects Versions: 3.0.0 >Reporter: Allen Wittenauer >Assignee: Allen Wittenauer > Attachments: HADOOP-11746-00.patch, HADOOP-11746-01.patch, > HADOOP-11746-02.patch, HADOOP-11746-03.patch, HADOOP-11746-04.patch, > HADOOP-11746-05.patch, HADOOP-11746-06.patch, HADOOP-11746-07.patch, > HADOOP-11746-09.patch > > > This code is bad and you should feel bad. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HADOOP-11801) Update BUILDING.txt
[ https://issues.apache.org/jira/browse/HADOOP-11801?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14482355#comment-14482355 ] Gabor Liptak commented on HADOOP-11801: --- Arpit, Ubuntu 14.04 comes with protobuf 2.5.0* http://packages.ubuntu.com/trusty/libprotobuf-dev and it likely stays 2.5.0* A specific version can be installed by running like: sudo apt-get -y install libprotobuf-dev=2.5.0-9ubuntu1 protobuf-compiler=2.5.0-9ubuntu1 but we will not get any patches this way ... > Update BUILDING.txt > --- > > Key: HADOOP-11801 > URL: https://issues.apache.org/jira/browse/HADOOP-11801 > Project: Hadoop Common > Issue Type: Improvement > Components: documentation >Reporter: Gabor Liptak >Priority: Minor > Attachments: HADOOP-11801.patch > > > ProtocolBuffer is packaged in Ubuntu -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Created] (HADOOP-11809) Building hadoop on windows 64 bit, windows 7.1 SDK : \hadoop-common\target\findbugsXml.xml does not exist
Umesh Kant created HADOOP-11809: --- Summary: Building hadoop on windows 64 bit, windows 7.1 SDK : \hadoop-common\target\findbugsXml.xml does not exist Key: HADOOP-11809 URL: https://issues.apache.org/jira/browse/HADOOP-11809 Project: Hadoop Common Issue Type: Bug Components: build Affects Versions: 2.6.0 Reporter: Umesh Kant I am trying to build hadoop 2.6.0 on Windows 7 64 bit, Windows 7.1 SDK. I have gone through Build.txt file and have did follow all the pre-requisites for build on windows. Still when I try to build, I am getting following error: Maven command: mvn package -X -Pdist -Pdocs -Psrc -Dtar -DskipTests -Pnative-win findbugs:findbugs [INFO] BUILD FAILURE [INFO] [INFO] Total time: 04:35 min [INFO] Finished at: 2015-04-03T23:16:57-04:00 [INFO] Final Memory: 123M/1435M [INFO] [ERROR] Failed to execute goal org.apache.maven.plugins:maven-antrun-plugin:1.7: run (site) on project hadoop-common: An Ant BuildException has occured: input fi le C:\H\hadoop-2.6.0-src\hadoop-common-project\hadoop-common\target\findbugsXml. xml does not exist [ERROR] around Ant part .. @ 44:232 in C:\H\hadoop-2.6.0-src\hadoop-common-project\hado op-common\target\antrun\build-main.xml [ERROR] -> [Help 1] org.apache.maven.lifecycle.LifecycleExecutionException: Failed to execute goal o rg.apache.maven.plugins:maven-antrun-plugin:1.7:run (site) on project hadoop-com mon: An Ant BuildException has occured: input file C:\H\hadoop-2.6.0-src\hadoop- common-project\hadoop-common\target\findbugsXml.xml does not exist around Ant part .. @ 44:232 in C:\H\hadoop-2.6.0-src\hadoop-common-project\hadoop-commo n\target\antrun\build-main.xml at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor .java:216) at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor .java:153) at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor .java:145) at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProje ct(LifecycleModuleBuilder.java:116) at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProje ct(LifecycleModuleBuilder.java:80) at org.apache.maven.lifecycle.internal.builder.singlethreaded.SingleThre adedBuilder.build(SingleThreadedBuilder.java:51) at org.apache.maven.lifecycle.internal.LifecycleStarter.execute(Lifecycl eStarter.java:128) at org.apache.maven.DefaultMaven.doExecute(DefaultMaven.java:307) at org.apache.maven.DefaultMaven.doExecute(DefaultMaven.java:193) at org.apache.maven.DefaultMaven.execute(DefaultMaven.java:106) at org.apache.maven.cli.MavenCli.execute(MavenCli.java:862) at org.apache.maven.cli.MavenCli.doMain(MavenCli.java:286) at org.apache.maven.cli.MavenCli.main(MavenCli.java:197) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl. java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAcces sorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:483) at org.codehaus.plexus.classworlds.launcher.Launcher.launchEnhanced(Laun cher.java:289) at org.codehaus.plexus.classworlds.launcher.Launcher.launch(Launcher.jav a:229) at org.codehaus.plexus.classworlds.launcher.Launcher.mainWithExitCode(La uncher.java:415) at org.codehaus.plexus.classworlds.launcher.Launcher.main(Launcher.java: 356) Caused by: org.apache.maven.plugin.MojoExecutionException: An Ant BuildException has occured: input file C:\H\hadoop-2.6.0-src\hadoop-common-project\hadoop-comm on\target\findbugsXml.xml does not exist around Ant part .. @ 44:232 in C:\H\hadoop-2.6.0-src\hadoop-common-project\hadoop-commo n\target\antrun\build-main.xml at org.apache.maven.plugin.antrun.AntRunMojo.execute(AntRunMojo.java:355 ) at org.apache.maven.plugin.DefaultBuildPluginManager.executeMojo(Default BuildPluginManager.java:134) at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor .java:208) ... 20 more Caused by: C:\H\hadoop-2.6.0-src\hadoop-common-project\hadoop-common\target\antr un\build-main.xml:44: input file C:\H\hadoop-2.6.0-src\hadoop-common-project\had oop-common\target\findbugsXml.xml does not exist at org.apache.tools.ant.taskdefs.XSLTProcess.handleError(XSLTProcess.jav a:1243) at org.apache.tools.ant.taskdefs.XSLTProcess.execute(XSLTProcess.java:33 5) at org.apache.tools.ant.UnknownElement.execute(UnknownElement.java:291) at sun.reflect.GeneratedMethodAccessor39.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(
[jira] [Commented] (HADOOP-11805) Better to rename some raw erasure coders
[ https://issues.apache.org/jira/browse/HADOOP-11805?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14482209#comment-14482209 ] Kai Zheng commented on HADOOP-11805: Yes, you're right. For native ones, I would come up in this naming style {{NativeXORRawErasureCoder}} and {{NativeRSRawErasureCoder}} (using ISA-L) later. Without _Native_ prefix, it would be meant to be in implemented Java. > Better to rename some raw erasure coders > > > Key: HADOOP-11805 > URL: https://issues.apache.org/jira/browse/HADOOP-11805 > Project: Hadoop Common > Issue Type: Sub-task > Components: io >Reporter: Kai Zheng >Assignee: Kai Zheng > Attachments: HADOOP-11805-v1.patch > > > When work on more coders, it was found better to rename some existing raw > coders for consistency, and more meaningful. As a result, we may have: > XORRawErasureCoder, in Java > NativeXORRawErasureCoder, in native > RSRawErasureCoder, in Java > NativeRSRawErasureCoder, in native and using ISA-L -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HADOOP-11805) Better to rename some raw erasure coders
[ https://issues.apache.org/jira/browse/HADOOP-11805?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14482173#comment-14482173 ] Zhe Zhang commented on HADOOP-11805: Thanks for the work Kai. So the main updates are: # Xor... -> XOR... #JRS... -> RS... Is that right? If so the main direction sounds good to me. > Better to rename some raw erasure coders > > > Key: HADOOP-11805 > URL: https://issues.apache.org/jira/browse/HADOOP-11805 > Project: Hadoop Common > Issue Type: Sub-task > Components: io >Reporter: Kai Zheng >Assignee: Kai Zheng > Attachments: HADOOP-11805-v1.patch > > > When work on more coders, it was found better to rename some existing raw > coders for consistency, and more meaningful. As a result, we may have: > XORRawErasureCoder, in Java > NativeXORRawErasureCoder, in native > RSRawErasureCoder, in Java > NativeRSRawErasureCoder, in native and using ISA-L -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HADOOP-11746) rewrite test-patch.sh
[ https://issues.apache.org/jira/browse/HADOOP-11746?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14482083#comment-14482083 ] Hadoop QA commented on HADOOP-11746: {color:red}-1 overall{color}. Here are the results of testing the latest attachment http://issues.apache.org/jira/secure/attachment/12723454/HADOOP-11746-08.patch against trunk revision 3fb5abf. {color:red}-1 @author{color}. The patch appears to contain 13 @author tags which the Hadoop community has agreed to not allow in code contributions. {color:green}+1 tests included{color}. The patch appears to include 4 new or modified test files. {color:green}+1 javac{color}. The applied patch does not increase the total number of javac compiler warnings. {color:green}+1 javadoc{color}. There were no new javadoc warning messages. {color:green}+1 eclipse:eclipse{color}. The patch built with eclipse:eclipse. {color:green}+1 findbugs{color}. The patch does not introduce any new Findbugs (version 2.0.3) warnings. {color:red}-1 release audit{color}. The applied patch generated 1 release audit warnings. {color:green}+1 core tests{color}. The patch passed unit tests in . Test results: https://builds.apache.org/job/PreCommit-HADOOP-Build/6068//testReport/ Release audit warnings: https://builds.apache.org/job/PreCommit-HADOOP-Build/6068//artifact/patchprocess/patchReleaseAuditProblems.txt Console output: https://builds.apache.org/job/PreCommit-HADOOP-Build/6068//console This message is automatically generated. > rewrite test-patch.sh > - > > Key: HADOOP-11746 > URL: https://issues.apache.org/jira/browse/HADOOP-11746 > Project: Hadoop Common > Issue Type: Test > Components: build, test >Affects Versions: 3.0.0 >Reporter: Allen Wittenauer >Assignee: Allen Wittenauer > Attachments: HADOOP-11746-00.patch, HADOOP-11746-01.patch, > HADOOP-11746-02.patch, HADOOP-11746-03.patch, HADOOP-11746-04.patch, > HADOOP-11746-05.patch, HADOOP-11746-06.patch, HADOOP-11746-07.patch, > HADOOP-11746-09.patch > > > This code is bad and you should feel bad. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HADOOP-11746) rewrite test-patch.sh
[ https://issues.apache.org/jira/browse/HADOOP-11746?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Allen Wittenauer updated HADOOP-11746: -- Attachment: (was: HADOOP-11746-08.patch) > rewrite test-patch.sh > - > > Key: HADOOP-11746 > URL: https://issues.apache.org/jira/browse/HADOOP-11746 > Project: Hadoop Common > Issue Type: Test > Components: build, test >Affects Versions: 3.0.0 >Reporter: Allen Wittenauer >Assignee: Allen Wittenauer > Attachments: HADOOP-11746-00.patch, HADOOP-11746-01.patch, > HADOOP-11746-02.patch, HADOOP-11746-03.patch, HADOOP-11746-04.patch, > HADOOP-11746-05.patch, HADOOP-11746-06.patch, HADOOP-11746-07.patch, > HADOOP-11746-09.patch > > > This code is bad and you should feel bad. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HADOOP-11746) rewrite test-patch.sh
[ https://issues.apache.org/jira/browse/HADOOP-11746?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Allen Wittenauer updated HADOOP-11746: -- Attachment: HADOOP-11746-09.patch -09: * -08 was the wrong file. woops. > rewrite test-patch.sh > - > > Key: HADOOP-11746 > URL: https://issues.apache.org/jira/browse/HADOOP-11746 > Project: Hadoop Common > Issue Type: Test > Components: build, test >Affects Versions: 3.0.0 >Reporter: Allen Wittenauer >Assignee: Allen Wittenauer > Attachments: HADOOP-11746-00.patch, HADOOP-11746-01.patch, > HADOOP-11746-02.patch, HADOOP-11746-03.patch, HADOOP-11746-04.patch, > HADOOP-11746-05.patch, HADOOP-11746-06.patch, HADOOP-11746-07.patch, > HADOOP-11746-08.patch, HADOOP-11746-09.patch > > > This code is bad and you should feel bad. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HADOOP-11746) rewrite test-patch.sh
[ https://issues.apache.org/jira/browse/HADOOP-11746?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Allen Wittenauer updated HADOOP-11746: -- Status: Patch Available (was: Open) > rewrite test-patch.sh > - > > Key: HADOOP-11746 > URL: https://issues.apache.org/jira/browse/HADOOP-11746 > Project: Hadoop Common > Issue Type: Test > Components: build, test >Affects Versions: 3.0.0 >Reporter: Allen Wittenauer >Assignee: Allen Wittenauer > Attachments: HADOOP-11746-00.patch, HADOOP-11746-01.patch, > HADOOP-11746-02.patch, HADOOP-11746-03.patch, HADOOP-11746-04.patch, > HADOOP-11746-05.patch, HADOOP-11746-06.patch, HADOOP-11746-07.patch, > HADOOP-11746-08.patch > > > This code is bad and you should feel bad. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HADOOP-11746) rewrite test-patch.sh
[ https://issues.apache.org/jira/browse/HADOOP-11746?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Allen Wittenauer updated HADOOP-11746: -- Status: Open (was: Patch Available) > rewrite test-patch.sh > - > > Key: HADOOP-11746 > URL: https://issues.apache.org/jira/browse/HADOOP-11746 > Project: Hadoop Common > Issue Type: Test > Components: build, test >Affects Versions: 3.0.0 >Reporter: Allen Wittenauer >Assignee: Allen Wittenauer > Attachments: HADOOP-11746-00.patch, HADOOP-11746-01.patch, > HADOOP-11746-02.patch, HADOOP-11746-03.patch, HADOOP-11746-04.patch, > HADOOP-11746-05.patch, HADOOP-11746-06.patch, HADOOP-11746-07.patch, > HADOOP-11746-08.patch > > > This code is bad and you should feel bad. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HADOOP-11746) rewrite test-patch.sh
[ https://issues.apache.org/jira/browse/HADOOP-11746?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Allen Wittenauer updated HADOOP-11746: -- Attachment: HADOOP-11746-08.patch -08: * borrow some code from the folks at HBase to help make the checkstyle check more useful * add an EOL whitespace check bz i'm tired of checking for it myself. :D * fix some extraneous output when no tests have failed > rewrite test-patch.sh > - > > Key: HADOOP-11746 > URL: https://issues.apache.org/jira/browse/HADOOP-11746 > Project: Hadoop Common > Issue Type: Test > Components: build, test >Affects Versions: 3.0.0 >Reporter: Allen Wittenauer >Assignee: Allen Wittenauer > Attachments: HADOOP-11746-00.patch, HADOOP-11746-01.patch, > HADOOP-11746-02.patch, HADOOP-11746-03.patch, HADOOP-11746-04.patch, > HADOOP-11746-05.patch, HADOOP-11746-06.patch, HADOOP-11746-07.patch, > HADOOP-11746-08.patch > > > This code is bad and you should feel bad. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HADOOP-11252) RPC client write does not time out by default
[ https://issues.apache.org/jira/browse/HADOOP-11252?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14482010#comment-14482010 ] Ray Chiang commented on HADOOP-11252: - +1 (non-binding). The code changes look fine to me. As an easy test of its effect, I'm able to cause problems on my cluster by setting the value to 1ms. Reasonable values run fine for me. I'd like to see the new property properly documented. From what I can see, core-default.xml looks like the right place. > RPC client write does not time out by default > - > > Key: HADOOP-11252 > URL: https://issues.apache.org/jira/browse/HADOOP-11252 > Project: Hadoop Common > Issue Type: Bug > Components: ipc >Affects Versions: 2.5.0 >Reporter: Wilfred Spiegelenburg >Assignee: Wilfred Spiegelenburg >Priority: Critical > Attachments: HADOOP-11252.patch > > > The RPC client has a default timeout set to 0 when no timeout is passed in. > This means that the network connection created will not timeout when used to > write data. The issue has shown in YARN-2578 and HDFS-4858. Timeouts for > writes then fall back to the tcp level retry (configured via tcp_retries2) > and timeouts between the 15-30 minutes. Which is too long for a default > behaviour. > Using 0 as the default value for timeout is incorrect. We should use a sane > value for the timeout and the "ipc.ping.interval" configuration value is a > logical choice for it. The default behaviour should be changed from 0 to the > value read for the ping interval from the Configuration. > Fixing it in common makes more sense than finding and changing all other > points in the code that do not pass in a timeout. > Offending code lines: > https://github.com/apache/hadoop/blob/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ipc/RPC.java#L488 > and > https://github.com/apache/hadoop/blob/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ipc/RPC.java#L350 -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HADOOP-11090) [Umbrella] Support Java 8 in Hadoop
[ https://issues.apache.org/jira/browse/HADOOP-11090?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14481998#comment-14481998 ] Rick Kellogg commented on HADOOP-11090: --- Suggest we look into using Animal Sniffer (http://mojo.codehaus.org/animal-sniffer/) to ensure API compatibility between releases. The folks at Spring Source use it to help with Java 6-8 compatibility as well. See: http://spring.io/blog/2015/04/03/how-spring-achieves-compatibility-with-java-6-7-and-8 > [Umbrella] Support Java 8 in Hadoop > --- > > Key: HADOOP-11090 > URL: https://issues.apache.org/jira/browse/HADOOP-11090 > Project: Hadoop Common > Issue Type: New Feature >Reporter: Mohammad Kamrul Islam >Assignee: Mohammad Kamrul Islam > > Java 8 is coming quickly to various clusters. Making sure Hadoop seamlessly > works with Java 8 is important for the Apache community. > > This JIRA is to track the issues/experiences encountered during Java 8 > migration. If you find a potential bug , please create a separate JIRA either > as a sub-task or linked into this JIRA. > If you find a Hadoop or JVM configuration tuning, you can create a JIRA as > well. Or you can add a comment here. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HADOOP-11801) Update BUILDING.txt
[ https://issues.apache.org/jira/browse/HADOOP-11801?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14481869#comment-14481869 ] Arpit Agarwal commented on HADOOP-11801: Thanks for contributing the doc update [~gliptak]. Is there a package specific to protobuf 2.5.0? > Update BUILDING.txt > --- > > Key: HADOOP-11801 > URL: https://issues.apache.org/jira/browse/HADOOP-11801 > Project: Hadoop Common > Issue Type: Improvement > Components: documentation >Reporter: Gabor Liptak >Priority: Minor > Attachments: HADOOP-11801.patch > > > ProtocolBuffer is packaged in Ubuntu -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HADOOP-11776) jdiff is broken in Hadoop 2
[ https://issues.apache.org/jira/browse/HADOOP-11776?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14481438#comment-14481438 ] Li Lu commented on HADOOP-11776: Thanks [~vinodkv] for the review and commit. Yes we do have more things to do for tools to check API compatibility. Given the current status our long term goal may be replacing it with some other tool. > jdiff is broken in Hadoop 2 > --- > > Key: HADOOP-11776 > URL: https://issues.apache.org/jira/browse/HADOOP-11776 > Project: Hadoop Common > Issue Type: Bug >Affects Versions: 2.6.0 >Reporter: Li Lu >Assignee: Li Lu >Priority: Blocker > Fix For: 2.7.0 > > Attachments: HADOOP-11776-040115.patch > > > Seems like we haven't touch the API files from jdiff under dev-support for a > while. For now we're missing the jdiff API files for hadoop 2. We're also > missing YARN when generating the jdiff API files. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HADOOP-11731) Rework the changelog and releasenotes
[ https://issues.apache.org/jira/browse/HADOOP-11731?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14481354#comment-14481354 ] Colin Patrick McCabe commented on HADOOP-11731: --- bq. Steve said: I don't think CHANGES.TXT works that well. We may think it does, but that's because without the tooling to validate it, stuff doesn't get added and so it can omit a lot of work. Then there's the problem of merging across branches, and dealing with race conditions/commit conflict between other people's work and yours. Absolutely. CHANGES.txt does not work that well for many reasons. It's often incorrect, as Allen pointed out, since it's entirely manually created. It creates many spurious merge conflicts. bq. Tsz Wo Nicholas Sze wrote: Again, I do not oppose using the new tool. However, we do need a transition period to see if it indeed works well. I agree. However, given that we are going to be maintaining branch-2 for at least another 2 years (we don't even have a 3.x release roadmap yet), it seems like that transition period should be in the 2.8 timeframe. It would also be nice to move to this system for point releases. Otherwise, we end up doing CHANGES.txt anyway for backports. > Rework the changelog and releasenotes > - > > Key: HADOOP-11731 > URL: https://issues.apache.org/jira/browse/HADOOP-11731 > Project: Hadoop Common > Issue Type: New Feature > Components: documentation >Affects Versions: 3.0.0 >Reporter: Allen Wittenauer >Assignee: Allen Wittenauer > Fix For: 3.0.0 > > Attachments: HADOOP-11731-00.patch, HADOOP-11731-01.patch, > HADOOP-11731-03.patch, HADOOP-11731-04.patch, HADOOP-11731-05.patch, > HADOOP-11731-06.patch, HADOOP-11731-07.patch > > > The current way we generate these build artifacts is awful. Plus they are > ugly and, in the case of release notes, very hard to pick out what is > important. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HADOOP-11758) Add options to filter out too much granular tracing spans
[ https://issues.apache.org/jira/browse/HADOOP-11758?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14481317#comment-14481317 ] Colin Patrick McCabe commented on HADOOP-11758: --- Thanks for attaching this. I've been meaning to put up some more real-world data from my hdfs cluster myself. It looks like we still have a problem with overly long span names in a few places... e.g., {{org.apache.hadoop.hdfs.protocol.ClientProtocol.complete}} should really be {{ClientProtocol#complete}}. Let me see if I can find where it's doing this... > Add options to filter out too much granular tracing spans > - > > Key: HADOOP-11758 > URL: https://issues.apache.org/jira/browse/HADOOP-11758 > Project: Hadoop Common > Issue Type: Improvement > Components: tracing >Reporter: Masatake Iwasaki >Assignee: Masatake Iwasaki >Priority: Minor > Attachments: testWriteTraceHooks-HDFS-8026.html, > testWriteTraceHooks.html > > > in order to avoid queue in span receiver spills -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HADOOP-10366) [Doc] wrap value of io.serializations in core-default.xml to fit better in browser
[ https://issues.apache.org/jira/browse/HADOOP-10366?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14481315#comment-14481315 ] Hadoop QA commented on HADOOP-10366: {color:green}+1 overall{color}. Here are the results of testing the latest attachment http://issues.apache.org/jira/secure/attachment/12723357/HADOOP-10366-wrap01.patch against trunk revision 53959e6. {color:green}+1 @author{color}. The patch does not contain any @author tags. {color:green}+0 tests included{color}. The patch appears to be a documentation patch that doesn't require tests. {color:green}+1 javac{color}. The applied patch does not increase the total number of javac compiler warnings. {color:green}+1 javadoc{color}. There were no new javadoc warning messages. {color:green}+1 eclipse:eclipse{color}. The patch built with eclipse:eclipse. {color:green}+1 findbugs{color}. The patch does not introduce any new Findbugs (version 2.0.3) warnings. {color:green}+1 release audit{color}. The applied patch does not increase the total number of release audit warnings. {color:green}+1 core tests{color}. The patch passed unit tests in hadoop-common-project/hadoop-common. Test results: https://builds.apache.org/job/PreCommit-HADOOP-Build/6067//testReport/ Console output: https://builds.apache.org/job/PreCommit-HADOOP-Build/6067//console This message is automatically generated. > [Doc] wrap value of io.serializations in core-default.xml to fit better in > browser > -- > > Key: HADOOP-10366 > URL: https://issues.apache.org/jira/browse/HADOOP-10366 > Project: Hadoop Common > Issue Type: Improvement > Components: documentation >Affects Versions: 3.0.0 >Reporter: Chengwei Yang >Assignee: kanaka kumar avvaru >Priority: Trivial > Labels: documentation, newbie, patch > Attachments: HADOOP-10366-wrap01, HADOOP-10366-wrap01.patch, > HADOOP-10366.patch > > > The io.serialization property in core-default.xml has a very long value in a > single line, as below > {code} > > io.serializations > > org.apache.hadoop.io.serializer.WritableSerialization,org.apache.hadoop.io.serializer.avro.AvroSpecificSerialization,org.apache.hadoop.io.serializer.avro.AvroReflectSerialization > A list of serialization classes that can be used for > obtaining serializers and deserializers. > > {code} > which not only break the code style (a very long line) but also not fit well > in browser. Due to this single very long line, the "description" column can > not show in browser by default -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HADOOP-10366) [Doc] wrap value of io.serializations in core-default.xml to fit better in browser
[ https://issues.apache.org/jira/browse/HADOOP-10366?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] kanaka kumar avvaru updated HADOOP-10366: - Status: Patch Available (was: Open) > [Doc] wrap value of io.serializations in core-default.xml to fit better in > browser > -- > > Key: HADOOP-10366 > URL: https://issues.apache.org/jira/browse/HADOOP-10366 > Project: Hadoop Common > Issue Type: Improvement > Components: documentation >Affects Versions: 3.0.0 >Reporter: Chengwei Yang >Assignee: kanaka kumar avvaru >Priority: Trivial > Labels: documentation, newbie, patch > Attachments: HADOOP-10366-wrap01, HADOOP-10366-wrap01.patch, > HADOOP-10366.patch > > > The io.serialization property in core-default.xml has a very long value in a > single line, as below > {code} > > io.serializations > > org.apache.hadoop.io.serializer.WritableSerialization,org.apache.hadoop.io.serializer.avro.AvroSpecificSerialization,org.apache.hadoop.io.serializer.avro.AvroReflectSerialization > A list of serialization classes that can be used for > obtaining serializers and deserializers. > > {code} > which not only break the code style (a very long line) but also not fit well > in browser. Due to this single very long line, the "description" column can > not show in browser by default -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HADOOP-10366) [Doc] wrap value of io.serializations in core-default.xml to fit better in browser
[ https://issues.apache.org/jira/browse/HADOOP-10366?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14481263#comment-14481263 ] kanaka kumar avvaru commented on HADOOP-10366: -- Attached a patch HADOOP-10366-wrap01.patch as per the suggestions in the earlier comments to wrap the value with line size of 80. > [Doc] wrap value of io.serializations in core-default.xml to fit better in > browser > -- > > Key: HADOOP-10366 > URL: https://issues.apache.org/jira/browse/HADOOP-10366 > Project: Hadoop Common > Issue Type: Improvement > Components: documentation >Affects Versions: 3.0.0 >Reporter: Chengwei Yang >Assignee: kanaka kumar avvaru >Priority: Trivial > Labels: documentation, newbie, patch > Attachments: HADOOP-10366-wrap01, HADOOP-10366-wrap01.patch, > HADOOP-10366.patch > > > The io.serialization property in core-default.xml has a very long value in a > single line, as below > {code} > > io.serializations > > org.apache.hadoop.io.serializer.WritableSerialization,org.apache.hadoop.io.serializer.avro.AvroSpecificSerialization,org.apache.hadoop.io.serializer.avro.AvroReflectSerialization > A list of serialization classes that can be used for > obtaining serializers and deserializers. > > {code} > which not only break the code style (a very long line) but also not fit well > in browser. Due to this single very long line, the "description" column can > not show in browser by default -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HADOOP-10366) [Doc] wrap value of io.serializations in core-default.xml to fit better in browser
[ https://issues.apache.org/jira/browse/HADOOP-10366?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] kanaka kumar avvaru updated HADOOP-10366: - Attachment: HADOOP-10366-wrap01.patch > [Doc] wrap value of io.serializations in core-default.xml to fit better in > browser > -- > > Key: HADOOP-10366 > URL: https://issues.apache.org/jira/browse/HADOOP-10366 > Project: Hadoop Common > Issue Type: Improvement > Components: documentation >Affects Versions: 3.0.0 >Reporter: Chengwei Yang >Assignee: kanaka kumar avvaru >Priority: Trivial > Labels: documentation, newbie, patch > Attachments: HADOOP-10366-wrap01, HADOOP-10366-wrap01.patch, > HADOOP-10366.patch > > > The io.serialization property in core-default.xml has a very long value in a > single line, as below > {code} > > io.serializations > > org.apache.hadoop.io.serializer.WritableSerialization,org.apache.hadoop.io.serializer.avro.AvroSpecificSerialization,org.apache.hadoop.io.serializer.avro.AvroReflectSerialization > A list of serialization classes that can be used for > obtaining serializers and deserializers. > > {code} > which not only break the code style (a very long line) but also not fit well > in browser. Due to this single very long line, the "description" column can > not show in browser by default -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HADOOP-10366) [Doc] wrap value of io.serializations in core-default.xml to fit better in browser
[ https://issues.apache.org/jira/browse/HADOOP-10366?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] kanaka kumar avvaru updated HADOOP-10366: - Attachment: HADOOP-10366-wrap01 > [Doc] wrap value of io.serializations in core-default.xml to fit better in > browser > -- > > Key: HADOOP-10366 > URL: https://issues.apache.org/jira/browse/HADOOP-10366 > Project: Hadoop Common > Issue Type: Improvement > Components: documentation >Affects Versions: 3.0.0 >Reporter: Chengwei Yang >Assignee: kanaka kumar avvaru >Priority: Trivial > Labels: documentation, newbie, patch > Attachments: HADOOP-10366-wrap01, HADOOP-10366.patch > > > The io.serialization property in core-default.xml has a very long value in a > single line, as below > {code} > > io.serializations > > org.apache.hadoop.io.serializer.WritableSerialization,org.apache.hadoop.io.serializer.avro.AvroSpecificSerialization,org.apache.hadoop.io.serializer.avro.AvroReflectSerialization > A list of serialization classes that can be used for > obtaining serializers and deserializers. > > {code} > which not only break the code style (a very long line) but also not fit well > in browser. Due to this single very long line, the "description" column can > not show in browser by default -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Assigned] (HADOOP-10366) [Doc] wrap value of io.serializations in core-default.xml to fit better in browser
[ https://issues.apache.org/jira/browse/HADOOP-10366?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] kanaka kumar avvaru reassigned HADOOP-10366: Assignee: kanaka kumar avvaru > [Doc] wrap value of io.serializations in core-default.xml to fit better in > browser > -- > > Key: HADOOP-10366 > URL: https://issues.apache.org/jira/browse/HADOOP-10366 > Project: Hadoop Common > Issue Type: Improvement > Components: documentation >Affects Versions: 3.0.0 >Reporter: Chengwei Yang >Assignee: kanaka kumar avvaru >Priority: Trivial > Labels: documentation, newbie, patch > Attachments: HADOOP-10366.patch > > > The io.serialization property in core-default.xml has a very long value in a > single line, as below > {code} > > io.serializations > > org.apache.hadoop.io.serializer.WritableSerialization,org.apache.hadoop.io.serializer.avro.AvroSpecificSerialization,org.apache.hadoop.io.serializer.avro.AvroReflectSerialization > A list of serialization classes that can be used for > obtaining serializers and deserializers. > > {code} > which not only break the code style (a very long line) but also not fit well > in browser. Due to this single very long line, the "description" column can > not show in browser by default -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HADOOP-10366) [Doc] wrap value of io.serializations in core-default.xml to fit better in browser
[ https://issues.apache.org/jira/browse/HADOOP-10366?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14481250#comment-14481250 ] kanaka kumar avvaru commented on HADOOP-10366: -- As the issue is stale from about a year, I would like to work on it and give a patch. [~chengwei-yang] if you would like to continue work on it, please feel free to assign it back to you. > [Doc] wrap value of io.serializations in core-default.xml to fit better in > browser > -- > > Key: HADOOP-10366 > URL: https://issues.apache.org/jira/browse/HADOOP-10366 > Project: Hadoop Common > Issue Type: Improvement > Components: documentation >Affects Versions: 3.0.0 >Reporter: Chengwei Yang >Priority: Trivial > Labels: documentation, newbie, patch > Attachments: HADOOP-10366.patch > > > The io.serialization property in core-default.xml has a very long value in a > single line, as below > {code} > > io.serializations > > org.apache.hadoop.io.serializer.WritableSerialization,org.apache.hadoop.io.serializer.avro.AvroSpecificSerialization,org.apache.hadoop.io.serializer.avro.AvroReflectSerialization > A list of serialization classes that can be used for > obtaining serializers and deserializers. > > {code} > which not only break the code style (a very long line) but also not fit well > in browser. Due to this single very long line, the "description" column can > not show in browser by default -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HADOOP-11772) RPC Invoker relies on static ClientCache which has synchronized(this) blocks
[ https://issues.apache.org/jira/browse/HADOOP-11772?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14481126#comment-14481126 ] Hadoop QA commented on HADOOP-11772: {color:green}+1 overall{color}. Here are the results of testing the latest attachment http://issues.apache.org/jira/secure/attachment/12723325/HADOOP-11772-wip-002.patch against trunk revision 96d7211. {color:green}+1 @author{color}. The patch does not contain any @author tags. {color:green}+1 tests included{color}. The patch appears to include 1 new or modified test files. {color:green}+1 javac{color}. The applied patch does not increase the total number of javac compiler warnings. {color:green}+1 javadoc{color}. There were no new javadoc warning messages. {color:green}+1 eclipse:eclipse{color}. The patch built with eclipse:eclipse. {color:green}+1 findbugs{color}. The patch does not introduce any new Findbugs (version 2.0.3) warnings. {color:green}+1 release audit{color}. The applied patch does not increase the total number of release audit warnings. {color:green}+1 core tests{color}. The patch passed unit tests in hadoop-common-project/hadoop-common. Test results: https://builds.apache.org/job/PreCommit-HADOOP-Build/6066//testReport/ Console output: https://builds.apache.org/job/PreCommit-HADOOP-Build/6066//console This message is automatically generated. > RPC Invoker relies on static ClientCache which has synchronized(this) blocks > > > Key: HADOOP-11772 > URL: https://issues.apache.org/jira/browse/HADOOP-11772 > Project: Hadoop Common > Issue Type: Sub-task > Components: ipc, performance >Reporter: Gopal V >Assignee: Akira AJISAKA > Attachments: HADOOP-11772-001.patch, HADOOP-11772-wip-001.patch, > HADOOP-11772-wip-002.patch, dfs-sync-ipc.png, sync-client-bt.png, > sync-client-threads.png > > > {code} > private static ClientCache CLIENTS=new ClientCache(); > ... > this.client = CLIENTS.getClient(conf, factory); > {code} > Meanwhile in ClientCache > {code} > public synchronized Client getClient(Configuration conf, > SocketFactory factory, Class valueClass) { > ... >Client client = clients.get(factory); > if (client == null) { > client = new Client(valueClass, conf, factory); > clients.put(factory, client); > } else { > client.incCount(); > } > {code} > All invokers end up calling these methods, resulting in IPC clients choking > up. > !sync-client-threads.png! > !sync-client-bt.png! > !dfs-sync-ipc.png! -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HADOOP-11665) Provide and unify cross platform byteorder support in native code
[ https://issues.apache.org/jira/browse/HADOOP-11665?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14481093#comment-14481093 ] Ayappan commented on HADOOP-11665: -- Any update here? This issue has been lingering around for a long time. > Provide and unify cross platform byteorder support in native code > - > > Key: HADOOP-11665 > URL: https://issues.apache.org/jira/browse/HADOOP-11665 > Project: Hadoop Common > Issue Type: Bug > Components: native, util >Affects Versions: 2.4.1, 2.6.0 > Environment: PowerPC Big Endian & other Big Endian platforms >Reporter: Binglin Chang >Assignee: Binglin Chang > Attachments: HADOOP-11665.001.patch > > -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HADOOP-11772) RPC Invoker relies on static ClientCache which has synchronized(this) blocks
[ https://issues.apache.org/jira/browse/HADOOP-11772?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Akira AJISAKA updated HADOOP-11772: --- Attachment: HADOOP-11772-wip-002.patch Fixed the test failure. > RPC Invoker relies on static ClientCache which has synchronized(this) blocks > > > Key: HADOOP-11772 > URL: https://issues.apache.org/jira/browse/HADOOP-11772 > Project: Hadoop Common > Issue Type: Sub-task > Components: ipc, performance >Reporter: Gopal V >Assignee: Akira AJISAKA > Attachments: HADOOP-11772-001.patch, HADOOP-11772-wip-001.patch, > HADOOP-11772-wip-002.patch, dfs-sync-ipc.png, sync-client-bt.png, > sync-client-threads.png > > > {code} > private static ClientCache CLIENTS=new ClientCache(); > ... > this.client = CLIENTS.getClient(conf, factory); > {code} > Meanwhile in ClientCache > {code} > public synchronized Client getClient(Configuration conf, > SocketFactory factory, Class valueClass) { > ... >Client client = clients.get(factory); > if (client == null) { > client = new Client(valueClass, conf, factory); > clients.put(factory, client); > } else { > client.incCount(); > } > {code} > All invokers end up calling these methods, resulting in IPC clients choking > up. > !sync-client-threads.png! > !sync-client-bt.png! > !dfs-sync-ipc.png! -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HADOOP-11649) Allow to configure multiple erasure codecs
[ https://issues.apache.org/jira/browse/HADOOP-11649?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14481053#comment-14481053 ] Vinayakumar B commented on HADOOP-11649: One question I have is, 1. java ServiceLoader will load all the implementations of specified type in the entire classpath. Right? Will it miss any impl to load? 2. The codes specified via configurations also should be in the classpath to get successfully added. IMO these also will be already loaded by java ServiceLoader. right? If all the codecs are loaded by ServiceLoader, then there is no need of configuration to add more codecs. If the above is true, then configuration to specify codes, can be used as a enabler to use only codecs specified in configuration, not all available in classpath( which are loaded by ServiceLoader). Coming to patch, 1. conf name specified in {{public static final String IO_ERASURE_CODECS_KEY = "io.erasure.codecs";}} and {{core-default.xml}} are different. 2. ErasureCodec.java, when there is a explicit dependency mentioned on this Jira, then this class need not be included in the patch. 3. ErasureCodecLoader#load(..) should be changed based on the ans of above question 4. {{io.erasurecode.codec.rs.rawcoder}} adding this configuration seems unrelated to this Jira > Allow to configure multiple erasure codecs > -- > > Key: HADOOP-11649 > URL: https://issues.apache.org/jira/browse/HADOOP-11649 > Project: Hadoop Common > Issue Type: Sub-task > Components: io >Reporter: Kai Zheng >Assignee: Kai Zheng > Attachments: HADOOP-11649-v1.patch > > > This is to allow to configure erasure codec and coder in core-site > configuration file. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HADOOP-11645) Erasure Codec API covering the essential aspects for an erasure code
[ https://issues.apache.org/jira/browse/HADOOP-11645?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14481047#comment-14481047 ] Vinayakumar B commented on HADOOP-11645: Seems like this needs a rebase. > Erasure Codec API covering the essential aspects for an erasure code > > > Key: HADOOP-11645 > URL: https://issues.apache.org/jira/browse/HADOOP-11645 > Project: Hadoop Common > Issue Type: Sub-task >Reporter: Kai Zheng >Assignee: Kai Zheng > Attachments: HADOOP-11645-v1.patch, HADOOP-11645-v2.patch > > > This is to define the even higher level API *ErasureCodec* to possiblly > consider all the essential aspects for an erasure code, as discussed in in > HDFS-7337 in details. Generally, it will cover the necessary configurations > about which *RawErasureCoder* to use for the code scheme, how to form and > layout the BlockGroup, and etc. It will also discuss how an *ErasureCodec* > will be used in both client and DataNode, in all the supported modes related > to EC. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HADOOP-11645) Erasure Codec API covering the essential aspects for an erasure code
[ https://issues.apache.org/jira/browse/HADOOP-11645?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14481046#comment-14481046 ] Vinayakumar B commented on HADOOP-11645: +1, patch looks good. > Erasure Codec API covering the essential aspects for an erasure code > > > Key: HADOOP-11645 > URL: https://issues.apache.org/jira/browse/HADOOP-11645 > Project: Hadoop Common > Issue Type: Sub-task >Reporter: Kai Zheng >Assignee: Kai Zheng > Attachments: HADOOP-11645-v1.patch, HADOOP-11645-v2.patch > > > This is to define the even higher level API *ErasureCodec* to possiblly > consider all the essential aspects for an erasure code, as discussed in in > HDFS-7337 in details. Generally, it will cover the necessary configurations > about which *RawErasureCoder* to use for the code scheme, how to form and > layout the BlockGroup, and etc. It will also discuss how an *ErasureCodec* > will be used in both client and DataNode, in all the supported modes related > to EC. -- This message was sent by Atlassian JIRA (v6.3.4#6332)