svn commit: r1837968 - in /hadoop/common/site/main/publish: bylaws.pdf committer_criteria.pdf index.pdf issue_tracking.pdf linkmap.pdf mailing_lists.pdf privacy_policy.pdf releases.html releases.pdf v
Author: vinodkv Date: Mon Aug 13 16:51:15 2018 New Revision: 1837968 URL: http://svn.apache.org/viewvc?rev=1837968=rev Log: Checking in generated site files. Modified: hadoop/common/site/main/publish/bylaws.pdf hadoop/common/site/main/publish/committer_criteria.pdf hadoop/common/site/main/publish/index.pdf hadoop/common/site/main/publish/issue_tracking.pdf hadoop/common/site/main/publish/linkmap.pdf hadoop/common/site/main/publish/mailing_lists.pdf hadoop/common/site/main/publish/privacy_policy.pdf hadoop/common/site/main/publish/releases.html hadoop/common/site/main/publish/releases.pdf hadoop/common/site/main/publish/version_control.pdf hadoop/common/site/main/publish/versioning.pdf hadoop/common/site/main/publish/who.pdf Modified: hadoop/common/site/main/publish/bylaws.pdf URL: http://svn.apache.org/viewvc/hadoop/common/site/main/publish/bylaws.pdf?rev=1837968=1837967=1837968=diff == Binary files - no diff available. Modified: hadoop/common/site/main/publish/committer_criteria.pdf URL: http://svn.apache.org/viewvc/hadoop/common/site/main/publish/committer_criteria.pdf?rev=1837968=1837967=1837968=diff == Binary files - no diff available. Modified: hadoop/common/site/main/publish/index.pdf URL: http://svn.apache.org/viewvc/hadoop/common/site/main/publish/index.pdf?rev=1837968=1837967=1837968=diff == Binary files - no diff available. Modified: hadoop/common/site/main/publish/issue_tracking.pdf URL: http://svn.apache.org/viewvc/hadoop/common/site/main/publish/issue_tracking.pdf?rev=1837968=1837967=1837968=diff == Binary files - no diff available. Modified: hadoop/common/site/main/publish/linkmap.pdf URL: http://svn.apache.org/viewvc/hadoop/common/site/main/publish/linkmap.pdf?rev=1837968=1837967=1837968=diff == Binary files - no diff available. Modified: hadoop/common/site/main/publish/mailing_lists.pdf URL: http://svn.apache.org/viewvc/hadoop/common/site/main/publish/mailing_lists.pdf?rev=1837968=1837967=1837968=diff == Binary files - no diff available. Modified: hadoop/common/site/main/publish/privacy_policy.pdf URL: http://svn.apache.org/viewvc/hadoop/common/site/main/publish/privacy_policy.pdf?rev=1837968=1837967=1837968=diff == Binary files - no diff available. Modified: hadoop/common/site/main/publish/releases.html URL: http://svn.apache.org/viewvc/hadoop/common/site/main/publish/releases.html?rev=1837968=1837967=1837968=diff == --- hadoop/common/site/main/publish/releases.html (original) +++ hadoop/common/site/main/publish/releases.html Mon Aug 13 16:51:15 2018 @@ -307,16 +307,6 @@ document.write("Last Published: " + docu -2.7.7 - 20 July, 2018 - http://www.apache.org/dyn/closer.cgi/hadoop/common/hadoop-2.7.7/hadoop-2.7.7.tar.gz;>binary - https://dist.apache.org/repos/dist/release/hadoop/common/hadoop-2.7.7/hadoop-2.7.7.tar.gz.asc;>signature - https://dist.apache.org/repos/dist/release/hadoop/common/hadoop-2.7.7/hadoop-2.7.7.tar.gz.mds;>checksum file - - - - - 3.0.3 31 May, 2018 http://www.apache.org/dyn/closer.cgi/hadoop/common/hadoop-3.0.3/hadoop-3.0.3-src.tar.gz;>source @@ -327,21 +317,11 @@ document.write("Last Published: " + docu -2.8.4 - 15 May, 2018 - http://www.apache.org/dyn/closer.cgi/hadoop/common/hadoop-2.8.4/hadoop-2.8.4-src.tar.gz;>source - https://dist.apache.org/repos/dist/release/hadoop/common/hadoop-2.8.4/hadoop-2.8.4-src.tar.gz.asc;>signature - https://dist.apache.org/repos/dist/release/hadoop/common/hadoop-2.8.4/hadoop-2.8.4-src.tar.gz.mds;>checksum file - - - - - - http://www.apache.org/dyn/closer.cgi/hadoop/common/hadoop-2.8.4/hadoop-2.8.4.tar.gz;>binary - https://dist.apache.org/repos/dist/release/hadoop/common/hadoop-2.8.4/hadoop-2.8.4.tar.gz.asc;>signature - https://dist.apache.org/repos/dist/release/hadoop/common/hadoop-2.8.4/hadoop-2.8.4.tar.gz.mds;>checksum file + http://www.apache.org/dyn/closer.cgi/hadoop/common/hadoop-3.0.3/hadoop-3.0.3.tar.gz;>binary + https://dist.apache.org/repos/dist/release/hadoop/common/hadoop-3.0.3/hadoop-3.0.3.tar.gz.asc;>signature + https://dist.apache.org/repos/dist/release/hadoop/com
svn commit: r1837967 - /hadoop/common/site/main/author/src/documentation/content/xdocs/releases.xml
Author: vinodkv Date: Mon Aug 13 16:46:09 2018 New Revision: 1837967 URL: http://svn.apache.org/viewvc?rev=1837967=rev Log: Added the missing links to the 3.0.3 binary bits. Modified: hadoop/common/site/main/author/src/documentation/content/xdocs/releases.xml Modified: hadoop/common/site/main/author/src/documentation/content/xdocs/releases.xml URL: http://svn.apache.org/viewvc/hadoop/common/site/main/author/src/documentation/content/xdocs/releases.xml?rev=1837967=1837966=1837967=diff == --- hadoop/common/site/main/author/src/documentation/content/xdocs/releases.xml (original) +++ hadoop/common/site/main/author/src/documentation/content/xdocs/releases.xml Mon Aug 13 16:46:09 2018 @@ -51,6 +51,13 @@ https://dist.apache.org/repos/dist/release/hadoop/common/hadoop-3.0.3/hadoop-3.0.3-src.tar.gz.mds;>checksum file + + + http://www.apache.org/dyn/closer.cgi/hadoop/common/hadoop-3.0.3/hadoop-3.0.3.tar.gz;>binary + https://dist.apache.org/repos/dist/release/hadoop/common/hadoop-3.0.3/hadoop-3.0.3.tar.gz.asc;>signature + https://dist.apache.org/repos/dist/release/hadoop/common/hadoop-3.0.3/hadoop-3.0.3.tar.gz.mds;>checksum file + + 2.9.1 3 May, 2018 http://www.apache.org/dyn/closer.cgi/hadoop/common/hadoop-2.9.1/hadoop-2.9.1-src.tar.gz;>source - To unsubscribe, e-mail: common-commits-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-commits-h...@hadoop.apache.org
svn commit: r1837966 - /hadoop/common/site/main/author/src/documentation/content/xdocs/releases.xml
Author: vinodkv Date: Mon Aug 13 16:36:53 2018 New Revision: 1837966 URL: http://svn.apache.org/viewvc?rev=1837966=rev Log: Added the missing links to the 2.7.7 source bits. Modified: hadoop/common/site/main/author/src/documentation/content/xdocs/releases.xml Modified: hadoop/common/site/main/author/src/documentation/content/xdocs/releases.xml URL: http://svn.apache.org/viewvc/hadoop/common/site/main/author/src/documentation/content/xdocs/releases.xml?rev=1837966=1837965=1837966=diff == --- hadoop/common/site/main/author/src/documentation/content/xdocs/releases.xml (original) +++ hadoop/common/site/main/author/src/documentation/content/xdocs/releases.xml Mon Aug 13 16:36:53 2018 @@ -81,6 +81,13 @@ 2.7.7 20 July, 2018 + http://www.apache.org/dyn/closer.cgi/hadoop/common/hadoop-2.7.7/hadoop-2.7.7-src.tar.gz;>source + https://dist.apache.org/repos/dist/release/hadoop/common/hadoop-2.7.7/hadoop-2.7.7-src.tar.gz.asc;>signature + https://dist.apache.org/repos/dist/release/hadoop/common/hadoop-2.7.7/hadoop-2.7.7-src.tar.gz.mds;>checksum file + + + + http://www.apache.org/dyn/closer.cgi/hadoop/common/hadoop-2.7.7/hadoop-2.7.7.tar.gz;>binary https://dist.apache.org/repos/dist/release/hadoop/common/hadoop-2.7.7/hadoop-2.7.7.tar.gz.asc;>signature https://dist.apache.org/repos/dist/release/hadoop/common/hadoop-2.7.7/hadoop-2.7.7.tar.gz.mds;>checksum file - To unsubscribe, e-mail: common-commits-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-commits-h...@hadoop.apache.org
svn commit: r1837963 - /hadoop/common/site/main/author/src/documentation/content/xdocs/releases.xml
Author: vinodkv Date: Mon Aug 13 16:18:40 2018 New Revision: 1837963 URL: http://svn.apache.org/viewvc?rev=1837963=rev Log: Fixing releases.html to only show the latest release in every line. Also fixing their order. Modified: hadoop/common/site/main/author/src/documentation/content/xdocs/releases.xml Modified: hadoop/common/site/main/author/src/documentation/content/xdocs/releases.xml URL: http://svn.apache.org/viewvc/hadoop/common/site/main/author/src/documentation/content/xdocs/releases.xml?rev=1837963=1837962=1837963=diff == --- hadoop/common/site/main/author/src/documentation/content/xdocs/releases.xml (original) +++ hadoop/common/site/main/author/src/documentation/content/xdocs/releases.xml Mon Aug 13 16:18:40 2018 @@ -44,13 +44,6 @@ https://dist.apache.org/repos/dist/release/hadoop/common/hadoop-3.1.1/hadoop-3.1.1.tar.gz.mds;>checksum file - 2.7.7 - 20 July, 2018 - http://www.apache.org/dyn/closer.cgi/hadoop/common/hadoop-2.7.7/hadoop-2.7.7.tar.gz;>binary - https://dist.apache.org/repos/dist/release/hadoop/common/hadoop-2.7.7/hadoop-2.7.7.tar.gz.asc;>signature - https://dist.apache.org/repos/dist/release/hadoop/common/hadoop-2.7.7/hadoop-2.7.7.tar.gz.mds;>checksum file - - 3.0.3 31 May, 2018 http://www.apache.org/dyn/closer.cgi/hadoop/common/hadoop-3.0.3/hadoop-3.0.3-src.tar.gz;>source @@ -58,20 +51,6 @@ https://dist.apache.org/repos/dist/release/hadoop/common/hadoop-3.0.3/hadoop-3.0.3-src.tar.gz.mds;>checksum file - 2.8.4 - 15 May, 2018 - http://www.apache.org/dyn/closer.cgi/hadoop/common/hadoop-2.8.4/hadoop-2.8.4-src.tar.gz;>source - https://dist.apache.org/repos/dist/release/hadoop/common/hadoop-2.8.4/hadoop-2.8.4-src.tar.gz.asc;>signature - https://dist.apache.org/repos/dist/release/hadoop/common/hadoop-2.8.4/hadoop-2.8.4-src.tar.gz.mds;>checksum file - - - - - http://www.apache.org/dyn/closer.cgi/hadoop/common/hadoop-2.8.4/hadoop-2.8.4.tar.gz;>binary - https://dist.apache.org/repos/dist/release/hadoop/common/hadoop-2.8.4/hadoop-2.8.4.tar.gz.asc;>signature - https://dist.apache.org/repos/dist/release/hadoop/common/hadoop-2.8.4/hadoop-2.8.4.tar.gz.mds;>checksum file - - 2.9.1 3 May, 2018 http://www.apache.org/dyn/closer.cgi/hadoop/common/hadoop-2.9.1/hadoop-2.9.1-src.tar.gz;>source @@ -86,74 +65,25 @@ https://dist.apache.org/repos/dist/release/hadoop/common/hadoop-2.9.1/hadoop-2.9.1.tar.gz.mds;>checksum file - 3.1.0 - 6 Apr, 2018 - http://www.apache.org/dyn/closer.cgi/hadoop/common/hadoop-3.1.0/hadoop-3.1.0-src.tar.gz;>source - https://dist.apache.org/repos/dist/release/hadoop/common/hadoop-3.1.0/hadoop-3.1.0-src.tar.gz.asc;>signature - https://dist.apache.org/repos/dist/release/hadoop/common/hadoop-3.1.0/hadoop-3.1.0-src.tar.gz.mds;>checksum file - - - - - http://www.apache.org/dyn/closer.cgi/hadoop/common/hadoop-3.1.0/hadoop-3.1.0.tar.gz;>binary - https://dist.apache.org/repos/dist/release/hadoop/common/hadoop-3.1.0/hadoop-3.1.0.tar.gz.asc;>signature - https://dist.apache.org/repos/dist/release/hadoop/common/hadoop-3.1.0/hadoop-3.1.0.tar.gz.mds;>checksum file - - - 3.0.2 - 21 April, 2018 - http://www.apache.org/dyn/closer.cgi/hadoop/common/hadoop-3.0.2/hadoop-3.0.2-src.tar.gz;>source - https://dist.apache.org/repos/dist/release/hadoop/common/hadoop-3.0.2/hadoop-3.0.2-src.tar.gz.asc;>signature - https://dist.apache.org/repos/dist/release/hadoop/common/hadoop-3.0.2/hadoop-3.0.2-src.tar.gz.mds;>checksum file - - - - - http://www.apache.org/dyn/closer.cgi/hadoop/common/hadoop-3.0.2/hadoop-3.0.2.tar.gz;>binary - https://dist.apache.org/repos/dist/release/hadoop/common/hadoop-3.0.2/hadoop-3.0.2.tar.gz.asc;>signature - https://dist.apache.org/repos/dist/release/hadoop/common/hadoop-3.0.2/hadoop-3.0.2.tar.gz.mds;>checksum file - - - 2.9.0 - 17 November, 2017 - http://www.apache.org/dyn/closer.cgi/hadoop/common/hadoop-2.9.0/hadoop-2.9.0-src.tar.gz;>source - https://dist.apache.org/repos/dist/release/hadoop/common/hadoop-2.9.0/hadoop-2.9.0-src.tar.gz.asc;>signature - https://dist.apache.org/repos/dist/release/hadoop/common/hadoop-2.9.0/hadoop-2.9.0-src.tar.gz.mds;>checksum file - - - - - http://www.apache.org/dyn/closer.cgi/hadoop/common/hadoop-2.9.0/hadoop-2.9.0.tar.gz;>binary - https://dist.apache.org/repos/
hadoop git commit: YARN-8197. Fixed AM IP Filter and Webapp proxy to redirect app tracking-URLs correctly when UI is secure. Contributed by Sunil Govindan.
Repository: hadoop Updated Branches: refs/heads/branch-3.1 442dd87dc -> a468400ba YARN-8197. Fixed AM IP Filter and Webapp proxy to redirect app tracking-URLs correctly when UI is secure. Contributed by Sunil Govindan. (cherry picked from commit 6b74f5d7fc509c55c331249256eec78b7e53b6ce) Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/a468400b Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/a468400b Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/a468400b Branch: refs/heads/branch-3.1 Commit: a468400bad1a5156a0007453ea69820631725ffa Parents: 442dd87 Author: Vinod Kumar Vavilapalli (I am also known as @tshooter.) Authored: Thu May 31 16:48:33 2018 -0700 Committer: Vinod Kumar Vavilapalli (I am also known as @tshooter.) Committed: Thu May 31 17:04:45 2018 -0700 -- .../hadoop-yarn-server-web-proxy/pom.xml| 13 ++ .../server/webproxy/amfilter/AmIpFilter.java| 19 ++- .../webproxy/amfilter/TestSecureAmFilter.java | 159 +++ .../src/test/resources/krb5.conf| 33 4 files changed, 220 insertions(+), 4 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hadoop/blob/a468400b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-web-proxy/pom.xml -- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-web-proxy/pom.xml b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-web-proxy/pom.xml index 8658e44..8035f57 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-web-proxy/pom.xml +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-web-proxy/pom.xml @@ -52,6 +52,19 @@ + org.apache.hadoop + hadoop-auth + test-jar + test + + + + org.apache.hadoop + hadoop-minikdc + provided + + + org.mockito mockito-all test http://git-wip-us.apache.org/repos/asf/hadoop/blob/a468400b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-web-proxy/src/main/java/org/apache/hadoop/yarn/server/webproxy/amfilter/AmIpFilter.java -- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-web-proxy/src/main/java/org/apache/hadoop/yarn/server/webproxy/amfilter/AmIpFilter.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-web-proxy/src/main/java/org/apache/hadoop/yarn/server/webproxy/amfilter/AmIpFilter.java index ac4afa8..b5e7ef6 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-web-proxy/src/main/java/org/apache/hadoop/yarn/server/webproxy/amfilter/AmIpFilter.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-web-proxy/src/main/java/org/apache/hadoop/yarn/server/webproxy/amfilter/AmIpFilter.java @@ -20,6 +20,7 @@ package org.apache.hadoop.yarn.server.webproxy.amfilter; import com.google.common.annotations.VisibleForTesting; import org.apache.hadoop.classification.InterfaceAudience.Public; +import org.apache.hadoop.security.UserGroupInformation; import org.apache.hadoop.yarn.server.webproxy.ProxyUtils; import org.apache.hadoop.yarn.server.webproxy.WebAppProxyServlet; import org.slf4j.Logger; @@ -214,15 +215,25 @@ public class AmIpFilter implements Filter { return addr; } - private boolean isValidUrl(String url) { + @VisibleForTesting + public boolean isValidUrl(String url) { boolean isValid = false; try { - HttpURLConnection conn = - (HttpURLConnection) new URL(url).openConnection(); + HttpURLConnection conn = (HttpURLConnection) new URL(url) + .openConnection(); conn.connect(); isValid = conn.getResponseCode() == HttpURLConnection.HTTP_OK; + // If security is enabled, any valid RM which can give 401 Unauthorized is + // good enough to access. Since AM doesn't have enough credential, auth + // cannot be completed and hence 401 is fine in such case. + if (!isValid && UserGroupInformation.isSecurityEnabled()) { +isValid = (conn +.getResponseCode() == HttpURLConnection.HTTP_UNAUTHORIZED) +|| (conn.getResponseCode() == HttpURLConnection.HTTP_FORBIDDEN); +return isValid; + } } catch (Exception e) { - LOG.debug("Failed to connect to " + url + ": " + e.toString()); + LOG.warn("Failed to connect to " + url + ": " + e.toString()); } return isValid; }
hadoop git commit: YARN-8197. Fixed AM IP Filter and Webapp proxy to redirect app tracking-URLs correctly when UI is secure. Contributed by Sunil Govindan.
Repository: hadoop Updated Branches: refs/heads/trunk 32671d871 -> 6b74f5d7f YARN-8197. Fixed AM IP Filter and Webapp proxy to redirect app tracking-URLs correctly when UI is secure. Contributed by Sunil Govindan. Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/6b74f5d7 Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/6b74f5d7 Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/6b74f5d7 Branch: refs/heads/trunk Commit: 6b74f5d7fc509c55c331249256eec78b7e53b6ce Parents: 32671d8 Author: Vinod Kumar Vavilapalli (I am also known as @tshooter.) Authored: Thu May 31 16:48:33 2018 -0700 Committer: Vinod Kumar Vavilapalli (I am also known as @tshooter.) Committed: Thu May 31 16:48:33 2018 -0700 -- .../hadoop-yarn-server-web-proxy/pom.xml| 13 ++ .../server/webproxy/amfilter/AmIpFilter.java| 19 ++- .../webproxy/amfilter/TestSecureAmFilter.java | 159 +++ .../src/test/resources/krb5.conf| 33 4 files changed, 220 insertions(+), 4 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hadoop/blob/6b74f5d7/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-web-proxy/pom.xml -- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-web-proxy/pom.xml b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-web-proxy/pom.xml index 61e0429..0d1b92b 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-web-proxy/pom.xml +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-web-proxy/pom.xml @@ -52,6 +52,19 @@ + org.apache.hadoop + hadoop-auth + test-jar + test + + + + org.apache.hadoop + hadoop-minikdc + provided + + + org.mockito mockito-all test http://git-wip-us.apache.org/repos/asf/hadoop/blob/6b74f5d7/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-web-proxy/src/main/java/org/apache/hadoop/yarn/server/webproxy/amfilter/AmIpFilter.java -- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-web-proxy/src/main/java/org/apache/hadoop/yarn/server/webproxy/amfilter/AmIpFilter.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-web-proxy/src/main/java/org/apache/hadoop/yarn/server/webproxy/amfilter/AmIpFilter.java index bd425a7..c965283 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-web-proxy/src/main/java/org/apache/hadoop/yarn/server/webproxy/amfilter/AmIpFilter.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-web-proxy/src/main/java/org/apache/hadoop/yarn/server/webproxy/amfilter/AmIpFilter.java @@ -20,6 +20,7 @@ package org.apache.hadoop.yarn.server.webproxy.amfilter; import com.google.common.annotations.VisibleForTesting; import org.apache.hadoop.classification.InterfaceAudience.Public; +import org.apache.hadoop.security.UserGroupInformation; import org.apache.hadoop.util.Time; import org.apache.hadoop.yarn.server.webproxy.ProxyUtils; import org.apache.hadoop.yarn.server.webproxy.WebAppProxyServlet; @@ -216,15 +217,25 @@ public class AmIpFilter implements Filter { return addr; } - private boolean isValidUrl(String url) { + @VisibleForTesting + public boolean isValidUrl(String url) { boolean isValid = false; try { - HttpURLConnection conn = - (HttpURLConnection) new URL(url).openConnection(); + HttpURLConnection conn = (HttpURLConnection) new URL(url) + .openConnection(); conn.connect(); isValid = conn.getResponseCode() == HttpURLConnection.HTTP_OK; + // If security is enabled, any valid RM which can give 401 Unauthorized is + // good enough to access. Since AM doesn't have enough credential, auth + // cannot be completed and hence 401 is fine in such case. + if (!isValid && UserGroupInformation.isSecurityEnabled()) { +isValid = (conn +.getResponseCode() == HttpURLConnection.HTTP_UNAUTHORIZED) +|| (conn.getResponseCode() == HttpURLConnection.HTTP_FORBIDDEN); +return isValid; + } } catch (Exception e) { - LOG.debug("Failed to connect to " + url + ": " + e.toString()); + LOG.warn("Failed to connect to " + url + ": " + e.toString()); } return isValid; } http://git-wip-us.apache.org/repos/asf/hadoop/blob/6b74f5d7/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-web-proxy/src/test/java/org/apache/hadoop/yarn/server/webproxy/amfilter/TestSecureAmFilter.java
hadoop git commit: YARN-8128. Documenting the per-node per-app file limit in YARN log aggregation. Contributed by Xuan Gong.
Repository: hadoop Updated Branches: refs/heads/branch-3.1 170f1040d -> 0460201c3 YARN-8128. Documenting the per-node per-app file limit in YARN log aggregation. Contributed by Xuan Gong. (cherry picked from commit 328f0847e3d98cad8c368d57499f31081c153237) Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/0460201c Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/0460201c Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/0460201c Branch: refs/heads/branch-3.1 Commit: 0460201c3a505144202bd899ceb1586beb6e6ab6 Parents: 170f104 Author: Vinod Kumar Vavilapalli (I am also known as @tshooter.)Authored: Thu May 17 12:47:10 2018 -0700 Committer: Vinod Kumar Vavilapalli (I am also known as @tshooter.) Committed: Thu May 17 13:09:09 2018 -0700 -- .../apache/hadoop/yarn/conf/YarnConfiguration.java | 10 ++ .../LogAggregationFileController.java | 17 ++--- .../src/main/resources/yarn-default.xml| 9 + 3 files changed, 25 insertions(+), 11 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hadoop/blob/0460201c/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/java/org/apache/hadoop/yarn/conf/YarnConfiguration.java -- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/java/org/apache/hadoop/yarn/conf/YarnConfiguration.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/java/org/apache/hadoop/yarn/conf/YarnConfiguration.java index 208f85c..7683821 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/java/org/apache/hadoop/yarn/conf/YarnConfiguration.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/java/org/apache/hadoop/yarn/conf/YarnConfiguration.java @@ -1357,6 +1357,16 @@ public class YarnConfiguration extends Configuration { NM_PREFIX + "log-aggregation.roll-monitoring-interval-seconds"; public static final long DEFAULT_NM_LOG_AGGREGATION_ROLL_MONITORING_INTERVAL_SECONDS = -1; + + /** + * Define how many aggregated log files per application per NM we can have + * in remote file system. + */ + public static final String NM_LOG_AGGREGATION_NUM_LOG_FILES_SIZE_PER_APP + = NM_PREFIX + "log-aggregation.num-log-files-per-app"; + public static final int + DEFAULT_NM_LOG_AGGREGATION_NUM_LOG_FILES_SIZE_PER_APP = 30; + /** * Number of threads used in log cleanup. Only applicable if Log aggregation * is disabled http://git-wip-us.apache.org/repos/asf/hadoop/blob/0460201c/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/logaggregation/filecontroller/LogAggregationFileController.java -- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/logaggregation/filecontroller/LogAggregationFileController.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/logaggregation/filecontroller/LogAggregationFileController.java index aeef574..5ac89e9 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/logaggregation/filecontroller/LogAggregationFileController.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/logaggregation/filecontroller/LogAggregationFileController.java @@ -95,13 +95,6 @@ public abstract class LogAggregationFileController { protected static final FsPermission APP_LOG_FILE_UMASK = FsPermission .createImmutable((short) (0640 ^ 0777)); - // This is temporary solution. The configuration will be deleted once - // we find a more scalable method to only write a single log file per LRS. - private static final String NM_LOG_AGGREGATION_NUM_LOG_FILES_SIZE_PER_APP - = YarnConfiguration.NM_PREFIX + "log-aggregation.num-log-files-per-app"; - private static final int - DEFAULT_NM_LOG_AGGREGATION_NUM_LOG_FILES_SIZE_PER_APP = 30; - // This is temporary solution. The configuration will be deleted once we have // the FileSystem API to check whether append operation is supported or not. public static final String LOG_AGGREGATION_FS_SUPPORT_APPEND @@ -122,12 +115,14 @@ public abstract class LogAggregationFileController { */ public void initialize(Configuration conf, String controllerName) { this.conf = conf; -int configuredRentionSize = -conf.getInt(NM_LOG_AGGREGATION_NUM_LOG_FILES_SIZE_PER_APP, -DEFAULT_NM_LOG_AGGREGATION_NUM_LOG_FILES_SIZE_PER_APP); +int configuredRentionSize = conf.getInt( +
hadoop git commit: YARN-8128. Documenting the per-node per-app file limit in YARN log aggregation. Contributed by Xuan Gong.
Repository: hadoop Updated Branches: refs/heads/trunk a2cdffb95 -> 328f0847e YARN-8128. Documenting the per-node per-app file limit in YARN log aggregation. Contributed by Xuan Gong. Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/328f0847 Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/328f0847 Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/328f0847 Branch: refs/heads/trunk Commit: 328f0847e3d98cad8c368d57499f31081c153237 Parents: a2cdffb Author: Vinod Kumar Vavilapalli (I am also known as @tshooter.)Authored: Thu May 17 12:47:10 2018 -0700 Committer: Vinod Kumar Vavilapalli (I am also known as @tshooter.) Committed: Thu May 17 12:47:10 2018 -0700 -- .../apache/hadoop/yarn/conf/YarnConfiguration.java | 10 ++ .../LogAggregationFileController.java | 17 ++--- .../src/main/resources/yarn-default.xml| 9 + 3 files changed, 25 insertions(+), 11 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hadoop/blob/328f0847/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/java/org/apache/hadoop/yarn/conf/YarnConfiguration.java -- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/java/org/apache/hadoop/yarn/conf/YarnConfiguration.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/java/org/apache/hadoop/yarn/conf/YarnConfiguration.java index 5ba2e05..8e56cb8 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/java/org/apache/hadoop/yarn/conf/YarnConfiguration.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/java/org/apache/hadoop/yarn/conf/YarnConfiguration.java @@ -1377,6 +1377,16 @@ public class YarnConfiguration extends Configuration { NM_PREFIX + "log-aggregation.roll-monitoring-interval-seconds"; public static final long DEFAULT_NM_LOG_AGGREGATION_ROLL_MONITORING_INTERVAL_SECONDS = -1; + + /** + * Define how many aggregated log files per application per NM we can have + * in remote file system. + */ + public static final String NM_LOG_AGGREGATION_NUM_LOG_FILES_SIZE_PER_APP + = NM_PREFIX + "log-aggregation.num-log-files-per-app"; + public static final int + DEFAULT_NM_LOG_AGGREGATION_NUM_LOG_FILES_SIZE_PER_APP = 30; + /** * Number of threads used in log cleanup. Only applicable if Log aggregation * is disabled http://git-wip-us.apache.org/repos/asf/hadoop/blob/328f0847/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/logaggregation/filecontroller/LogAggregationFileController.java -- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/logaggregation/filecontroller/LogAggregationFileController.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/logaggregation/filecontroller/LogAggregationFileController.java index aeef574..5ac89e9 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/logaggregation/filecontroller/LogAggregationFileController.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/logaggregation/filecontroller/LogAggregationFileController.java @@ -95,13 +95,6 @@ public abstract class LogAggregationFileController { protected static final FsPermission APP_LOG_FILE_UMASK = FsPermission .createImmutable((short) (0640 ^ 0777)); - // This is temporary solution. The configuration will be deleted once - // we find a more scalable method to only write a single log file per LRS. - private static final String NM_LOG_AGGREGATION_NUM_LOG_FILES_SIZE_PER_APP - = YarnConfiguration.NM_PREFIX + "log-aggregation.num-log-files-per-app"; - private static final int - DEFAULT_NM_LOG_AGGREGATION_NUM_LOG_FILES_SIZE_PER_APP = 30; - // This is temporary solution. The configuration will be deleted once we have // the FileSystem API to check whether append operation is supported or not. public static final String LOG_AGGREGATION_FS_SUPPORT_APPEND @@ -122,12 +115,14 @@ public abstract class LogAggregationFileController { */ public void initialize(Configuration conf, String controllerName) { this.conf = conf; -int configuredRentionSize = -conf.getInt(NM_LOG_AGGREGATION_NUM_LOG_FILES_SIZE_PER_APP, -DEFAULT_NM_LOG_AGGREGATION_NUM_LOG_FILES_SIZE_PER_APP); +int configuredRentionSize = conf.getInt( +YarnConfiguration.NM_LOG_AGGREGATION_NUM_LOG_FILES_SIZE_PER_APP, +YarnConfiguration +
hadoop git commit: YARN-8249. Fixed few REST APIs in RMWebServices to have static-user check. Contributed by Sunil Govindan.
Repository: hadoop Updated Branches: refs/heads/branch-3.1 0506c762b -> 4c0a6b67f YARN-8249. Fixed few REST APIs in RMWebServices to have static-user check. Contributed by Sunil Govindan. (cherry picked from commit d76fbbc9b82e720d7d5188f9ae2f56a8d78f3a98) Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/4c0a6b67 Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/4c0a6b67 Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/4c0a6b67 Branch: refs/heads/branch-3.1 Commit: 4c0a6b67f3f6c76cf92f8e4f5a529e6387c5aa27 Parents: 0506c76 Author: Vinod Kumar Vavilapalli (I am also known as @tshooter.)Authored: Thu May 10 19:03:23 2018 -0700 Committer: Vinod Kumar Vavilapalli (I am also known as @tshooter.) Committed: Thu May 10 19:07:23 2018 -0700 -- .../webapp/AHSWebServices.java | 18 +- .../hadoop/yarn/server/webapp/WebServices.java | 2 +- .../resourcemanager/webapp/RMWebServices.java | 302 --- .../webapp/TestRMWebServices.java | 2 +- ...tRMWebServicesHttpStaticUserPermissions.java | 12 +- 5 files changed, 142 insertions(+), 194 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hadoop/blob/4c0a6b67/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/src/main/java/org/apache/hadoop/yarn/server/applicationhistoryservice/webapp/AHSWebServices.java -- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/src/main/java/org/apache/hadoop/yarn/server/applicationhistoryservice/webapp/AHSWebServices.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/src/main/java/org/apache/hadoop/yarn/server/applicationhistoryservice/webapp/AHSWebServices.java index 755127b..9aa71a7 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/src/main/java/org/apache/hadoop/yarn/server/applicationhistoryservice/webapp/AHSWebServices.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/src/main/java/org/apache/hadoop/yarn/server/applicationhistoryservice/webapp/AHSWebServices.java @@ -111,7 +111,7 @@ public class AHSWebServices extends WebServices { public TimelineAbout about( @Context HttpServletRequest req, @Context HttpServletResponse res) { -init(res); +initForReadableEndpoints(res); return TimelineUtils.createTimelineAbout("Generic History Service API"); } @@ -141,7 +141,7 @@ public class AHSWebServices extends WebServices { @QueryParam("finishedTimeBegin") String finishBegin, @QueryParam("finishedTimeEnd") String finishEnd, @QueryParam("applicationTypes") Set applicationTypes) { -init(res); +initForReadableEndpoints(res); validateStates(stateQuery, statesQuery); return super.getApps(req, res, stateQuery, statesQuery, finalStatusQuery, userQuery, queueQuery, count, startedBegin, startedEnd, finishBegin, @@ -155,7 +155,7 @@ public class AHSWebServices extends WebServices { @Override public AppInfo getApp(@Context HttpServletRequest req, @Context HttpServletResponse res, @PathParam("appid") String appId) { -init(res); +initForReadableEndpoints(res); return super.getApp(req, res, appId); } @@ -166,7 +166,7 @@ public class AHSWebServices extends WebServices { @Override public AppAttemptsInfo getAppAttempts(@Context HttpServletRequest req, @Context HttpServletResponse res, @PathParam("appid") String appId) { -init(res); +initForReadableEndpoints(res); return super.getAppAttempts(req, res, appId); } @@ -178,7 +178,7 @@ public class AHSWebServices extends WebServices { public AppAttemptInfo getAppAttempt(@Context HttpServletRequest req, @Context HttpServletResponse res, @PathParam("appid") String appId, @PathParam("appattemptid") String appAttemptId) { -init(res); +initForReadableEndpoints(res); return super.getAppAttempt(req, res, appId, appAttemptId); } @@ -190,7 +190,7 @@ public class AHSWebServices extends WebServices { public ContainersInfo getContainers(@Context HttpServletRequest req, @Context HttpServletResponse res, @PathParam("appid") String appId, @PathParam("appattemptid") String appAttemptId) { -init(res); +initForReadableEndpoints(res); return super.getContainers(req, res, appId, appAttemptId); } @@ -203,7 +203,7 @@ public class AHSWebServices extends WebServices { @Context HttpServletResponse res, @PathParam("appid") String appId,
hadoop git commit: YARN-8249. Fixed few REST APIs in RMWebServices to have static-user check. Contributed by Sunil Govindan.
Repository: hadoop Updated Branches: refs/heads/trunk 84b305f11 -> d76fbbc9b YARN-8249. Fixed few REST APIs in RMWebServices to have static-user check. Contributed by Sunil Govindan. Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/d76fbbc9 Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/d76fbbc9 Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/d76fbbc9 Branch: refs/heads/trunk Commit: d76fbbc9b82e720d7d5188f9ae2f56a8d78f3a98 Parents: 84b305f Author: Vinod Kumar Vavilapalli (I am also known as @tshooter.)Authored: Thu May 10 19:03:23 2018 -0700 Committer: Vinod Kumar Vavilapalli (I am also known as @tshooter.) Committed: Thu May 10 19:05:35 2018 -0700 -- .../webapp/AHSWebServices.java | 18 +- .../hadoop/yarn/server/webapp/WebServices.java | 2 +- .../resourcemanager/webapp/RMWebServices.java | 302 --- .../webapp/TestRMWebServices.java | 2 +- ...tRMWebServicesHttpStaticUserPermissions.java | 12 +- 5 files changed, 142 insertions(+), 194 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hadoop/blob/d76fbbc9/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/src/main/java/org/apache/hadoop/yarn/server/applicationhistoryservice/webapp/AHSWebServices.java -- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/src/main/java/org/apache/hadoop/yarn/server/applicationhistoryservice/webapp/AHSWebServices.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/src/main/java/org/apache/hadoop/yarn/server/applicationhistoryservice/webapp/AHSWebServices.java index 755127b..9aa71a7 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/src/main/java/org/apache/hadoop/yarn/server/applicationhistoryservice/webapp/AHSWebServices.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/src/main/java/org/apache/hadoop/yarn/server/applicationhistoryservice/webapp/AHSWebServices.java @@ -111,7 +111,7 @@ public class AHSWebServices extends WebServices { public TimelineAbout about( @Context HttpServletRequest req, @Context HttpServletResponse res) { -init(res); +initForReadableEndpoints(res); return TimelineUtils.createTimelineAbout("Generic History Service API"); } @@ -141,7 +141,7 @@ public class AHSWebServices extends WebServices { @QueryParam("finishedTimeBegin") String finishBegin, @QueryParam("finishedTimeEnd") String finishEnd, @QueryParam("applicationTypes") Set applicationTypes) { -init(res); +initForReadableEndpoints(res); validateStates(stateQuery, statesQuery); return super.getApps(req, res, stateQuery, statesQuery, finalStatusQuery, userQuery, queueQuery, count, startedBegin, startedEnd, finishBegin, @@ -155,7 +155,7 @@ public class AHSWebServices extends WebServices { @Override public AppInfo getApp(@Context HttpServletRequest req, @Context HttpServletResponse res, @PathParam("appid") String appId) { -init(res); +initForReadableEndpoints(res); return super.getApp(req, res, appId); } @@ -166,7 +166,7 @@ public class AHSWebServices extends WebServices { @Override public AppAttemptsInfo getAppAttempts(@Context HttpServletRequest req, @Context HttpServletResponse res, @PathParam("appid") String appId) { -init(res); +initForReadableEndpoints(res); return super.getAppAttempts(req, res, appId); } @@ -178,7 +178,7 @@ public class AHSWebServices extends WebServices { public AppAttemptInfo getAppAttempt(@Context HttpServletRequest req, @Context HttpServletResponse res, @PathParam("appid") String appId, @PathParam("appattemptid") String appAttemptId) { -init(res); +initForReadableEndpoints(res); return super.getAppAttempt(req, res, appId, appAttemptId); } @@ -190,7 +190,7 @@ public class AHSWebServices extends WebServices { public ContainersInfo getContainers(@Context HttpServletRequest req, @Context HttpServletResponse res, @PathParam("appid") String appId, @PathParam("appattemptid") String appAttemptId) { -init(res); +initForReadableEndpoints(res); return super.getContainers(req, res, appId, appAttemptId); } @@ -203,7 +203,7 @@ public class AHSWebServices extends WebServices { @Context HttpServletResponse res, @PathParam("appid") String appId, @PathParam("appattemptid") String appAttemptId, @PathParam("containerid") String
hadoop git commit: YARN-7734. Fix UT failure TestContainerLogsPage#testContainerLogPageAccess. Contributed by Tao Yang.
Repository: hadoop Updated Branches: refs/heads/branch-3.1 a13b3edfd -> c3ca10c22 YARN-7734. Fix UT failure TestContainerLogsPage#testContainerLogPageAccess. Contributed by Tao Yang. (cherry picked from commit 411993f6e5723c8cba8100bff0269418e46f6367) Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/c3ca10c2 Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/c3ca10c2 Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/c3ca10c2 Branch: refs/heads/branch-3.1 Commit: c3ca10c22a3179ec3b1f699cccf33884e494653d Parents: a13b3ed Author: Weiwei YangAuthored: Wed Mar 28 17:35:17 2018 +0800 Committer: Vinod Kumar Vavilapalli (I am also known as @tshooter.) Committed: Wed Apr 18 19:40:34 2018 -0700 -- .../yarn/server/nodemanager/webapp/TestContainerLogsPage.java | 1 + 1 file changed, 1 insertion(+) -- http://git-wip-us.apache.org/repos/asf/hadoop/blob/c3ca10c2/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/test/java/org/apache/hadoop/yarn/server/nodemanager/webapp/TestContainerLogsPage.java -- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/test/java/org/apache/hadoop/yarn/server/nodemanager/webapp/TestContainerLogsPage.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/test/java/org/apache/hadoop/yarn/server/nodemanager/webapp/TestContainerLogsPage.java index 8dc06c7..ece1af4 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/test/java/org/apache/hadoop/yarn/server/nodemanager/webapp/TestContainerLogsPage.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/test/java/org/apache/hadoop/yarn/server/nodemanager/webapp/TestContainerLogsPage.java @@ -259,6 +259,7 @@ public class TestContainerLogsPage { new ConcurrentHashMap (); when(context.getContainers()).thenReturn(containers); when(context.getLocalDirsHandler()).thenReturn(dirsHandler); + when(context.getConf()).thenReturn(conf); MockContainer container = new MockContainer(appAttemptId, new AsyncDispatcher(), conf, user, appId, 1); - To unsubscribe, e-mail: common-commits-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-commits-h...@hadoop.apache.org
[2/2] hadoop git commit: Preparing for 3.1.0 development
Preparing for 3.1.0 development Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/0bd99621 Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/0bd99621 Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/0bd99621 Branch: refs/heads/branch-3.1 Commit: 0bd99621c5c63adccfbcbf8c4dd9d2b7cec7b9b1 Parents: 49c747a Author: Vinod Kumar Vavilapalli (I am also known as @tshooter.)Authored: Mon Mar 19 15:16:54 2018 -0700 Committer: Vinod Kumar Vavilapalli (I am also known as @tshooter.) Committed: Mon Mar 19 15:16:54 2018 -0700 -- hadoop-assemblies/pom.xml| 4 ++-- hadoop-build-tools/pom.xml | 2 +- hadoop-client-modules/hadoop-client-api/pom.xml | 4 ++-- hadoop-client-modules/hadoop-client-check-invariants/pom.xml | 4 ++-- .../hadoop-client-check-test-invariants/pom.xml | 4 ++-- hadoop-client-modules/hadoop-client-integration-tests/pom.xml| 4 ++-- hadoop-client-modules/hadoop-client-minicluster/pom.xml | 4 ++-- hadoop-client-modules/hadoop-client-runtime/pom.xml | 4 ++-- hadoop-client-modules/hadoop-client/pom.xml | 4 ++-- hadoop-client-modules/pom.xml| 2 +- hadoop-cloud-storage-project/hadoop-cloud-storage/pom.xml| 4 ++-- hadoop-cloud-storage-project/pom.xml | 4 ++-- hadoop-common-project/hadoop-annotations/pom.xml | 4 ++-- hadoop-common-project/hadoop-auth-examples/pom.xml | 4 ++-- hadoop-common-project/hadoop-auth/pom.xml| 4 ++-- hadoop-common-project/hadoop-common/pom.xml | 4 ++-- hadoop-common-project/hadoop-kms/pom.xml | 4 ++-- hadoop-common-project/hadoop-minikdc/pom.xml | 4 ++-- hadoop-common-project/hadoop-nfs/pom.xml | 4 ++-- hadoop-common-project/pom.xml| 4 ++-- hadoop-dist/pom.xml | 4 ++-- hadoop-hdfs-project/hadoop-hdfs-client/pom.xml | 4 ++-- hadoop-hdfs-project/hadoop-hdfs-httpfs/pom.xml | 4 ++-- hadoop-hdfs-project/hadoop-hdfs-native-client/pom.xml| 4 ++-- hadoop-hdfs-project/hadoop-hdfs-nfs/pom.xml | 4 ++-- hadoop-hdfs-project/hadoop-hdfs/pom.xml | 4 ++-- hadoop-hdfs-project/pom.xml | 4 ++-- .../hadoop-mapreduce-client/hadoop-mapreduce-client-app/pom.xml | 4 ++-- .../hadoop-mapreduce-client-common/pom.xml | 4 ++-- .../hadoop-mapreduce-client/hadoop-mapreduce-client-core/pom.xml | 4 ++-- .../hadoop-mapreduce-client-hs-plugins/pom.xml | 4 ++-- .../hadoop-mapreduce-client/hadoop-mapreduce-client-hs/pom.xml | 4 ++-- .../hadoop-mapreduce-client-jobclient/pom.xml| 4 ++-- .../hadoop-mapreduce-client-nativetask/pom.xml | 4 ++-- .../hadoop-mapreduce-client-shuffle/pom.xml | 4 ++-- .../hadoop-mapreduce-client-uploader/pom.xml | 4 ++-- hadoop-mapreduce-project/hadoop-mapreduce-client/pom.xml | 4 ++-- hadoop-mapreduce-project/hadoop-mapreduce-examples/pom.xml | 4 ++-- hadoop-mapreduce-project/pom.xml | 4 ++-- hadoop-maven-plugins/pom.xml | 2 +- hadoop-minicluster/pom.xml | 4 ++-- hadoop-project-dist/pom.xml | 4 ++-- hadoop-project/pom.xml | 4 ++-- hadoop-tools/hadoop-aliyun/pom.xml | 2 +- hadoop-tools/hadoop-archive-logs/pom.xml | 4 ++-- hadoop-tools/hadoop-archives/pom.xml | 4 ++-- hadoop-tools/hadoop-aws/pom.xml | 4 ++-- hadoop-tools/hadoop-azure-datalake/pom.xml | 2 +- hadoop-tools/hadoop-azure/pom.xml| 2 +- hadoop-tools/hadoop-datajoin/pom.xml | 4 ++-- hadoop-tools/hadoop-distcp/pom.xml | 4 ++-- hadoop-tools/hadoop-extras/pom.xml | 4 ++-- hadoop-tools/hadoop-fs2img/pom.xml | 4 ++-- hadoop-tools/hadoop-gridmix/pom.xml | 4 ++-- hadoop-tools/hadoop-kafka/pom.xml| 4 ++-- hadoop-tools/hadoop-openstack/pom.xml| 4 ++-- hadoop-tools/hadoop-pipes/pom.xml| 4 ++--
[1/2] hadoop git commit: Preparing for 3.1.0 development
Repository: hadoop Updated Branches: refs/heads/branch-3.1 49c747ab1 -> 0bd99621c http://git-wip-us.apache.org/repos/asf/hadoop/blob/0bd99621/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice-hbase/hadoop-yarn-server-timelineservice-hbase-server/hadoop-yarn-server-timelineservice-hbase-server-2/pom.xml -- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice-hbase/hadoop-yarn-server-timelineservice-hbase-server/hadoop-yarn-server-timelineservice-hbase-server-2/pom.xml b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice-hbase/hadoop-yarn-server-timelineservice-hbase-server/hadoop-yarn-server-timelineservice-hbase-server-2/pom.xml index 77a4add..430fc75 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice-hbase/hadoop-yarn-server-timelineservice-hbase-server/hadoop-yarn-server-timelineservice-hbase-server-2/pom.xml +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice-hbase/hadoop-yarn-server-timelineservice-hbase-server/hadoop-yarn-server-timelineservice-hbase-server-2/pom.xml @@ -22,13 +22,13 @@ hadoop-yarn-server-timelineservice-hbase-server org.apache.hadoop -3.2.0-SNAPSHOT +3.1.0-SNAPSHOT 4.0.0 hadoop-yarn-server-timelineservice-hbase-server-2 Apache Hadoop YARN TimelineService HBase Server 2.0 - 3.2.0-SNAPSHOT + 3.1.0-SNAPSHOT http://git-wip-us.apache.org/repos/asf/hadoop/blob/0bd99621/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice-hbase/hadoop-yarn-server-timelineservice-hbase-server/pom.xml -- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice-hbase/hadoop-yarn-server-timelineservice-hbase-server/pom.xml b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice-hbase/hadoop-yarn-server-timelineservice-hbase-server/pom.xml index dd2928c..ab48ff1 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice-hbase/hadoop-yarn-server-timelineservice-hbase-server/pom.xml +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice-hbase/hadoop-yarn-server-timelineservice-hbase-server/pom.xml @@ -22,12 +22,12 @@ hadoop-yarn-server-timelineservice-hbase org.apache.hadoop -3.2.0-SNAPSHOT +3.1.0-SNAPSHOT 4.0.0 hadoop-yarn-server-timelineservice-hbase-server - 3.2.0-SNAPSHOT + 3.1.0-SNAPSHOT Apache Hadoop YARN TimelineService HBase Servers pom http://git-wip-us.apache.org/repos/asf/hadoop/blob/0bd99621/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice-hbase/pom.xml -- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice-hbase/pom.xml b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice-hbase/pom.xml index 7e5a803..e0110ba 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice-hbase/pom.xml +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice-hbase/pom.xml @@ -22,12 +22,12 @@ hadoop-yarn-server org.apache.hadoop -3.2.0-SNAPSHOT +3.1.0-SNAPSHOT 4.0.0 hadoop-yarn-server-timelineservice-hbase - 3.2.0-SNAPSHOT + 3.1.0-SNAPSHOT Apache Hadoop YARN TimelineService HBase Backend pom http://git-wip-us.apache.org/repos/asf/hadoop/blob/0bd99621/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice/pom.xml -- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice/pom.xml b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice/pom.xml index dd47fd5..4236c51 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice/pom.xml +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice/pom.xml @@ -22,11 +22,11 @@ hadoop-yarn-server org.apache.hadoop -3.2.0-SNAPSHOT +3.1.0-SNAPSHOT 4.0.0 hadoop-yarn-server-timelineservice - 3.2.0-SNAPSHOT + 3.1.0-SNAPSHOT Apache Hadoop YARN Timeline Service http://git-wip-us.apache.org/repos/asf/hadoop/blob/0bd99621/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-web-proxy/pom.xml -- diff --git
hadoop git commit: YARN-7565. Addendum to fix an incompatible change. Contributed by Wangda Tan.
Repository: hadoop Updated Branches: refs/heads/trunk 17ba74be2 -> 95d4ec7fc YARN-7565. Addendum to fix an incompatible change. Contributed by Wangda Tan. Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/95d4ec7f Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/95d4ec7f Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/95d4ec7f Branch: refs/heads/trunk Commit: 95d4ec7fc07605d1ed6eabf066cd5413eb3fe465 Parents: 17ba74b Author: Vinod Kumar Vavilapalli (I am also known as @tshooter.)Authored: Thu Dec 14 17:07:17 2017 -0800 Committer: Vinod Kumar Vavilapalli (I am also known as @tshooter.) Committed: Thu Dec 14 17:10:37 2017 -0800 -- .../apache/hadoop/yarn/client/api/async/AMRMClientAsync.java | 2 -- .../yarn/client/api/async/impl/AMRMClientAsyncImpl.java | 7 +-- 2 files changed, 5 insertions(+), 4 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hadoop/blob/95d4ec7f/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/api/async/AMRMClientAsync.java -- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/api/async/AMRMClientAsync.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/api/async/AMRMClientAsync.java index a3d7959..2b82ad6 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/api/async/AMRMClientAsync.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/api/async/AMRMClientAsync.java @@ -539,7 +539,5 @@ extends AbstractService { * @param e */ void onError(Throwable e); - -void onContainersReceivedFromPreviousAttempts(List containers); } } http://git-wip-us.apache.org/repos/asf/hadoop/blob/95d4ec7f/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/api/async/impl/AMRMClientAsyncImpl.java -- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/api/async/impl/AMRMClientAsyncImpl.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/api/async/impl/AMRMClientAsyncImpl.java index cafb153..33b0aba 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/api/async/impl/AMRMClientAsyncImpl.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/api/async/impl/AMRMClientAsyncImpl.java @@ -360,8 +360,11 @@ extends AMRMClientAsync { } if (!response.getContainersFromPreviousAttempts().isEmpty()) { -handler.onContainersReceivedFromPreviousAttempts( -response.getContainersFromPreviousAttempts()); +if (handler instanceof AMRMClientAsync.AbstractCallbackHandler) { + ((AMRMClientAsync.AbstractCallbackHandler) handler) + .onContainersReceivedFromPreviousAttempts( + response.getContainersFromPreviousAttempts()); +} } progress = handler.getProgress(); } catch (Throwable ex) { - To unsubscribe, e-mail: common-commits-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-commits-h...@hadoop.apache.org
hadoop git commit: HADOOP-15059. Undoing the switch of Credentials to PB format as default - done via HADOOP-12563 for supporting 2.x to 3.x upgrades.
Repository: hadoop Updated Branches: refs/heads/branch-3.0.0 c27951e98 -> b1712f9c1 HADOOP-15059. Undoing the switch of Credentials to PB format as default - done via HADOOP-12563 for supporting 2.x to 3.x upgrades. (cherry picked from commit f19638333b11da6dcab9a964e73a49947b8390fd) Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/b1712f9c Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/b1712f9c Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/b1712f9c Branch: refs/heads/branch-3.0.0 Commit: b1712f9c1f8026d979df1c66e14bed87410e1546 Parents: c27951e Author: Vinod Kumar Vavilapalli (I am also known as @tshooter.)Authored: Fri Dec 8 08:00:21 2017 -0800 Committer: Vinod Kumar Vavilapalli (I am also known as @tshooter.) Committed: Fri Dec 8 08:03:58 2017 -0800 -- .../org/apache/hadoop/security/Credentials.java | 112 --- .../hadoop/security/token/DtFileOperations.java | 10 +- .../hadoop/security/token/TestDtUtilShell.java | 15 +-- .../hdfs/tools/DelegationTokenFetcher.java | 3 +- 4 files changed, 86 insertions(+), 54 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hadoop/blob/b1712f9c/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/Credentials.java -- diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/Credentials.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/Credentials.java index 4d58981..3e51249 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/Credentials.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/Credentials.java @@ -27,7 +27,6 @@ import java.io.DataOutput; import java.io.DataOutputStream; import java.io.File; import java.io.FileInputStream; -import java.io.FileOutputStream; import java.io.IOException; import java.nio.charset.StandardCharsets; import java.util.Arrays; @@ -60,6 +59,28 @@ import org.slf4j.LoggerFactory; @InterfaceAudience.Public @InterfaceStability.Evolving public class Credentials implements Writable { + public enum SerializedFormat { +WRITABLE((byte) 0x00), +PROTOBUF((byte) 0x01); + +// Caching to avoid reconstructing the array each time. +private static final SerializedFormat[] FORMATS = values(); + +final byte value; + +SerializedFormat(byte val) { + this.value = val; +} + +public static SerializedFormat valueOf(int val) { + try { +return FORMATS[val]; + } catch (ArrayIndexOutOfBoundsException e) { +throw new IllegalArgumentException("Unknown credential format: " + val); + } +} + } + private static final Logger LOG = LoggerFactory.getLogger(Credentials.class); private Map secretKeysMap = new HashMap (); @@ -224,63 +245,74 @@ public class Credentials implements Writable { if (!Arrays.equals(magic, TOKEN_STORAGE_MAGIC)) { throw new IOException("Bad header found in token storage."); } -byte version = in.readByte(); -if (version != TOKEN_STORAGE_VERSION && -version != OLD_TOKEN_STORAGE_VERSION) { - throw new IOException("Unknown version " + version + -" in token storage."); +SerializedFormat format; +try { + format = SerializedFormat.valueOf(in.readByte()); +} catch (IllegalArgumentException e) { + throw new IOException(e); } -if (version == OLD_TOKEN_STORAGE_VERSION) { +switch (format) { +case WRITABLE: readFields(in); -} else if (version == TOKEN_STORAGE_VERSION) { + break; +case PROTOBUF: readProto(in); + break; +default: + throw new IOException("Unsupported format " + format); } } private static final byte[] TOKEN_STORAGE_MAGIC = "HDTS".getBytes(StandardCharsets.UTF_8); - private static final byte TOKEN_STORAGE_VERSION = 1; - - /** - * For backward compatibility. - */ - private static final byte OLD_TOKEN_STORAGE_VERSION = 0; - public void writeTokenStorageToStream(DataOutputStream os) throws IOException { -os.write(TOKEN_STORAGE_MAGIC); -os.write(TOKEN_STORAGE_VERSION); -writeProto(os); +// by default store in the oldest supported format for compatibility +writeTokenStorageToStream(os, SerializedFormat.WRITABLE); } - public void writeTokenStorageFile(Path filename, -Configuration conf) throws IOException { -FSDataOutputStream os = filename.getFileSystem(conf).create(filename); -writeTokenStorageToStream(os); -os.close(); + public void
hadoop git commit: HADOOP-15059. Undoing the switch of Credentials to PB format as default - done via HADOOP-12563 for supporting 2.x to 3.x upgrades.
Repository: hadoop Updated Branches: refs/heads/branch-3.0 abaabb5de -> ba874d678 HADOOP-15059. Undoing the switch of Credentials to PB format as default - done via HADOOP-12563 for supporting 2.x to 3.x upgrades. (cherry picked from commit f19638333b11da6dcab9a964e73a49947b8390fd) Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/ba874d67 Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/ba874d67 Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/ba874d67 Branch: refs/heads/branch-3.0 Commit: ba874d6787508e7b718cd0e62ecb6d6d2c833b08 Parents: abaabb5 Author: Vinod Kumar Vavilapalli (I am also known as @tshooter.)Authored: Fri Dec 8 08:00:21 2017 -0800 Committer: Vinod Kumar Vavilapalli (I am also known as @tshooter.) Committed: Fri Dec 8 08:03:23 2017 -0800 -- .../org/apache/hadoop/security/Credentials.java | 112 --- .../hadoop/security/token/DtFileOperations.java | 10 +- .../hadoop/security/token/TestDtUtilShell.java | 15 +-- .../hdfs/tools/DelegationTokenFetcher.java | 3 +- 4 files changed, 86 insertions(+), 54 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hadoop/blob/ba874d67/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/Credentials.java -- diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/Credentials.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/Credentials.java index 4d58981..3e51249 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/Credentials.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/Credentials.java @@ -27,7 +27,6 @@ import java.io.DataOutput; import java.io.DataOutputStream; import java.io.File; import java.io.FileInputStream; -import java.io.FileOutputStream; import java.io.IOException; import java.nio.charset.StandardCharsets; import java.util.Arrays; @@ -60,6 +59,28 @@ import org.slf4j.LoggerFactory; @InterfaceAudience.Public @InterfaceStability.Evolving public class Credentials implements Writable { + public enum SerializedFormat { +WRITABLE((byte) 0x00), +PROTOBUF((byte) 0x01); + +// Caching to avoid reconstructing the array each time. +private static final SerializedFormat[] FORMATS = values(); + +final byte value; + +SerializedFormat(byte val) { + this.value = val; +} + +public static SerializedFormat valueOf(int val) { + try { +return FORMATS[val]; + } catch (ArrayIndexOutOfBoundsException e) { +throw new IllegalArgumentException("Unknown credential format: " + val); + } +} + } + private static final Logger LOG = LoggerFactory.getLogger(Credentials.class); private Map secretKeysMap = new HashMap (); @@ -224,63 +245,74 @@ public class Credentials implements Writable { if (!Arrays.equals(magic, TOKEN_STORAGE_MAGIC)) { throw new IOException("Bad header found in token storage."); } -byte version = in.readByte(); -if (version != TOKEN_STORAGE_VERSION && -version != OLD_TOKEN_STORAGE_VERSION) { - throw new IOException("Unknown version " + version + -" in token storage."); +SerializedFormat format; +try { + format = SerializedFormat.valueOf(in.readByte()); +} catch (IllegalArgumentException e) { + throw new IOException(e); } -if (version == OLD_TOKEN_STORAGE_VERSION) { +switch (format) { +case WRITABLE: readFields(in); -} else if (version == TOKEN_STORAGE_VERSION) { + break; +case PROTOBUF: readProto(in); + break; +default: + throw new IOException("Unsupported format " + format); } } private static final byte[] TOKEN_STORAGE_MAGIC = "HDTS".getBytes(StandardCharsets.UTF_8); - private static final byte TOKEN_STORAGE_VERSION = 1; - - /** - * For backward compatibility. - */ - private static final byte OLD_TOKEN_STORAGE_VERSION = 0; - public void writeTokenStorageToStream(DataOutputStream os) throws IOException { -os.write(TOKEN_STORAGE_MAGIC); -os.write(TOKEN_STORAGE_VERSION); -writeProto(os); +// by default store in the oldest supported format for compatibility +writeTokenStorageToStream(os, SerializedFormat.WRITABLE); } - public void writeTokenStorageFile(Path filename, -Configuration conf) throws IOException { -FSDataOutputStream os = filename.getFileSystem(conf).create(filename); -writeTokenStorageToStream(os); -os.close(); + public void
hadoop git commit: HADOOP-15059. Undoing the switch of Credentials to PB format as default - done via HADOOP-12563 for supporting 2.x to 3.x upgrades.
Repository: hadoop Updated Branches: refs/heads/trunk ce04340ec -> f19638333 HADOOP-15059. Undoing the switch of Credentials to PB format as default - done via HADOOP-12563 for supporting 2.x to 3.x upgrades. Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/f1963833 Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/f1963833 Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/f1963833 Branch: refs/heads/trunk Commit: f19638333b11da6dcab9a964e73a49947b8390fd Parents: ce04340 Author: Vinod Kumar Vavilapalli (I am also known as @tshooter.)Authored: Fri Dec 8 08:00:21 2017 -0800 Committer: Vinod Kumar Vavilapalli (I am also known as @tshooter.) Committed: Fri Dec 8 08:00:21 2017 -0800 -- .../org/apache/hadoop/security/Credentials.java | 112 --- .../hadoop/security/token/DtFileOperations.java | 10 +- .../hadoop/security/token/TestDtUtilShell.java | 15 +-- .../hdfs/tools/DelegationTokenFetcher.java | 3 +- 4 files changed, 86 insertions(+), 54 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hadoop/blob/f1963833/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/Credentials.java -- diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/Credentials.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/Credentials.java index 4d58981..3e51249 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/Credentials.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/Credentials.java @@ -27,7 +27,6 @@ import java.io.DataOutput; import java.io.DataOutputStream; import java.io.File; import java.io.FileInputStream; -import java.io.FileOutputStream; import java.io.IOException; import java.nio.charset.StandardCharsets; import java.util.Arrays; @@ -60,6 +59,28 @@ import org.slf4j.LoggerFactory; @InterfaceAudience.Public @InterfaceStability.Evolving public class Credentials implements Writable { + public enum SerializedFormat { +WRITABLE((byte) 0x00), +PROTOBUF((byte) 0x01); + +// Caching to avoid reconstructing the array each time. +private static final SerializedFormat[] FORMATS = values(); + +final byte value; + +SerializedFormat(byte val) { + this.value = val; +} + +public static SerializedFormat valueOf(int val) { + try { +return FORMATS[val]; + } catch (ArrayIndexOutOfBoundsException e) { +throw new IllegalArgumentException("Unknown credential format: " + val); + } +} + } + private static final Logger LOG = LoggerFactory.getLogger(Credentials.class); private Map secretKeysMap = new HashMap (); @@ -224,63 +245,74 @@ public class Credentials implements Writable { if (!Arrays.equals(magic, TOKEN_STORAGE_MAGIC)) { throw new IOException("Bad header found in token storage."); } -byte version = in.readByte(); -if (version != TOKEN_STORAGE_VERSION && -version != OLD_TOKEN_STORAGE_VERSION) { - throw new IOException("Unknown version " + version + -" in token storage."); +SerializedFormat format; +try { + format = SerializedFormat.valueOf(in.readByte()); +} catch (IllegalArgumentException e) { + throw new IOException(e); } -if (version == OLD_TOKEN_STORAGE_VERSION) { +switch (format) { +case WRITABLE: readFields(in); -} else if (version == TOKEN_STORAGE_VERSION) { + break; +case PROTOBUF: readProto(in); + break; +default: + throw new IOException("Unsupported format " + format); } } private static final byte[] TOKEN_STORAGE_MAGIC = "HDTS".getBytes(StandardCharsets.UTF_8); - private static final byte TOKEN_STORAGE_VERSION = 1; - - /** - * For backward compatibility. - */ - private static final byte OLD_TOKEN_STORAGE_VERSION = 0; - public void writeTokenStorageToStream(DataOutputStream os) throws IOException { -os.write(TOKEN_STORAGE_MAGIC); -os.write(TOKEN_STORAGE_VERSION); -writeProto(os); +// by default store in the oldest supported format for compatibility +writeTokenStorageToStream(os, SerializedFormat.WRITABLE); } - public void writeTokenStorageFile(Path filename, -Configuration conf) throws IOException { -FSDataOutputStream os = filename.getFileSystem(conf).create(filename); -writeTokenStorageToStream(os); -os.close(); + public void writeTokenStorageToStream(DataOutputStream os, + SerializedFormat format) throws
svn commit: r1802426 [2/2] - in /hadoop/common/site/main: author/src/documentation/content/xdocs/ publish/ publish/docs/r2.8.1/ publish/docs/r2.8.1/api/ publish/docs/r2.8.1/api/org/ publish/docs/r2.8.
Modified: hadoop/common/site/main/author/src/documentation/content/xdocs/index.xml URL: http://svn.apache.org/viewvc/hadoop/common/site/main/author/src/documentation/content/xdocs/index.xml?rev=1802426=1802425=1802426=diff == --- hadoop/common/site/main/author/src/documentation/content/xdocs/index.xml (original) +++ hadoop/common/site/main/author/src/documentation/content/xdocs/index.xml Wed Jul 19 20:02:27 2017 @@ -152,7 +152,18 @@ -26 May 2017: Release 3.0.0-alpha3 available +08 June, 2017: Release 2.8.1 available + + This is a security release in the 2.8.0 release line. It consists of 2.8.0 plus security fixes. Users on 2.8.0 are encouraged to upgrade to 2.8.1. + + + Please note that 2.8.x release line continues to be not yet ready for production use. Critical issues are being ironed out via testing and downstream + adoption. Production users should wait for a subsequent release in the 2.8.x line. + + + + +26 May, 2017: Release 3.0.0-alpha3 available This is a security release in the 3.0.0 release line. It consists of alpha2 plus security fixes, along with necessary build-related fixes. Users on 3.0.0-alpha1 and 3.0.0-alpha2 are encouraged to upgrade to 3.0.0-alpha3. Modified: hadoop/common/site/main/author/src/documentation/content/xdocs/releases.xml URL: http://svn.apache.org/viewvc/hadoop/common/site/main/author/src/documentation/content/xdocs/releases.xml?rev=1802426=1802425=1802426=diff == --- hadoop/common/site/main/author/src/documentation/content/xdocs/releases.xml (original) +++ hadoop/common/site/main/author/src/documentation/content/xdocs/releases.xml Wed Jul 19 20:02:27 2017 @@ -44,18 +44,18 @@ https://dist.apache.org/repos/dist/release/hadoop/common/hadoop-3.0.0-alpha4/hadoop-3.0.0-alpha4.tar.gz.mds;>checksum file - 2.8.0 - 22 March, 2017 - http://www.apache.org/dyn/closer.cgi/hadoop/common/hadoop-2.8.0/hadoop-2.8.0-src.tar.gz;>source - https://dist.apache.org/repos/dist/release/hadoop/common/hadoop-2.8.0/hadoop-2.8.0-src.tar.gz.asc;>signature - https://dist.apache.org/repos/dist/release/hadoop/common/hadoop-2.8.0/hadoop-2.8.0-src.tar.gz.mds;>77B6A9A5 F1324A00.. + 2.8.1 + 08 June, 2017 + http://www.apache.org/dyn/closer.cgi/hadoop/common/hadoop-2.8.1/hadoop-2.8.1-src.tar.gz;>source + https://dist.apache.org/repos/dist/release/hadoop/common/hadoop-2.8.1/hadoop-2.8.1-src.tar.gz.asc;>signature + https://dist.apache.org/repos/dist/release/hadoop/common/hadoop-2.8.1/hadoop-2.8.1-src.tar.gz.mds;>0748C0E2 519382F2.. - http://www.apache.org/dyn/closer.cgi/hadoop/common/hadoop-2.8.0/hadoop-2.8.0.tar.gz;>binary - https://dist.apache.org/repos/dist/release/hadoop/common/hadoop-2.8.0/hadoop-2.8.0.tar.gz.asc;>signature - https://dist.apache.org/repos/dist/release/hadoop/common/hadoop-2.8.0/hadoop-2.8.0.tar.gz.mds;>3C0C6053 651970C3.. + http://www.apache.org/dyn/closer.cgi/hadoop/common/hadoop-2.8.1/hadoop-2.8.1.tar.gz;>binary + https://dist.apache.org/repos/dist/release/hadoop/common/hadoop-2.8.1/hadoop-2.8.1.tar.gz.asc;>signature + https://dist.apache.org/repos/dist/release/hadoop/common/hadoop-2.8.1/hadoop-2.8.1.tar.gz.mds;>B5BE5275 78EF2C85.. 2.7.3 @@ -142,6 +142,18 @@ The alpha4 http://hadoop.apache.org/docs/r3.0.0-alpha4/hadoop-project-dist/hadoop-common/release/3.0.0-alpha4/RELEASENOTES.3.0.0-alpha4.html;>release notes and http://hadoop.apache.org/docs/r3.0.0-alpha4/hadoop-project-dist/hadoop-common/release/3.0.0-alpha4/CHANGES.3.0.0-alpha4.html;>changelog detail the changes since 3.0.0-alpha3. + + +08 June, 2017: Release 2.8.1 available + + This is a security release in the 2.8.0 release line. It consists of 2.8.0 plus security fixes. Users on 2.8.0 are encouraged to upgrade to 2.8.1. + + + Please note that 2.8.x release line continues to be not yet ready for production use. Critical issues are being ironed out via testing and downstream + adoption. Production users should wait for a subsequent release in the 2.8.x line. + + + 26 May 2017: Release 3.0.0-alpha3 available Modified: hadoop/common/site/main/author/src/documentation/content/xdocs/site.xml URL: http://svn.apache.org/viewvc/hadoop/common/site/main/author/src/documentation/content/xdocs/site.xml?rev=1802426=1802425=1802426=diff == ---
svn commit: r1802426 [1/2] - in /hadoop/common/site/main: author/src/documentation/content/xdocs/ publish/ publish/docs/r2.8.1/ publish/docs/r2.8.1/api/ publish/docs/r2.8.1/api/org/ publish/docs/r2.8.
Author: vinodkv Date: Wed Jul 19 20:02:27 2017 New Revision: 1802426 URL: http://svn.apache.org/viewvc?rev=1802426=rev Log: Updating site for release 2.8.1. Added: hadoop/common/site/main/publish/docs/r2.8.1/ hadoop/common/site/main/publish/docs/r2.8.1/api/ hadoop/common/site/main/publish/docs/r2.8.1/api/org/ hadoop/common/site/main/publish/docs/r2.8.1/api/org/apache/ hadoop/common/site/main/publish/docs/r2.8.1/api/org/apache/hadoop/ hadoop/common/site/main/publish/docs/r2.8.1/api/org/apache/hadoop/ant/ hadoop/common/site/main/publish/docs/r2.8.1/api/org/apache/hadoop/ant/condition/ hadoop/common/site/main/publish/docs/r2.8.1/api/org/apache/hadoop/class-use/ hadoop/common/site/main/publish/docs/r2.8.1/api/org/apache/hadoop/classification/ hadoop/common/site/main/publish/docs/r2.8.1/api/org/apache/hadoop/classification/class-use/ hadoop/common/site/main/publish/docs/r2.8.1/api/org/apache/hadoop/conf/ hadoop/common/site/main/publish/docs/r2.8.1/api/org/apache/hadoop/conf/class-use/ hadoop/common/site/main/publish/docs/r2.8.1/api/org/apache/hadoop/contrib/ hadoop/common/site/main/publish/docs/r2.8.1/api/org/apache/hadoop/contrib/bkjournal/ hadoop/common/site/main/publish/docs/r2.8.1/api/org/apache/hadoop/contrib/utils/ hadoop/common/site/main/publish/docs/r2.8.1/api/org/apache/hadoop/contrib/utils/join/ hadoop/common/site/main/publish/docs/r2.8.1/api/org/apache/hadoop/crypto/ hadoop/common/site/main/publish/docs/r2.8.1/api/org/apache/hadoop/crypto/key/ hadoop/common/site/main/publish/docs/r2.8.1/api/org/apache/hadoop/crypto/key/class-use/ hadoop/common/site/main/publish/docs/r2.8.1/api/org/apache/hadoop/crypto/key/kms/ hadoop/common/site/main/publish/docs/r2.8.1/api/org/apache/hadoop/crypto/key/kms/server/ hadoop/common/site/main/publish/docs/r2.8.1/api/org/apache/hadoop/crypto/random/ hadoop/common/site/main/publish/docs/r2.8.1/api/org/apache/hadoop/examples/ hadoop/common/site/main/publish/docs/r2.8.1/api/org/apache/hadoop/examples/dancing/ hadoop/common/site/main/publish/docs/r2.8.1/api/org/apache/hadoop/examples/pi/ hadoop/common/site/main/publish/docs/r2.8.1/api/org/apache/hadoop/examples/pi/math/ hadoop/common/site/main/publish/docs/r2.8.1/api/org/apache/hadoop/examples/terasort/ hadoop/common/site/main/publish/docs/r2.8.1/api/org/apache/hadoop/filecache/ hadoop/common/site/main/publish/docs/r2.8.1/api/org/apache/hadoop/filecache/class-use/ hadoop/common/site/main/publish/docs/r2.8.1/api/org/apache/hadoop/fs/ hadoop/common/site/main/publish/docs/r2.8.1/api/org/apache/hadoop/fs/adl/ hadoop/common/site/main/publish/docs/r2.8.1/api/org/apache/hadoop/fs/adl/class-use/ hadoop/common/site/main/publish/docs/r2.8.1/api/org/apache/hadoop/fs/adl/oauth2/ hadoop/common/site/main/publish/docs/r2.8.1/api/org/apache/hadoop/fs/adl/oauth2/class-use/ hadoop/common/site/main/publish/docs/r2.8.1/api/org/apache/hadoop/fs/azure/ hadoop/common/site/main/publish/docs/r2.8.1/api/org/apache/hadoop/fs/azure/class-use/ hadoop/common/site/main/publish/docs/r2.8.1/api/org/apache/hadoop/fs/azure/metrics/ hadoop/common/site/main/publish/docs/r2.8.1/api/org/apache/hadoop/fs/azure/metrics/class-use/ hadoop/common/site/main/publish/docs/r2.8.1/api/org/apache/hadoop/fs/class-use/ hadoop/common/site/main/publish/docs/r2.8.1/api/org/apache/hadoop/fs/crypto/ hadoop/common/site/main/publish/docs/r2.8.1/api/org/apache/hadoop/fs/ftp/ hadoop/common/site/main/publish/docs/r2.8.1/api/org/apache/hadoop/fs/ftp/class-use/ hadoop/common/site/main/publish/docs/r2.8.1/api/org/apache/hadoop/fs/http/ hadoop/common/site/main/publish/docs/r2.8.1/api/org/apache/hadoop/fs/http/client/ hadoop/common/site/main/publish/docs/r2.8.1/api/org/apache/hadoop/fs/http/server/ hadoop/common/site/main/publish/docs/r2.8.1/api/org/apache/hadoop/fs/permission/ hadoop/common/site/main/publish/docs/r2.8.1/api/org/apache/hadoop/fs/permission/class-use/ hadoop/common/site/main/publish/docs/r2.8.1/api/org/apache/hadoop/fs/s3/ hadoop/common/site/main/publish/docs/r2.8.1/api/org/apache/hadoop/fs/s3/class-use/ hadoop/common/site/main/publish/docs/r2.8.1/api/org/apache/hadoop/fs/s3a/ hadoop/common/site/main/publish/docs/r2.8.1/api/org/apache/hadoop/fs/s3a/class-use/ hadoop/common/site/main/publish/docs/r2.8.1/api/org/apache/hadoop/fs/s3native/ hadoop/common/site/main/publish/docs/r2.8.1/api/org/apache/hadoop/fs/s3native/class-use/ hadoop/common/site/main/publish/docs/r2.8.1/api/org/apache/hadoop/fs/sftp/ hadoop/common/site/main/publish/docs/r2.8.1/api/org/apache/hadoop/fs/shell/ hadoop/common/site/main/publish/docs/r2.8.1/api/org/apache/hadoop/fs/shell/find/ hadoop/common/site/main/publish/docs/r2.8.1/api/org/apache/hadoop/fs/swift/ hadoop/common/site/main/publish/docs/r2.8.1/api/org/apache/hadoop/fs/swift/auth
hadoop git commit: Preparing for 2.8.3 development
Repository: hadoop Updated Branches: refs/heads/branch-2.8 c60516aee -> 396de0fd8 Preparing for 2.8.3 development Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/396de0fd Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/396de0fd Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/396de0fd Branch: refs/heads/branch-2.8 Commit: 396de0fd8ff4311b859fbffa84ef771589c0751b Parents: c60516a Author: Vinod Kumar Vavilapalli (I am also known as @tshooter.)Authored: Wed Jun 7 15:38:20 2017 -0700 Committer: Vinod Kumar Vavilapalli (I am also known as @tshooter.) Committed: Wed Jun 7 15:38:20 2017 -0700 -- hadoop-assemblies/pom.xml| 4 ++-- hadoop-build-tools/pom.xml | 2 +- hadoop-client/pom.xml| 4 ++-- hadoop-common-project/hadoop-annotations/pom.xml | 4 ++-- hadoop-common-project/hadoop-auth-examples/pom.xml | 4 ++-- hadoop-common-project/hadoop-auth/pom.xml| 4 ++-- hadoop-common-project/hadoop-common/pom.xml | 4 ++-- hadoop-common-project/hadoop-kms/pom.xml | 4 ++-- hadoop-common-project/hadoop-minikdc/pom.xml | 4 ++-- hadoop-common-project/hadoop-nfs/pom.xml | 4 ++-- hadoop-common-project/pom.xml| 4 ++-- hadoop-dist/pom.xml | 4 ++-- hadoop-hdfs-project/hadoop-hdfs-client/pom.xml | 4 ++-- hadoop-hdfs-project/hadoop-hdfs-httpfs/pom.xml | 4 ++-- hadoop-hdfs-project/hadoop-hdfs-native-client/pom.xml| 4 ++-- hadoop-hdfs-project/hadoop-hdfs-nfs/pom.xml | 4 ++-- hadoop-hdfs-project/hadoop-hdfs/pom.xml | 4 ++-- hadoop-hdfs-project/hadoop-hdfs/src/contrib/bkjournal/pom.xml| 4 ++-- hadoop-hdfs-project/pom.xml | 4 ++-- .../hadoop-mapreduce-client/hadoop-mapreduce-client-app/pom.xml | 4 ++-- .../hadoop-mapreduce-client-common/pom.xml | 4 ++-- .../hadoop-mapreduce-client/hadoop-mapreduce-client-core/pom.xml | 4 ++-- .../hadoop-mapreduce-client-hs-plugins/pom.xml | 4 ++-- .../hadoop-mapreduce-client/hadoop-mapreduce-client-hs/pom.xml | 4 ++-- .../hadoop-mapreduce-client-jobclient/pom.xml| 4 ++-- .../hadoop-mapreduce-client-shuffle/pom.xml | 4 ++-- hadoop-mapreduce-project/hadoop-mapreduce-client/pom.xml | 4 ++-- hadoop-mapreduce-project/hadoop-mapreduce-examples/pom.xml | 4 ++-- hadoop-mapreduce-project/pom.xml | 4 ++-- hadoop-maven-plugins/pom.xml | 2 +- hadoop-minicluster/pom.xml | 4 ++-- hadoop-project-dist/pom.xml | 4 ++-- hadoop-project/pom.xml | 4 ++-- hadoop-tools/hadoop-ant/pom.xml | 4 ++-- hadoop-tools/hadoop-archive-logs/pom.xml | 4 ++-- hadoop-tools/hadoop-archives/pom.xml | 4 ++-- hadoop-tools/hadoop-aws/pom.xml | 4 ++-- hadoop-tools/hadoop-azure-datalake/pom.xml | 2 +- hadoop-tools/hadoop-azure/pom.xml| 2 +- hadoop-tools/hadoop-datajoin/pom.xml | 4 ++-- hadoop-tools/hadoop-distcp/pom.xml | 4 ++-- hadoop-tools/hadoop-extras/pom.xml | 4 ++-- hadoop-tools/hadoop-gridmix/pom.xml | 4 ++-- hadoop-tools/hadoop-openstack/pom.xml| 4 ++-- hadoop-tools/hadoop-pipes/pom.xml| 4 ++-- hadoop-tools/hadoop-rumen/pom.xml| 4 ++-- hadoop-tools/hadoop-sls/pom.xml | 4 ++-- hadoop-tools/hadoop-streaming/pom.xml| 4 ++-- hadoop-tools/hadoop-tools-dist/pom.xml | 4 ++-- hadoop-tools/pom.xml | 4 ++-- hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/pom.xml | 4 ++-- .../hadoop-yarn-applications-distributedshell/pom.xml| 4 ++-- .../hadoop-yarn-applications-unmanaged-am-launcher/pom.xml | 4 ++-- hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/pom.xml | 4 ++-- hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/pom.xml | 4 ++-- hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/pom.xml | 4 ++--
[hadoop] Git Push Summary
Repository: hadoop Updated Tags: refs/tags/rel/release-2.8.1 [created] f12e08d9d - To unsubscribe, e-mail: common-commits-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-commits-h...@hadoop.apache.org
[2/8] hadoop git commit: Preparing for 2.8.2 development: mvn versions:set -DnewVersion=2.8.2
http://git-wip-us.apache.org/repos/asf/hadoop/blob/c64fe9f0/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/pom.xml.versionsBackup -- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/pom.xml.versionsBackup b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/pom.xml.versionsBackup deleted file mode 100644 index f358741..000 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/pom.xml.versionsBackup +++ /dev/null @@ -1,147 +0,0 @@ - - -http://maven.apache.org/POM/4.0.0; - xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance; - xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 - http://maven.apache.org/xsd/maven-4.0.0.xsd;> - -hadoop-yarn -org.apache.hadoop -2.8.0 - - 4.0.0 - org.apache.hadoop - hadoop-yarn-api - 2.8.0 - Apache Hadoop YARN API - - - -${project.parent.basedir} -true -../dev-support - - - - - commons-lang - commons-lang - - - com.google.guava - guava - - - commons-logging - commons-logging - - - - - org.apache.hadoop - hadoop-common - provided - - - - - org.apache.hadoop - hadoop-annotations - - - com.google.protobuf - protobuf-java - - - - org.apache.hadoop - hadoop-common - test-jar - test - - - - junit - junit - test - - - - - - - ${basedir}/../hadoop-yarn-common/src/main/resources - - yarn-default.xml - -false - - - - -org.apache.hadoop -hadoop-maven-plugins - - -compile-protoc -generate-sources - - protoc - - - ${protobuf.version} - ${protoc.path} - - ${basedir}/../../../hadoop-common-project/hadoop-common/src/main/proto -${basedir}/src/main/proto -${basedir}/src/main/proto/server - - -${basedir}/src/main/proto - - yarn_protos.proto - yarn_service_protos.proto - applicationmaster_protocol.proto - applicationclient_protocol.proto - containermanagement_protocol.proto - server/yarn_server_resourcemanager_service_protos.proto - server/resourcemanager_administration_protocol.proto - application_history_client.proto - server/application_history_server.proto - client_SCM_protocol.proto - server/SCM_Admin_protocol.proto - - - ${project.build.directory}/generated-sources/java - - - - - - -maven-jar-plugin - - - - test-jar - -test-compile - - - - - - - http://git-wip-us.apache.org/repos/asf/hadoop/blob/c64fe9f0/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-applications-distributedshell/pom.xml -- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-applications-distributedshell/pom.xml b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-applications-distributedshell/pom.xml index b4dbcb8..f364d387 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-applications-distributedshell/pom.xml +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-applications-distributedshell/pom.xml @@ -19,12 +19,12 @@ hadoop-yarn-applications org.apache.hadoop -2.8.1 +2.8.2-SNAPSHOT 4.0.0 org.apache.hadoop hadoop-yarn-applications-distributedshell - 2.8.1 + 2.8.2-SNAPSHOT Apache Hadoop YARN DistributedShell http://git-wip-us.apache.org/repos/asf/hadoop/blob/c64fe9f0/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-applications-distributedshell/pom.xml.versionsBackup -- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-applications-distributedshell/pom.xml.versionsBackup b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-applications-distributedshell/pom.xml.versionsBackup deleted file mode 100644 index 9349b95..000 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-applications-distributedshell/pom.xml.versionsBackup +++ /dev/null @@ -1,188 +0,0 @@ - - -http://maven.apache.org/POM/4.0.0; - xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance; -
[3/8] hadoop git commit: Preparing for 2.8.2 development: mvn versions:set -DnewVersion=2.8.2
http://git-wip-us.apache.org/repos/asf/hadoop/blob/c64fe9f0/hadoop-tools/hadoop-azure-datalake/pom.xml.versionsBackup -- diff --git a/hadoop-tools/hadoop-azure-datalake/pom.xml.versionsBackup b/hadoop-tools/hadoop-azure-datalake/pom.xml.versionsBackup deleted file mode 100644 index a8c3b16..000 --- a/hadoop-tools/hadoop-azure-datalake/pom.xml.versionsBackup +++ /dev/null @@ -1,160 +0,0 @@ - - -http://maven.apache.org/POM/4.0.0; - xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance; - xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 http://maven.apache.org/xsd/maven-4.0.0.xsd;> - 4.0.0 - -org.apache.hadoop -hadoop-project -2.8.0 -../../hadoop-project - - org.apache.hadoop - hadoop-azure-datalake - Apache Hadoop Azure Data Lake support - -This module contains code to support integration with Azure Data Lake. - - jar - -2.4.0 -0.9.1 -UTF-8 -true - - - - -org.apache.maven.plugins -maven-project-info-reports-plugin - - - false - false - - - - -org.apache.maven.plugins -maven-jar-plugin - - - - test-jar - - - - - -org.apache.maven.plugins -maven-dependency-plugin - - -deplist -compile - - list - - - - ${project.basedir}/target/hadoop-tools-deps/${project.artifactId}.tools-optional.txt - - - - - - - - - - - - org.eclipse.m2e - lifecycle-mapping - 1.0.0 - - - - - -org.apache.maven.plugins - -maven-enforcer-plugin - -[1.0.0,) - - enforce - - - - - - - - - - - - - - - - - - - com.microsoft.azure - azure-data-lake-store-sdk - 2.1.4 - - - -org.apache.hadoop - hadoop-common - - - com.squareup.okhttp - okhttp - 2.4.0 - - - junit - junit - test - - - com.eclipsesource.minimal-json - minimal-json - 0.9.1 - test - - - org.apache.hadoop - hadoop-common - test - test-jar - - - com.squareup.okhttp - mockwebserver - 2.4.0 - test - - - http://git-wip-us.apache.org/repos/asf/hadoop/blob/c64fe9f0/hadoop-tools/hadoop-azure/pom.xml -- diff --git a/hadoop-tools/hadoop-azure/pom.xml b/hadoop-tools/hadoop-azure/pom.xml index fc6d8ba..f5f36e84 100644 --- a/hadoop-tools/hadoop-azure/pom.xml +++ b/hadoop-tools/hadoop-azure/pom.xml @@ -19,7 +19,7 @@ org.apache.hadoop hadoop-project -2.8.1 +2.8.2-SNAPSHOT ../../hadoop-project hadoop-azure http://git-wip-us.apache.org/repos/asf/hadoop/blob/c64fe9f0/hadoop-tools/hadoop-azure/pom.xml.versionsBackup -- diff --git a/hadoop-tools/hadoop-azure/pom.xml.versionsBackup b/hadoop-tools/hadoop-azure/pom.xml.versionsBackup deleted file mode 100644 index 5076c93..000 --- a/hadoop-tools/hadoop-azure/pom.xml.versionsBackup +++ /dev/null @@ -1,204 +0,0 @@ - - -http://maven.apache.org/POM/4.0.0; - xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance; - xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 http://maven.apache.org/xsd/maven-4.0.0.xsd;> - 4.0.0 - -org.apache.hadoop -hadoop-project -2.8.0 -../../hadoop-project - - hadoop-azure - Apache Hadoop Azure support - -This module contains code to support integration with Azure. -Currently this consists of a filesystem client to read data from -and write data to Azure Storage. - - jar - - -UTF-8 -true - - - - - -org.codehaus.mojo -findbugs-maven-plugin - - true - true - ${basedir}/dev-support/findbugs-exclude.xml - - Max - - - -org.apache.maven.plugins -maven-project-info-reports-plugin - - - false - false - - - - - org.apache.maven.plugins - maven-checkstyle-plugin - - -src/config/checkstyle.xml - - - - - -org.apache.maven.plugins -maven-jar-plugin - -
[8/8] hadoop git commit: Preparing for 2.8.2 development: mvn versions:set -DnewVersion=2.8.2
Preparing for 2.8.2 development: mvn versions:set -DnewVersion=2.8.2 Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/c64fe9f0 Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/c64fe9f0 Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/c64fe9f0 Branch: refs/heads/branch-2.8.2 Commit: c64fe9f07be9dcb01fabdd9b93f1fd3247f86f96 Parents: c24e51a Author: Vinod Kumar Vavilapalli (I am also known as @tshooter.)Authored: Wed Jun 7 14:54:39 2017 -0700 Committer: Vinod Kumar Vavilapalli (I am also known as @tshooter.) Committed: Wed Jun 7 14:54:39 2017 -0700 -- hadoop-assemblies/pom.xml |4 +- hadoop-assemblies/pom.xml.versionsBackup| 85 -- hadoop-build-tools/pom.xml |2 +- hadoop-build-tools/pom.xml.versionsBackup | 114 -- hadoop-client/pom.xml |4 +- hadoop-client/pom.xml.versionsBackup| 312 .../hadoop-annotations/pom.xml |4 +- .../hadoop-annotations/pom.xml.versionsBackup | 74 - .../hadoop-auth-examples/pom.xml|4 +- .../hadoop-auth-examples/pom.xml.versionsBackup | 96 -- hadoop-common-project/hadoop-auth/pom.xml |4 +- .../hadoop-auth/pom.xml.versionsBackup | 281 hadoop-common-project/hadoop-common/pom.xml |4 +- .../hadoop-common/pom.xml.versionsBackup| 874 --- hadoop-common-project/hadoop-kms/pom.xml|4 +- .../hadoop-kms/pom.xml.versionsBackup | 442 -- hadoop-common-project/hadoop-minikdc/pom.xml|4 +- .../hadoop-minikdc/pom.xml.versionsBackup | 155 -- hadoop-common-project/hadoop-nfs/pom.xml|4 +- .../hadoop-nfs/pom.xml.versionsBackup | 152 -- hadoop-common-project/pom.xml |4 +- hadoop-common-project/pom.xml.versionsBackup| 60 - hadoop-dist/pom.xml |4 +- hadoop-dist/pom.xml.versionsBackup | 199 --- hadoop-hdfs-project/hadoop-hdfs-client/pom.xml |4 +- .../hadoop-hdfs-client/pom.xml.versionsBackup | 171 --- hadoop-hdfs-project/hadoop-hdfs-httpfs/pom.xml |4 +- .../hadoop-hdfs-httpfs/pom.xml.versionsBackup | 571 --- .../hadoop-hdfs-native-client/pom.xml |4 +- .../pom.xml.versionsBackup | 232 --- hadoop-hdfs-project/hadoop-hdfs-nfs/pom.xml |4 +- .../hadoop-hdfs-nfs/pom.xml.versionsBackup | 233 --- hadoop-hdfs-project/hadoop-hdfs/pom.xml |4 +- .../hadoop-hdfs/pom.xml.versionsBackup | 479 -- .../hadoop-hdfs/src/contrib/bkjournal/pom.xml |4 +- .../contrib/bkjournal/pom.xml.versionsBackup| 175 --- hadoop-hdfs-project/pom.xml |4 +- hadoop-hdfs-project/pom.xml.versionsBackup | 59 - .../hadoop-mapreduce-client-app/pom.xml |4 +- .../pom.xml.versionsBackup | 141 -- .../hadoop-mapreduce-client-common/pom.xml |4 +- .../pom.xml.versionsBackup | 92 -- .../hadoop-mapreduce-client-core/pom.xml|4 +- .../pom.xml.versionsBackup | 119 -- .../hadoop-mapreduce-client-hs-plugins/pom.xml |4 +- .../pom.xml.versionsBackup | 46 - .../hadoop-mapreduce-client-hs/pom.xml |4 +- .../pom.xml.versionsBackup | 87 -- .../hadoop-mapreduce-client-jobclient/pom.xml |4 +- .../pom.xml.versionsBackup | 182 --- .../hadoop-mapreduce-client-shuffle/pom.xml |4 +- .../pom.xml.versionsBackup | 86 -- .../hadoop-mapreduce-client/pom.xml |4 +- .../pom.xml.versionsBackup | 318 .../hadoop-mapreduce-examples/pom.xml |4 +- .../pom.xml.versionsBackup | 154 -- hadoop-mapreduce-project/pom.xml|4 +- hadoop-mapreduce-project/pom.xml.versionsBackup | 286 hadoop-maven-plugins/pom.xml|2 +- hadoop-maven-plugins/pom.xml.versionsBackup | 85 -- hadoop-minicluster/pom.xml |4 +- hadoop-minicluster/pom.xml.versionsBackup | 105 -- hadoop-project-dist/pom.xml |4 +- hadoop-project-dist/pom.xml.versionsBackup | 425 -- hadoop-project/pom.xml |4 +- hadoop-project/pom.xml.versionsBackup | 1415 -- hadoop-tools/hadoop-ant/pom.xml |4 +- hadoop-tools/hadoop-ant/pom.xml.versionsBackup | 56 - hadoop-tools/hadoop-archive-logs/pom.xml|4 +- .../hadoop-archive-logs/pom.xml.versionsBackup | 189 --- hadoop-tools/hadoop-archives/pom.xml|4
[4/8] hadoop git commit: Preparing for 2.8.2 development: mvn versions:set -DnewVersion=2.8.2
http://git-wip-us.apache.org/repos/asf/hadoop/blob/c64fe9f0/hadoop-project/pom.xml.versionsBackup -- diff --git a/hadoop-project/pom.xml.versionsBackup b/hadoop-project/pom.xml.versionsBackup deleted file mode 100644 index f36fc2f..000 --- a/hadoop-project/pom.xml.versionsBackup +++ /dev/null @@ -1,1415 +0,0 @@ - - -http://maven.apache.org/POM/4.0.0; - xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance; - xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 - http://maven.apache.org/xsd/maven-4.0.0.xsd;> - 4.0.0 - -org.apache.hadoop -hadoop-main -2.8.0 - - org.apache.hadoop - hadoop-project - 2.8.0 - Apache Hadoop Project POM - Apache Hadoop Project POM - pom - - - -2017 - -false - true -6.1.26 -_ -_ - - -4 - - - -1.0.9 - -2.11.0 - -${project.version} - -1.0.13 - -${project.build.directory}/test-dir -${test.build.dir} - - - - ${basedir}/../../hadoop-common-project/hadoop-common/target -file:///dev/urandom - - -1.7.4 - - -1.9 - - -1.9.13 -2.2.3 - - - -2.5.0 -${env.HADOOP_PROTOC_PATH} - -3.4.6 -2.7.1 -3.0.0 - -6.0.48 -2.9.4 - - -2.0.0-M15 - - -1.7 - - - -[${javac.version},) -[3.0.2,) - - --Xmx2048m -XX:MaxPermSize=768m -XX:+HeapDumpOnOutOfMemoryError -2.17 - ${maven-surefire-plugin.version} - ${maven-surefire-plugin.version} - -2.5 -3.1 -2.5.1 -2.6 -2.5 -2.4 -2.3 - 2.7 -1.2 - 1.5 -1.9 -1.3.1 -1.0-beta-1 -1.0-alpha-8 -900 -1.10.6 - - - - - -jdiff -jdiff -${jdiff.version} - - -org.apache.hadoop -hadoop-assemblies -${project.version} - - -org.apache.hadoop -hadoop-annotations -${project.version} - - -org.apache.hadoop -hadoop-common -${project.version} - - -org.apache.hadoop -hadoop-common -${project.version} -test-jar - - -org.apache.hadoop -hadoop-auth -${project.version} - - -org.apache.hadoop -hadoop-auth -${project.version} -test-jar - - -org.apache.hadoop -hadoop-nfs -${project.version} - - -org.apache.hadoop -hadoop-hdfs -${project.version} - - -org.apache.hadoop -hadoop-hdfs-client -${project.version} - - -org.apache.hadoop -hadoop-hdfs -${project.version} -test-jar - - -org.apache.hadoop -hadoop-mapreduce-client-app -${project.version} - - -org.apache.hadoop -hadoop-mapreduce-client-app -${project.version} -test-jar - - -org.apache.hadoop -hadoop-mapreduce-client-common -${project.version} - - -org.apache.hadoop -hadoop-yarn-api -${project.version} - - - -org.apache.hadoop -hadoop-yarn-client -${project.version} - - - -org.apache.hadoop -hadoop-mapreduce-client-core -${project.version} - - - -org.apache.hadoop -hadoop-mapreduce-client-jobclient -${project.version} - - - -org.apache.hadoop -hadoop-mapreduce-client-shuffle -${project.version} - - - -org.apache.hadoop -hadoop-yarn -${project.version} -pom - - - -org.apache.hadoop -hadoop-yarn-server -${project.version} - - - -org.apache.hadoop -hadoop-yarn-server-web-proxy -${project.version} - - - -org.apache.hadoop -hadoop-yarn-server-common -${project.version} - - - -org.apache.hadoop - hadoop-yarn-server-tests -${project.version} -test-jar - - - -org.apache.hadoop -hadoop-yarn-common -${project.version} - - -org.apache.hadoop -hadoop-yarn-common -${project.version} -test-jar - - - -org.apache.hadoop -hadoop-yarn-registry -${project.version} - - -org.apache.hadoop -hadoop-yarn-server-nodemanager -${project.version} - - -org.apache.hadoop -hadoop-yarn-server-resourcemanager -${project.version} - - -org.apache.hadoop -hadoop-yarn-server-resourcemanager -${project.version} -test-jar - - - -org.apache.hadoop -
[6/8] hadoop git commit: Preparing for 2.8.2 development: mvn versions:set -DnewVersion=2.8.2
http://git-wip-us.apache.org/repos/asf/hadoop/blob/c64fe9f0/hadoop-hdfs-project/hadoop-hdfs-nfs/pom.xml.versionsBackup -- diff --git a/hadoop-hdfs-project/hadoop-hdfs-nfs/pom.xml.versionsBackup b/hadoop-hdfs-project/hadoop-hdfs-nfs/pom.xml.versionsBackup deleted file mode 100644 index b20e6e7..000 --- a/hadoop-hdfs-project/hadoop-hdfs-nfs/pom.xml.versionsBackup +++ /dev/null @@ -1,233 +0,0 @@ - - -http://maven.apache.org/POM/4.0.0; -xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance; -xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 -http://maven.apache.org/xsd/maven-4.0.0.xsd;> - 4.0.0 - -org.apache.hadoop -hadoop-project -2.8.0 -../../hadoop-project - - org.apache.hadoop - hadoop-hdfs-nfs - 2.8.0 - Apache Hadoop HDFS-NFS - Apache Hadoop HDFS-NFS - jar - - - - org.apache.hadoop - hadoop-annotations - provided - - - org.apache.hadoop - hadoop-auth - provided - - - org.apache.hadoop - hadoop-nfs - provided - - - io.netty - netty - compile - - - org.apache.hadoop - hadoop-hdfs - compile - - - org.apache.hadoop - hadoop-hdfs - test - test-jar - - - org.apache.hadoop - hadoop-common - provided - - - org.apache.hadoop - hadoop-common - test - test-jar - - - org.apache.zookeeper - zookeeper - test-jar - test - - - com.google.guava - guava - compile - - - org.mortbay.jetty - jetty - compile - - - org.mortbay.jetty - jetty-util - compile - - - com.sun.jersey - jersey-core - compile - - - com.sun.jersey - jersey-server - compile - - - commons-cli - commons-cli - compile - - - commons-codec - commons-codec - compile - - - commons-io - commons-io - compile - - - commons-lang - commons-lang - compile - - - commons-logging - commons-logging - compile - - - commons-daemon - commons-daemon - compile - - - log4j - log4j - compile - - - com.google.protobuf - protobuf-java - compile - - - javax.servlet - servlet-api - compile - - - junit - junit - test - - - org.mockito - mockito-all - test - - - org.slf4j - slf4j-log4j12 - provided - - - org.codehaus.jackson - jackson-core-asl - compile - - - org.codehaus.jackson - jackson-mapper-asl - compile - - - xmlenc - xmlenc - compile - - - org.bouncycastle - bcprov-jdk16 - test - - - - - - dist - -false - - - - -org.apache.maven.plugins -maven-assembly-plugin - - -org.apache.hadoop -hadoop-assemblies -${project.version} - - - - -dist -package - - single - - - ${project.artifactId}-${project.version} - false - false - - - ../../hadoop-assemblies/src/main/resources/assemblies/hadoop-hdfs-nfs-dist.xml - - - - - - - - - - - http://git-wip-us.apache.org/repos/asf/hadoop/blob/c64fe9f0/hadoop-hdfs-project/hadoop-hdfs/pom.xml -- diff --git a/hadoop-hdfs-project/hadoop-hdfs/pom.xml b/hadoop-hdfs-project/hadoop-hdfs/pom.xml index 6404cb9..2da1e2e 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/pom.xml +++ b/hadoop-hdfs-project/hadoop-hdfs/pom.xml @@ -20,12 +20,12 @@ http://maven.apache.org/xsd/maven-4.0.0.xsd;> org.apache.hadoop hadoop-project-dist -2.8.1 +2.8.2-SNAPSHOT ../../hadoop-project-dist org.apache.hadoop hadoop-hdfs - 2.8.1 + 2.8.2-SNAPSHOT Apache Hadoop HDFS Apache Hadoop HDFS jar http://git-wip-us.apache.org/repos/asf/hadoop/blob/c64fe9f0/hadoop-hdfs-project/hadoop-hdfs/pom.xml.versionsBackup -- diff --git a/hadoop-hdfs-project/hadoop-hdfs/pom.xml.versionsBackup b/hadoop-hdfs-project/hadoop-hdfs/pom.xml.versionsBackup deleted file mode 100644 index b3bd83f..000 --- a/hadoop-hdfs-project/hadoop-hdfs/pom.xml.versionsBackup +++ /dev/null @@ -1,479 +0,0 @@ - - -http://maven.apache.org/POM/4.0.0;
[7/8] hadoop git commit: Preparing for 2.8.2 development: mvn versions:set -DnewVersion=2.8.2
http://git-wip-us.apache.org/repos/asf/hadoop/blob/c64fe9f0/hadoop-common-project/hadoop-kms/pom.xml.versionsBackup -- diff --git a/hadoop-common-project/hadoop-kms/pom.xml.versionsBackup b/hadoop-common-project/hadoop-kms/pom.xml.versionsBackup deleted file mode 100644 index dac856f..000 --- a/hadoop-common-project/hadoop-kms/pom.xml.versionsBackup +++ /dev/null @@ -1,442 +0,0 @@ - - -http://maven.apache.org/POM/4.0.0; - xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance; - xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 - http://maven.apache.org/xsd/maven-4.0.0.xsd;> - 4.0.0 - -org.apache.hadoop -hadoop-project -2.8.0 -../../hadoop-project - - org.apache.hadoop - hadoop-kms - 2.8.0 - war - - Apache Hadoop KMS - Apache Hadoop KMS - - - - ${project.build.directory}/${project.artifactId}-${project.version}/share/hadoop/kms/tomcat - - - http://archive.apache.org/dist/tomcat/tomcat-6/v${tomcat.version}/bin/apache-tomcat-${tomcat.version}.tar.gz - - - - - - org.apache.hadoop - hadoop-minikdc - test - - - junit - junit - test - - - org.mockito - mockito-all - test - - - org.apache.hadoop - hadoop-auth - compile - - - com.google.guava - guava - compile - - - com.sun.jersey - jersey-core - compile - - - com.sun.jersey - jersey-server - compile - - - javax.servlet - servlet-api - provided - - - org.mortbay.jetty - jetty - test - - - org.apache.hadoop - hadoop-common - compile - - - javax.xml.stream - stax-api - - - commons-httpclient - commons-httpclient - - - tomcat - jasper-compiler - - - tomcat - jasper-runtime - - - javax.servlet - servlet-api - - - javax.servlet - jsp-api - - - javax.servlet.jsp - jsp-api - - - org.mortbay.jetty - jetty - - - org.mortbay.jetty - jetty-util - - - org.mortbay.jetty - jsp-api-2.1 - - - org.mortbay.jetty - servlet-api-2.5 - - - net.java.dev.jets3t - jets3t - - - org.eclipse.jdt - core - - - commons-el - commons-el - - - - - org.apache.hadoop - hadoop-common - test - test-jar - - - log4j - log4j - compile - - - org.slf4j - slf4j-api - compile - - - org.slf4j - slf4j-log4j12 - runtime - - - org.slf4j - jul-to-slf4j - compile - - - org.mortbay.jetty - jetty-util - compile - - - com.codahale.metrics - metrics-core - compile - - - org.apache.curator - curator-test - test - - - - org.bouncycastle - bcprov-jdk16 - test - - - - - - -org.apache.maven.plugins -maven-surefire-plugin - - always - 1 - 600 - - - listener - org.apache.hadoop.test.TimedOutTestsListener - - - -**/${test.exclude}.java -${test.exclude.pattern} - - - - -org.apache.maven.plugins -maven-antrun-plugin - - -create-web-xmls -generate-test-resources - - run - - - - - - - - - - - - - - -org.apache.maven.plugins -maven-war-plugin - - -default-war -prepare-package - - war - - - true - kms - ${project.build.directory}/kms - - - - - - -org.apache.maven.plugins -maven-jar-plugin - - -prepare-jar -prepare-package - - jar - - - classes - - - -prepare-test-jar -prepare-package - - test-jar - - - - - -org.codehaus.mojo -findbugs-maven-plugin - -
[5/8] hadoop git commit: Preparing for 2.8.2 development: mvn versions:set -DnewVersion=2.8.2
http://git-wip-us.apache.org/repos/asf/hadoop/blob/c64fe9f0/hadoop-mapreduce-project/hadoop-mapreduce-client/pom.xml.versionsBackup -- diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/pom.xml.versionsBackup b/hadoop-mapreduce-project/hadoop-mapreduce-client/pom.xml.versionsBackup deleted file mode 100644 index acbe3a2..000 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/pom.xml.versionsBackup +++ /dev/null @@ -1,318 +0,0 @@ - - -http://maven.apache.org/POM/4.0.0; - xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance; - xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 - http://maven.apache.org/xsd/maven-4.0.0.xsd;> - 4.0.0 - -org.apache.hadoop -hadoop-project -2.8.0 -../../hadoop-project - - org.apache.hadoop - hadoop-mapreduce-client - 2.8.0 - Apache Hadoop MapReduce Client - pom - - - ${basedir}/../../../hadoop-common-project/hadoop-common/target - - -../dev-support - - - - - com.google.protobuf - protobuf-java - - - org.apache.avro - avro - - - org.mortbay.jetty - jetty - - - org.apache.ant - ant - - - org.jboss.netty - netty - - - org.apache.velocity - velocity - - - org.slf4j - slf4j-api - - - paranamer-ant - com.thoughtworks.paranamer - - - - - org.apache.hadoop - hadoop-common - provided - - - - org.slf4j - slf4j-api - - - org.slf4j - slf4j-log4j12 - - - org.apache.hadoop - hadoop-annotations - - - org.mockito - mockito-all - test - - - org.apache.hadoop - hadoop-common - test-jar - test - - - org.apache.hadoop - hadoop-hdfs - test - - - com.google.inject.extensions - guice-servlet - - - junit - junit - test - - - io.netty - netty - - - commons-logging - commons-logging - provided - - - com.google.guava - guava - provided - - - commons-codec - commons-codec - provided - - - commons-cli - commons-cli - provided - - - commons-lang - commons-lang - provided - - - commons-collections - commons-collections - provided - - - com.sun.jersey.jersey-test-framework - jersey-test-framework-grizzly2 - test - - - - - - -org.codehaus.mojo -findbugs-maven-plugin - - true - true - ${mr.basedir}/dev-support/findbugs-exclude.xml - Max - - - - org.apache.maven.plugins - maven-surefire-plugin - - - - listener - org.apache.hadoop.test.TimedOutTestsListener - - - - - - - - - docs - -false - - -2.7.2 --unstable -512m - - - - -org.apache.maven.plugins -maven-javadoc-plugin - - - - javadoc - -prepare-package - - - - -org.apache.maven.plugins -maven-dependency-plugin - - -site -prepare-package - - copy - - - - - jdiff - jdiff - ${jdiff.version} - false - ${project.build.directory} - jdiff.jar - - - org.apache.hadoop - hadoop-annotations - ${project.version} - false - ${project.build.directory} - hadoop-annotations.jar - - - xerces - xercesImpl - ${xerces.version.jdiff} - false - ${project.build.directory} - xerces.jar - - - - - - - -org.apache.maven.plugins -maven-antrun-plugin - - -site -prepare-package - - run - - -
[1/8] hadoop git commit: Preparing for 2.8.2 development: mvn versions:set -DnewVersion=2.8.2
Repository: hadoop Updated Branches: refs/heads/branch-2.8.2 c24e51a35 -> c64fe9f07 http://git-wip-us.apache.org/repos/asf/hadoop/blob/c64fe9f0/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/pom.xml.versionsBackup -- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/pom.xml.versionsBackup b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/pom.xml.versionsBackup deleted file mode 100644 index 5415848..000 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/pom.xml.versionsBackup +++ /dev/null @@ -1,375 +0,0 @@ - - -http://maven.apache.org/POM/4.0.0; - xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance; - xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 - http://maven.apache.org/xsd/maven-4.0.0.xsd;> - -hadoop-yarn-server -org.apache.hadoop -2.8.0 - - 4.0.0 - org.apache.hadoop - hadoop-yarn-server-resourcemanager - 2.8.0 - Apache Hadoop YARN ResourceManager - - - -${project.parent.parent.basedir} - - - - - javax.servlet - servlet-api - - - - org.apache.hadoop - hadoop-common - provided - - - - org.slf4j - slf4j-api - - - - org.apache.hadoop - hadoop-annotations - - - org.mockito - mockito-all - test - - - - org.apache.hadoop - hadoop-common - test-jar - test - - - com.google.inject.extensions - guice-servlet - - - com.google.protobuf - protobuf-java - - - junit - junit - test - - - commons-io - commons-io - - - - org.apache.hadoop - hadoop-hdfs - test - - - com.google.inject - guice - - - com.sun.jersey.jersey-test-framework - jersey-test-framework-core - test - - - com.sun.jersey - jersey-json - - - com.sun.jersey.contribs - jersey-guice - - - - org.apache.hadoop - hadoop-yarn-common - test-jar - test - - - - org.apache.hadoop - hadoop-yarn-common - - - - org.apache.hadoop - hadoop-yarn-api - - - javax.xml.bind - jaxb-api - - - org.codehaus.jettison - jettison - - - commons-lang - commons-lang - - - com.sun.jersey - jersey-core - - - com.sun.jersey - jersey-client - - - org.mortbay.jetty - jetty-util - - - com.google.guava - guava - - - commons-logging - commons-logging - - - log4j - log4j - - - - - org.apache.hadoop - hadoop-yarn-server-common - - - - org.apache.hadoop - hadoop-yarn-server-applicationhistoryservice - ${project.version} - - - org.apache.hadoop - hadoop-yarn-server-web-proxy - - - org.apache.curator - curator-client - - - org.apache.curator - curator-test - - - org.apache.zookeeper - zookeeper - - - org.fusesource.leveldbjni - leveldbjni-all - - - org.apache.zookeeper - zookeeper - test-jar - - - - org.apache.hadoop - hadoop-hdfs - test-jar - test - - - - - org.apache.hadoop - hadoop-minikdc - test - - - - org.apache.hadoop - hadoop-auth - test - test-jar - ${project.version} - - - - - com.sun.jersey.jersey-test-framework - jersey-test-framework-grizzly2 - test - - - - - - - - ${basedir}/src/test/resources - - - ${basedir}/conf - - capacity-scheduler.xml - - - - - - - - - -maven-jar-plugin - - - - test-jar - -test-compile - - - - org.apache.hadoop.test.YarnTestDriver - - - -**/core-site.xml -**/yarn-site.xml - - - - - - - -org.apache.hadoop -hadoop-maven-plugins - - -compile-protoc -generate-sources - - protoc - - - ${protobuf.version} - ${protoc.path} - - ${basedir}/../../../../hadoop-common-project/hadoop-common/src/main/proto -${basedir}/../../hadoop-yarn-api/src/main/proto - ${basedir}/../../hadoop-yarn-common/src/main/proto -
[2/4] hadoop git commit: Sanitize arguments before launching Docker containers.
Sanitize arguments before launching Docker containers. (cherry picked from commit 23bbbce63e963284219a68f992f5febc73300d7d) Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/12ce6af1 Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/12ce6af1 Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/12ce6af1 Branch: refs/heads/branch-2.8.1-private Commit: 12ce6af15d36d9230fcdd1e651e645d0972f3eca Parents: bbe3b08 Author: Varun VasudevAuthored: Thu May 18 12:03:52 2017 +0530 Committer: Vinod Kumar Vavilapalli (I am also known as @tshooter.) Committed: Wed Jun 7 13:37:41 2017 -0700 -- .../impl/container-executor.c | 173 ++- .../impl/container-executor.h | 8 +- .../test/test-container-executor.c | 44 + 3 files changed, 223 insertions(+), 2 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hadoop/blob/12ce6af1/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/native/container-executor/impl/container-executor.c -- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/native/container-executor/impl/container-executor.c b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/native/container-executor/impl/container-executor.c index bdbcdfa..7aa36cb 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/native/container-executor/impl/container-executor.c +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/native/container-executor/impl/container-executor.c @@ -40,6 +40,7 @@ #include #include #include +#include #include "config.h" @@ -1117,7 +1118,170 @@ int initialize_app(const char *user, const char *app_id, return -1; } +static char* escape_single_quote(const char *str) { + int p = 0; + int i = 0; + char replacement[] = "'\"'\"'"; + size_t replacement_length = strlen(replacement); + size_t ret_size = strlen(str) * replacement_length + 1; + char *ret = (char *) calloc(ret_size, sizeof(char)); + if(ret == NULL) { +exit(OUT_OF_MEMORY); + } + while(str[p] != '\0') { +if(str[p] == '\'') { + strncat(ret, replacement, ret_size - strlen(ret)); + i += replacement_length; +} +else { + ret[i] = str[p]; + ret[i + 1] = '\0'; + i++; +} +p++; + } + return ret; +} + +static void quote_and_append_arg(char **str, size_t *size, const char* param, const char *arg) { + char *tmp = escape_single_quote(arg); + strcat(*str, param); + strcat(*str, "'"); + if(strlen(*str) + strlen(tmp) > *size) { +*str = (char *) realloc(*str, strlen(*str) + strlen(tmp) + 1024); +if(*str == NULL) { + exit(OUT_OF_MEMORY); +} +*size = strlen(*str) + strlen(tmp) + 1024; + } + strcat(*str, tmp); + strcat(*str, "' "); + free(tmp); +} + +char** tokenize_docker_command(const char *input, int *split_counter) { + char *line = (char *)calloc(strlen(input) + 1, sizeof(char)); + char **linesplit = (char **) malloc(sizeof(char *)); + char *p = NULL; + int c = 0; + *split_counter = 0; + strncpy(line, input, strlen(input)); + + p = strtok(line, " "); + while(p != NULL) { +linesplit[*split_counter] = p; +(*split_counter)++; +linesplit = realloc(linesplit, (sizeof(char *) * (*split_counter + 1))); +if(linesplit == NULL) { + fprintf(ERRORFILE, "Cannot allocate memory to parse docker command %s", + strerror(errno)); + fflush(ERRORFILE); + exit(OUT_OF_MEMORY); +} +p = strtok(NULL, " "); + } + linesplit[*split_counter] = NULL; + return linesplit; +} + +char* sanitize_docker_command(const char *line) { + static struct option long_options[] = { +{"name", required_argument, 0, 'n' }, +{"user", required_argument, 0, 'u' }, +{"rm", no_argument, 0, 'r' }, +{"workdir", required_argument, 0, 'w' }, +{"net", required_argument, 0, 'e' }, +{"cgroup-parent", required_argument, 0, 'g' }, +{"privileged", no_argument, 0, 'p' }, +{"cap-add", required_argument, 0, 'a' }, +{"cap-drop", required_argument, 0, 'o' }, +{"device", required_argument, 0, 'i' }, +{"detach", required_argument, 0, 't' }, +{0, 0, 0, 0} + }; + + int c = 0; + int option_index = 0; + char *output = NULL; + size_t output_size = 0; + char **linesplit; + int split_counter = 0; + int len = strlen(line); + + linesplit = tokenize_docker_command(line, _counter); + + output_size = len * 2; + output = (char *) calloc(output_size, sizeof(char)); + if(output == NULL) { +
[3/4] hadoop git commit: Fix Docker sanitization.
Fix Docker sanitization. Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/045f59fb Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/045f59fb Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/045f59fb Branch: refs/heads/branch-2.8.1-private Commit: 045f59fbb4b0965ab0e5d7ed2913f203dd19edcf Parents: 12ce6af Author: Varun VasudevAuthored: Wed May 24 16:03:28 2017 +0530 Committer: Vinod Kumar Vavilapalli (I am also known as @tshooter.) Committed: Wed Jun 7 13:37:51 2017 -0700 -- .../impl/container-executor.c | 6 +- .../test/test-container-executor.c | 20 +--- 2 files changed, 6 insertions(+), 20 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hadoop/blob/045f59fb/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/native/container-executor/impl/container-executor.c -- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/native/container-executor/impl/container-executor.c b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/native/container-executor/impl/container-executor.c index 7aa36cb..ddbf738 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/native/container-executor/impl/container-executor.c +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/native/container-executor/impl/container-executor.c @@ -1268,13 +1268,9 @@ char* sanitize_docker_command(const char *line) { } if(optind < split_counter) { -quote_and_append_arg(, _size, "", linesplit[optind++]); -strcat(output, "'"); while(optind < split_counter) { - strcat(output, linesplit[optind++]); - strcat(output, " "); + quote_and_append_arg(, _size, "", linesplit[optind++]); } -strcat(output, "'"); } return output; http://git-wip-us.apache.org/repos/asf/hadoop/blob/045f59fb/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/native/container-executor/test/test-container-executor.c -- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/native/container-executor/test/test-container-executor.c b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/native/container-executor/test/test-container-executor.c index fcc05a3..7a4dda2 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/native/container-executor/test/test-container-executor.c +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/native/container-executor/test/test-container-executor.c @@ -715,11 +715,6 @@ void test_run_container() { void test_sanitize_docker_command() { -/* - char *input[] = { -"run " - }; -*/ char *input[] = { "run --name=cname --user=nobody -d --workdir=/yarn/local/cdir --privileged --rm --device=/sys/fs/cgroup/device:/sys/fs/cgroup/device --detach=true --cgroup-parent=/sys/fs/cgroup/cpu/yarn/cid --net=host --cap-drop=ALL --cap-add=SYS_CHROOT --cap-add=MKNOD --cap-add=SETFCAP --cap-add=SETPCAP --cap-add=FSETID --cap-add=CHOWN --cap-add=AUDIT_WRITE --cap-add=SETGID --cap-add=NET_RAW --cap-add=FOWNER --cap-add=SETUID --cap-add=DAC_OVERRIDE --cap-add=KILL --cap-add=NET_BIND_SERVICE -v /sys/fs/cgroup:/sys/fs/cgroup:ro -v /yarn/local/cdir:/yarn/local/cdir -v /yarn/local/usercache/test/:/yarn/local/usercache/test/ ubuntu bash /yarn/local/usercache/test/appcache/aid/cid/launch_container.sh", "run --name=$CID --user=nobody -d --workdir=/yarn/local/cdir --privileged --rm --device=/sys/fs/cgroup/device:/sys/fs/cgroup/device --detach=true --cgroup-parent=/sys/fs/cgroup/cpu/yarn/cid --net=host --cap-drop=ALL --cap-add=SYS_CHROOT --cap-add=MKNOD --cap-add=SETFCAP --cap-add=SETPCAP --cap-add=FSETID --cap-add=CHOWN --cap-add=AUDIT_WRITE --cap-add=SETGID --cap-add=NET_RAW --cap-add=FOWNER --cap-add=SETUID --cap-add=DAC_OVERRIDE --cap-add=KILL --cap-add=NET_BIND_SERVICE -v /sys/fs/cgroup:/sys/fs/cgroup:ro -v /yarn/local/cdir:/yarn/local/cdir -v /yarn/local/usercache/test/:/yarn/local/usercache/test/ ubuntu bash /yarn/local/usercache/test/appcache/aid/cid/launch_container.sh", @@ -727,17 +722,12 @@ void test_sanitize_docker_command() { "run --name=cname --user=nobody -d --workdir=/yarn/local/cdir --privileged --rm --device=/sys/fs/cgroup/device:/sys/fs/cgroup/device
[4/4] hadoop git commit: Preparing for 2.8.1 release: mvn versions:set -DnewVersion=2.8.1
Preparing for 2.8.1 release: mvn versions:set -DnewVersion=2.8.1 Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/1e6296df Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/1e6296df Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/1e6296df Branch: refs/heads/branch-2.8.1-private Commit: 1e6296df38f9cd3d9581c8af58a2a03a6e4312be Parents: 045f59fb Author: @tshooter.) <@tshooter.)> Authored: Sat May 20 12:56:16 2017 -0700 Committer: Vinod Kumar Vavilapalli (I am also known as @tshooter.)Committed: Wed Jun 7 13:37:58 2017 -0700 -- hadoop-assemblies/pom.xml| 4 ++-- hadoop-build-tools/pom.xml | 2 +- hadoop-client/pom.xml| 4 ++-- hadoop-common-project/hadoop-annotations/pom.xml | 4 ++-- hadoop-common-project/hadoop-auth-examples/pom.xml | 4 ++-- hadoop-common-project/hadoop-auth/pom.xml| 4 ++-- hadoop-common-project/hadoop-common/pom.xml | 4 ++-- hadoop-common-project/hadoop-kms/pom.xml | 4 ++-- hadoop-common-project/hadoop-minikdc/pom.xml | 4 ++-- hadoop-common-project/hadoop-nfs/pom.xml | 4 ++-- hadoop-common-project/pom.xml| 4 ++-- hadoop-dist/pom.xml | 4 ++-- hadoop-hdfs-project/hadoop-hdfs-client/pom.xml | 4 ++-- hadoop-hdfs-project/hadoop-hdfs-httpfs/pom.xml | 4 ++-- hadoop-hdfs-project/hadoop-hdfs-native-client/pom.xml| 4 ++-- hadoop-hdfs-project/hadoop-hdfs-nfs/pom.xml | 4 ++-- hadoop-hdfs-project/hadoop-hdfs/pom.xml | 4 ++-- hadoop-hdfs-project/hadoop-hdfs/src/contrib/bkjournal/pom.xml| 4 ++-- hadoop-hdfs-project/pom.xml | 4 ++-- .../hadoop-mapreduce-client/hadoop-mapreduce-client-app/pom.xml | 4 ++-- .../hadoop-mapreduce-client-common/pom.xml | 4 ++-- .../hadoop-mapreduce-client/hadoop-mapreduce-client-core/pom.xml | 4 ++-- .../hadoop-mapreduce-client-hs-plugins/pom.xml | 4 ++-- .../hadoop-mapreduce-client/hadoop-mapreduce-client-hs/pom.xml | 4 ++-- .../hadoop-mapreduce-client-jobclient/pom.xml| 4 ++-- .../hadoop-mapreduce-client-shuffle/pom.xml | 4 ++-- hadoop-mapreduce-project/hadoop-mapreduce-client/pom.xml | 4 ++-- hadoop-mapreduce-project/hadoop-mapreduce-examples/pom.xml | 4 ++-- hadoop-mapreduce-project/pom.xml | 4 ++-- hadoop-maven-plugins/pom.xml | 2 +- hadoop-minicluster/pom.xml | 4 ++-- hadoop-project-dist/pom.xml | 4 ++-- hadoop-project/pom.xml | 4 ++-- hadoop-project/src/site/markdown/index.md.vm | 2 +- hadoop-tools/hadoop-ant/pom.xml | 4 ++-- hadoop-tools/hadoop-archive-logs/pom.xml | 4 ++-- hadoop-tools/hadoop-archives/pom.xml | 4 ++-- hadoop-tools/hadoop-aws/pom.xml | 4 ++-- hadoop-tools/hadoop-azure-datalake/pom.xml | 2 +- hadoop-tools/hadoop-azure/pom.xml| 2 +- hadoop-tools/hadoop-datajoin/pom.xml | 4 ++-- hadoop-tools/hadoop-distcp/pom.xml | 4 ++-- hadoop-tools/hadoop-extras/pom.xml | 4 ++-- hadoop-tools/hadoop-gridmix/pom.xml | 4 ++-- hadoop-tools/hadoop-openstack/pom.xml| 4 ++-- hadoop-tools/hadoop-pipes/pom.xml| 4 ++-- hadoop-tools/hadoop-rumen/pom.xml| 4 ++-- hadoop-tools/hadoop-sls/pom.xml | 4 ++-- hadoop-tools/hadoop-streaming/pom.xml| 4 ++-- hadoop-tools/hadoop-tools-dist/pom.xml | 4 ++-- hadoop-tools/pom.xml | 4 ++-- hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/pom.xml | 4 ++-- .../hadoop-yarn-applications-distributedshell/pom.xml| 4 ++-- .../hadoop-yarn-applications-unmanaged-am-launcher/pom.xml | 4 ++-- hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/pom.xml | 4 ++-- hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/pom.xml | 4 ++-- hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/pom.xml | 4 ++--
[1/4] hadoop git commit: Validate docker image name before launching container.
Repository: hadoop Updated Branches: refs/heads/branch-2.8.1-private [created] 1e6296df3 Validate docker image name before launching container. (cherry picked from commit 51e65cc7104bcccdfc2554f489c8a5c0e8feea37) Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/bbe3b085 Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/bbe3b085 Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/bbe3b085 Branch: refs/heads/branch-2.8.1-private Commit: bbe3b0857d383c5e4dc4a7ade90a88a3e24338b2 Parents: 91f2b7a Author: Varun VasudevAuthored: Thu May 18 11:53:16 2017 +0530 Committer: Vinod Kumar Vavilapalli (I am also known as @tshooter.) Committed: Wed Jun 7 13:37:34 2017 -0700 -- .../runtime/DockerLinuxContainerRuntime.java| 24 +--- .../runtime/TestDockerContainerRuntime.java | 29 2 files changed, 49 insertions(+), 4 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hadoop/blob/bbe3b085/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/containermanager/linux/runtime/DockerLinuxContainerRuntime.java -- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/containermanager/linux/runtime/DockerLinuxContainerRuntime.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/containermanager/linux/runtime/DockerLinuxContainerRuntime.java index c303e94..fc3376a 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/containermanager/linux/runtime/DockerLinuxContainerRuntime.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/containermanager/linux/runtime/DockerLinuxContainerRuntime.java @@ -51,6 +51,7 @@ import java.util.HashSet; import java.util.List; import java.util.Map; import java.util.Set; +import java.util.regex.Pattern; import static org.apache.hadoop.yarn.server.nodemanager.containermanager.linux.runtime.LinuxContainerRuntimeConstants.*; @@ -60,6 +61,12 @@ public class DockerLinuxContainerRuntime implements LinuxContainerRuntime { private static final Log LOG = LogFactory.getLog( DockerLinuxContainerRuntime.class); + // This validates that the image is a proper docker image + public static final String DOCKER_IMAGE_PATTERN = + "^(([a-zA-Z0-9.-]+)(:\\d+)?/)?([a-z0-9_./-]+)(:[\\w.-]+)?$"; + private static final Pattern dockerImagePattern = + Pattern.compile(DOCKER_IMAGE_PATTERN); + @InterfaceAudience.Private public static final String ENV_DOCKER_CONTAINER_IMAGE = "YARN_CONTAINER_RUNTIME_DOCKER_IMAGE"; @@ -216,10 +223,7 @@ public class DockerLinuxContainerRuntime implements LinuxContainerRuntime { .getEnvironment(); String imageName = environment.get(ENV_DOCKER_CONTAINER_IMAGE); -if (imageName == null) { - throw new ContainerExecutionException(ENV_DOCKER_CONTAINER_IMAGE - + " not set!"); -} +validateImageName(imageName); String containerIdStr = container.getContainerId().toString(); String runAsUser = ctx.getExecutionAttribute(RUN_AS_USER); @@ -354,4 +358,16 @@ public class DockerLinuxContainerRuntime implements LinuxContainerRuntime { throws ContainerExecutionException { } + + public static void validateImageName(String imageName) + throws ContainerExecutionException { +if (imageName == null || imageName.isEmpty()) { + throw new ContainerExecutionException( + ENV_DOCKER_CONTAINER_IMAGE + " not set!"); +} +if (!dockerImagePattern.matcher(imageName).matches()) { + throw new ContainerExecutionException("Image name '" + imageName + + "' doesn't match docker image name pattern"); +} + } } http://git-wip-us.apache.org/repos/asf/hadoop/blob/bbe3b085/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/test/java/org/apache/hadoop/yarn/server/nodemanager/containermanager/linux/runtime/TestDockerContainerRuntime.java -- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/test/java/org/apache/hadoop/yarn/server/nodemanager/containermanager/linux/runtime/TestDockerContainerRuntime.java
[hadoop] Git Push Summary
Repository: hadoop Updated Branches: refs/heads/branch-2.8.2 [created] c24e51a35 - To unsubscribe, e-mail: common-commits-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-commits-h...@hadoop.apache.org
svn commit: r14971 - in /release/hadoop/common: ./ hadoop-2.7.3/
Author: vinodkv Date: Thu Aug 25 19:25:04 2016 New Revision: 14971 Log: Publishing the bits for release 2.7.3 Added: release/hadoop/common/hadoop-2.7.3/ release/hadoop/common/hadoop-2.7.3/hadoop-2.7.3-src.tar.gz (with props) release/hadoop/common/hadoop-2.7.3/hadoop-2.7.3-src.tar.gz.asc (with props) release/hadoop/common/hadoop-2.7.3/hadoop-2.7.3-src.tar.gz.mds release/hadoop/common/hadoop-2.7.3/hadoop-2.7.3.tar.gz (with props) release/hadoop/common/hadoop-2.7.3/hadoop-2.7.3.tar.gz.asc (with props) release/hadoop/common/hadoop-2.7.3/hadoop-2.7.3.tar.gz.mds Modified: release/hadoop/common/current release/hadoop/common/stable2 Modified: release/hadoop/common/current == --- release/hadoop/common/current (original) +++ release/hadoop/common/current Thu Aug 25 19:25:04 2016 @@ -1 +1 @@ -link hadoop-2.7.2 \ No newline at end of file +link hadoop-2.7.3 \ No newline at end of file Added: release/hadoop/common/hadoop-2.7.3/hadoop-2.7.3-src.tar.gz == Binary file - no diff available. Propchange: release/hadoop/common/hadoop-2.7.3/hadoop-2.7.3-src.tar.gz -- svn:mime-type = application/x-gzip Added: release/hadoop/common/hadoop-2.7.3/hadoop-2.7.3-src.tar.gz.asc == Binary file - no diff available. Propchange: release/hadoop/common/hadoop-2.7.3/hadoop-2.7.3-src.tar.gz.asc -- svn:mime-type = application/pgp-signature Added: release/hadoop/common/hadoop-2.7.3/hadoop-2.7.3-src.tar.gz.mds == --- release/hadoop/common/hadoop-2.7.3/hadoop-2.7.3-src.tar.gz.mds (added) +++ release/hadoop/common/hadoop-2.7.3/hadoop-2.7.3-src.tar.gz.mds Thu Aug 25 19:25:04 2016 @@ -0,0 +1,17 @@ +hadoop-2.7.3-src.tar.gz:MD5 = 95 5C EA EB 83 17 E6 EB D6 DE 06 65 87 80 72 + 3F +hadoop-2.7.3-src.tar.gz: SHA1 = A067 EA2C 656D F6FF 064C 71DE CA01 9DC5 EA99 + 317F +hadoop-2.7.3-src.tar.gz: RMD160 = CE3F 00C7 4C03 D5F1 0FCB 2AB1 302B FFFD FCAE + 9A91 +hadoop-2.7.3-src.tar.gz: SHA224 = 41BA6C94 23E0CC92 200EFF93 29C059E4 E029700C + 8A662845 7A232A38 +hadoop-2.7.3-src.tar.gz: SHA256 = 227785DC 6E3E6EF8 CFD64393 B305D090 78A20970 + 3C9C0191 0A1BDDCF 86BE3054 +hadoop-2.7.3-src.tar.gz: SHA384 = 9DEB84BB 7787D8FC A0E45936 25B93958 A52B064E + CAE6CB7D 2DFEFFDF 54F79CFE 5C451B85 FA287339 + 6DCD59C8 47141FD4 +hadoop-2.7.3-src.tar.gz: SHA512 = 8451F89D 3CBB6728 88ABC67C 76A53B2D 50F44B88 + 78127C3E 361CB354 CD1B5A3A 2BC7D531 C1BA67E9 + BC3D17E5 C6AA496D 11969484 C12C86B5 6E8823CD + 1AB6482A Added: release/hadoop/common/hadoop-2.7.3/hadoop-2.7.3.tar.gz == Binary file - no diff available. Propchange: release/hadoop/common/hadoop-2.7.3/hadoop-2.7.3.tar.gz -- svn:mime-type = application/x-gzip Added: release/hadoop/common/hadoop-2.7.3/hadoop-2.7.3.tar.gz.asc == Binary file - no diff available. Propchange: release/hadoop/common/hadoop-2.7.3/hadoop-2.7.3.tar.gz.asc -- svn:mime-type = application/pgp-signature Added: release/hadoop/common/hadoop-2.7.3/hadoop-2.7.3.tar.gz.mds == --- release/hadoop/common/hadoop-2.7.3/hadoop-2.7.3.tar.gz.mds (added) +++ release/hadoop/common/hadoop-2.7.3/hadoop-2.7.3.tar.gz.mds Thu Aug 25 19:25:04 2016 @@ -0,0 +1,14 @@ +hadoop-2.7.3.tar.gz:MD5 = 34 55 BB 57 E4 B4 90 6B BE A6 7B 58 CC A7 8F A8 +hadoop-2.7.3.tar.gz: SHA1 = B84B 8989 3426 9C68 753E 4E03 6D21 395E 5A4A B5B1 +hadoop-2.7.3.tar.gz: RMD160 = 8FE4 A91E 8C67 2A33 C4E9 61FB 607A DBBD 1AE5 E03A +hadoop-2.7.3.tar.gz: SHA224 = 23AB1EAB B7648921 7101671C DCF9D774 7B84AD50 + 6A74E300 AE6617FA +hadoop-2.7.3.tar.gz: SHA256 = D489DF38 08244B90 6EB38F4D 081BA49E 50C4603D + B03EFD5E 594A1E98 B09259C2 +hadoop-2.7.3.tar.gz: SHA384 = EFB42E60 3AF4FFB2 BA9F4CF4 1B56F71B D3F3BD8F + 23331C25 27267762 FDEB67F0 F2B6F56D 797842DB + BB8C9F75 9DBA195D
[hadoop] Git Push Summary
Repository: hadoop Updated Tags: refs/tags/rel/release-2.7.3 [created] 269673cce - To unsubscribe, e-mail: common-commits-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-commits-h...@hadoop.apache.org
[1/2] hadoop git commit: HDFS-10692. Update JDiff report's base version for HDFS from 2.6.0 to 2.7.2. Contributed by Wangda Tan.
Repository: hadoop Updated Branches: refs/heads/branch-3.0.0-alpha1 67a4d3cce -> c462ee356 http://git-wip-us.apache.org/repos/asf/hadoop/blob/c462ee35/hadoop-hdfs-project/hadoop-hdfs/dev-support/jdiff/Apache_Hadoop_HDFS_2.7.2.xml -- diff --git a/hadoop-hdfs-project/hadoop-hdfs/dev-support/jdiff/Apache_Hadoop_HDFS_2.7.2.xml b/hadoop-hdfs-project/hadoop-hdfs/dev-support/jdiff/Apache_Hadoop_HDFS_2.7.2.xml new file mode 100644 index 000..028ba2d --- /dev/null +++ b/hadoop-hdfs-project/hadoop-hdfs/dev-support/jdiff/Apache_Hadoop_HDFS_2.7.2.xml @@ -0,0 +1,21505 @@ + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
[2/2] hadoop git commit: HDFS-10692. Update JDiff report's base version for HDFS from 2.6.0 to 2.7.2. Contributed by Wangda Tan.
HDFS-10692. Update JDiff report's base version for HDFS from 2.6.0 to 2.7.2. Contributed by Wangda Tan. (cherry picked from commit dc7a1c54f4447fb01980ae61d5d67e90c4f52f00) Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/c462ee35 Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/c462ee35 Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/c462ee35 Branch: refs/heads/branch-3.0.0-alpha1 Commit: c462ee356b1bbe89f152540d45f0257a01b56018 Parents: 67a4d3c Author: Vinod Kumar Vavilapalli (I am also known as @tshooter.)Authored: Mon Aug 22 11:35:18 2016 -0700 Committer: Vinod Kumar Vavilapalli (I am also known as @tshooter.) Committed: Mon Aug 22 11:38:09 2016 -0700 -- .../jdiff/Apache_Hadoop_HDFS_2.7.2.xml | 21505 + 1 file changed, 21505 insertions(+) -- - To unsubscribe, e-mail: common-commits-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-commits-h...@hadoop.apache.org
[1/2] hadoop git commit: HDFS-10692. Update JDiff report's base version for HDFS from 2.6.0 to 2.7.2. Contributed by Wangda Tan.
Repository: hadoop Updated Branches: refs/heads/branch-2 62c7d09ef -> 8a6a7c82f http://git-wip-us.apache.org/repos/asf/hadoop/blob/8a6a7c82/hadoop-hdfs-project/hadoop-hdfs/dev-support/jdiff/Apache_Hadoop_HDFS_2.7.2.xml -- diff --git a/hadoop-hdfs-project/hadoop-hdfs/dev-support/jdiff/Apache_Hadoop_HDFS_2.7.2.xml b/hadoop-hdfs-project/hadoop-hdfs/dev-support/jdiff/Apache_Hadoop_HDFS_2.7.2.xml new file mode 100644 index 000..028ba2d --- /dev/null +++ b/hadoop-hdfs-project/hadoop-hdfs/dev-support/jdiff/Apache_Hadoop_HDFS_2.7.2.xml @@ -0,0 +1,21505 @@ + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
[1/2] hadoop git commit: HDFS-10692. Update JDiff report's base version for HDFS from 2.6.0 to 2.7.2. Contributed by Wangda Tan.
Repository: hadoop Updated Branches: refs/heads/branch-2.8 e714d8219 -> c0c4cc8ee http://git-wip-us.apache.org/repos/asf/hadoop/blob/c0c4cc8e/hadoop-hdfs-project/hadoop-hdfs/dev-support/jdiff/Apache_Hadoop_HDFS_2.7.2.xml -- diff --git a/hadoop-hdfs-project/hadoop-hdfs/dev-support/jdiff/Apache_Hadoop_HDFS_2.7.2.xml b/hadoop-hdfs-project/hadoop-hdfs/dev-support/jdiff/Apache_Hadoop_HDFS_2.7.2.xml new file mode 100644 index 000..028ba2d --- /dev/null +++ b/hadoop-hdfs-project/hadoop-hdfs/dev-support/jdiff/Apache_Hadoop_HDFS_2.7.2.xml @@ -0,0 +1,21505 @@ + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
[2/2] hadoop git commit: HDFS-10692. Update JDiff report's base version for HDFS from 2.6.0 to 2.7.2. Contributed by Wangda Tan.
HDFS-10692. Update JDiff report's base version for HDFS from 2.6.0 to 2.7.2. Contributed by Wangda Tan. (cherry picked from commit dc7a1c54f4447fb01980ae61d5d67e90c4f52f00) Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/c0c4cc8e Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/c0c4cc8e Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/c0c4cc8e Branch: refs/heads/branch-2.8 Commit: c0c4cc8ee8840ccfa75cac73894f6867e6d3dc7c Parents: e714d82 Author: Vinod Kumar Vavilapalli (I am also known as @tshooter.)Authored: Mon Aug 22 11:35:18 2016 -0700 Committer: Vinod Kumar Vavilapalli (I am also known as @tshooter.) Committed: Mon Aug 22 11:37:14 2016 -0700 -- .../jdiff/Apache_Hadoop_HDFS_2.7.2.xml | 21505 + 1 file changed, 21505 insertions(+) -- - To unsubscribe, e-mail: common-commits-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-commits-h...@hadoop.apache.org
[2/2] hadoop git commit: HDFS-10692. Update JDiff report's base version for HDFS from 2.6.0 to 2.7.2. Contributed by Wangda Tan.
HDFS-10692. Update JDiff report's base version for HDFS from 2.6.0 to 2.7.2. Contributed by Wangda Tan. (cherry picked from commit dc7a1c54f4447fb01980ae61d5d67e90c4f52f00) Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/8a6a7c82 Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/8a6a7c82 Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/8a6a7c82 Branch: refs/heads/branch-2 Commit: 8a6a7c82f259c91915c4786c9440ffa0a7b148d5 Parents: 62c7d09 Author: Vinod Kumar Vavilapalli (I am also known as @tshooter.)Authored: Mon Aug 22 11:35:18 2016 -0700 Committer: Vinod Kumar Vavilapalli (I am also known as @tshooter.) Committed: Mon Aug 22 11:36:38 2016 -0700 -- .../jdiff/Apache_Hadoop_HDFS_2.7.2.xml | 21505 + 1 file changed, 21505 insertions(+) -- - To unsubscribe, e-mail: common-commits-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-commits-h...@hadoop.apache.org
[2/2] hadoop git commit: HDFS-10692. Update JDiff report's base version for HDFS from 2.6.0 to 2.7.2. Contributed by Wangda Tan.
HDFS-10692. Update JDiff report's base version for HDFS from 2.6.0 to 2.7.2. Contributed by Wangda Tan. Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/dc7a1c54 Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/dc7a1c54 Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/dc7a1c54 Branch: refs/heads/trunk Commit: dc7a1c54f4447fb01980ae61d5d67e90c4f52f00 Parents: 115ecb5 Author: Vinod Kumar Vavilapalli (I am also known as @tshooter.)Authored: Mon Aug 22 11:35:18 2016 -0700 Committer: Vinod Kumar Vavilapalli (I am also known as @tshooter.) Committed: Mon Aug 22 11:35:18 2016 -0700 -- .../jdiff/Apache_Hadoop_HDFS_2.7.2.xml | 21505 + 1 file changed, 21505 insertions(+) -- - To unsubscribe, e-mail: common-commits-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-commits-h...@hadoop.apache.org
[1/2] hadoop git commit: HDFS-10692. Update JDiff report's base version for HDFS from 2.6.0 to 2.7.2. Contributed by Wangda Tan.
Repository: hadoop Updated Branches: refs/heads/trunk 115ecb52a -> dc7a1c54f http://git-wip-us.apache.org/repos/asf/hadoop/blob/dc7a1c54/hadoop-hdfs-project/hadoop-hdfs/dev-support/jdiff/Apache_Hadoop_HDFS_2.7.2.xml -- diff --git a/hadoop-hdfs-project/hadoop-hdfs/dev-support/jdiff/Apache_Hadoop_HDFS_2.7.2.xml b/hadoop-hdfs-project/hadoop-hdfs/dev-support/jdiff/Apache_Hadoop_HDFS_2.7.2.xml new file mode 100644 index 000..028ba2d --- /dev/null +++ b/hadoop-hdfs-project/hadoop-hdfs/dev-support/jdiff/Apache_Hadoop_HDFS_2.7.2.xml @@ -0,0 +1,21505 @@ + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
[3/3] hadoop git commit: HADOOP-13428. Fixed hadoop-common build files to generate jdiff by working around a jdiff-bug. Contributed by Wangda Tan.
HADOOP-13428. Fixed hadoop-common build files to generate jdiff by working around a jdiff-bug. Contributed by Wangda Tan. (cherry picked from commit 99603e902244f17b04cfd55122f47355d070b588) Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/67a4d3cc Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/67a4d3cc Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/67a4d3cc Branch: refs/heads/branch-3.0.0-alpha1 Commit: 67a4d3cce1e39813cbc7ec3c6535fbc5b6dee96e Parents: 25f2325 Author: Vinod Kumar Vavilapalli (I am also known as @tshooter.)Authored: Fri Aug 19 19:08:53 2016 -0700 Committer: Vinod Kumar Vavilapalli (I am also known as @tshooter.) Committed: Fri Aug 19 19:31:17 2016 -0700 -- .../dev-support/jdiff-workaround.patch |98 + .../jdiff/Apache_Hadoop_Common_2.7.2.xml| 46648 + hadoop-common-project/hadoop-common/pom.xml | 2 + hadoop-project-dist/pom.xml |44 +- 4 files changed, 46789 insertions(+), 3 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hadoop/blob/67a4d3cc/hadoop-common-project/hadoop-common/dev-support/jdiff-workaround.patch -- diff --git a/hadoop-common-project/hadoop-common/dev-support/jdiff-workaround.patch b/hadoop-common-project/hadoop-common/dev-support/jdiff-workaround.patch new file mode 100644 index 000..8f87d40 --- /dev/null +++ b/hadoop-common-project/hadoop-common/dev-support/jdiff-workaround.patch @@ -0,0 +1,98 @@ +diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/MetricsSystem.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/MetricsSystem.java +index a277abd..ed7c709 100644 +--- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/MetricsSystem.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/MetricsSystem.java +@@ -43,18 +43,6 @@ + public abstract MetricsSystem init(String prefix); + + /** +- * Register a metrics source +- * @paramthe actual type of the source object +- * @param source object to register +- * @param name of the source. Must be unique or null (then extracted from +- * the annotations of the source object.) +- * @param desc the description of the source (or null. See above.) +- * @return the source object +- * @exception MetricsException +- */ +- public abstract T register(String name, String desc, T source); +- +- /** +* Unregister a metrics source +* @param name of the source. This is the name you use to call register() +*/ +@@ -77,18 +65,19 @@ +*/ + @InterfaceAudience.Private + public abstract MetricsSource getSource(String name); ++ + + /** +- * Register a metrics sink +- * @paramthe type of the sink +- * @param sink to register +- * @param name of the sink. Must be unique. +- * @param desc the description of the sink +- * @return the sink ++ * Register a metrics source ++ * @paramthe actual type of the source object ++ * @param source object to register ++ * @param name of the source. Must be unique or null (then extracted from ++ * the annotations of the source object.) ++ * @param desc the description of the source (or null. See above.) ++ * @return the source object +* @exception MetricsException +*/ +- public abstract +- T register(String name, String desc, T sink); ++ public abstract T register(String name, String desc, T source); + + /** +* Register a callback interface for JMX events +diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/impl/MetricsSystemImpl.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/impl/MetricsSystemImpl.java +index 6986edb..eeea81f 100644 +--- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/impl/MetricsSystemImpl.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/impl/MetricsSystemImpl.java +@@ -270,27 +270,6 @@ void registerSource(String name, String desc, MetricsSource source) { + LOG.debug("Registered source "+ name); + } + +- @Override public synchronized +- T register(final String name, final String description, final T sink) { +-LOG.debug(name +", "+ description); +-if (allSinks.containsKey(name)) { +- LOG.warn("Sink "+ name +" already exists!"); +- return sink; +-} +-allSinks.put(name, sink); +-if (config != null) { +- registerSink(name, description, sink); +-} +-// We want to re-register the sink to pick up new config +-//
[1/3] hadoop git commit: HADOOP-13428. Fixed hadoop-common build files to generate jdiff by working around a jdiff-bug. Contributed by Wangda Tan.
Repository: hadoop Updated Branches: refs/heads/branch-3.0.0-alpha1 25f23254e -> 67a4d3cce http://git-wip-us.apache.org/repos/asf/hadoop/blob/67a4d3cc/hadoop-common-project/hadoop-common/pom.xml -- diff --git a/hadoop-common-project/hadoop-common/pom.xml b/hadoop-common-project/hadoop-common/pom.xml index c13eec4..3885c70 100644 --- a/hadoop-common-project/hadoop-common/pom.xml +++ b/hadoop-common-project/hadoop-common/pom.xml @@ -34,6 +34,7 @@ src/test/resources/kdc common true +true ../etc/hadoop wsce-site.xml @@ -507,6 +508,7 @@ src/test/resources/test.har/_masterindex src/test/resources/test.har/part-0 src/test/resources/javakeystoreprovider.password +dev-support/jdiff-workaround.patch http://git-wip-us.apache.org/repos/asf/hadoop/blob/67a4d3cc/hadoop-project-dist/pom.xml -- diff --git a/hadoop-project-dist/pom.xml b/hadoop-project-dist/pom.xml index e386ff7..577923f 100644 --- a/hadoop-project-dist/pom.xml +++ b/hadoop-project-dist/pom.xml @@ -160,7 +160,7 @@ false -2.6.0 +2.7.2 -unstable @@ -173,7 +173,7 @@ org.apache.maven.plugins maven-javadoc-plugin - + javadoc @@ -241,6 +241,26 @@ org.apache.maven.plugins maven-antrun-plugin + + + +pre-site +prepare-package + + run + + + + + + + + + + + + site prepare-package @@ -249,7 +269,6 @@ - @@ -302,6 +321,25 @@ + + + +post-site +prepare-package + + run + + + + + + + + + + + + - To unsubscribe, e-mail: common-commits-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-commits-h...@hadoop.apache.org
[2/3] hadoop git commit: HADOOP-13428. Fixed hadoop-common build files to generate jdiff by working around a jdiff-bug. Contributed by Wangda Tan.
http://git-wip-us.apache.org/repos/asf/hadoop/blob/67a4d3cc/hadoop-common-project/hadoop-common/dev-support/jdiff/Apache_Hadoop_Common_2.7.2.xml -- diff --git a/hadoop-common-project/hadoop-common/dev-support/jdiff/Apache_Hadoop_Common_2.7.2.xml b/hadoop-common-project/hadoop-common/dev-support/jdiff/Apache_Hadoop_Common_2.7.2.xml new file mode 100644 index 000..5ef99b2 --- /dev/null +++ b/hadoop-common-project/hadoop-common/dev-support/jdiff/Apache_Hadoop_Common_2.7.2.xml @@ -0,0 +1,46648 @@ + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
[1/3] hadoop git commit: HADOOP-13428. Fixed hadoop-common build files to generate jdiff by working around a jdiff-bug. Contributed by Wangda Tan.
Repository: hadoop Updated Branches: refs/heads/branch-2.8 2e9d84d24 -> b2efb2040 http://git-wip-us.apache.org/repos/asf/hadoop/blob/b2efb204/hadoop-common-project/hadoop-common/pom.xml -- diff --git a/hadoop-common-project/hadoop-common/pom.xml b/hadoop-common-project/hadoop-common/pom.xml index 997a17f..9fc4afc 100644 --- a/hadoop-common-project/hadoop-common/pom.xml +++ b/hadoop-common-project/hadoop-common/pom.xml @@ -33,6 +33,7 @@ common true +true ../etc/hadoop wsce-site.xml @@ -493,6 +494,7 @@ src/test/resources/test.har/_masterindex src/test/resources/test.har/part-0 src/test/resources/javakeystoreprovider.password +dev-support/jdiff-workaround.patch http://git-wip-us.apache.org/repos/asf/hadoop/blob/b2efb204/hadoop-project-dist/pom.xml -- diff --git a/hadoop-project-dist/pom.xml b/hadoop-project-dist/pom.xml index 4cadfa5..123c04f 100644 --- a/hadoop-project-dist/pom.xml +++ b/hadoop-project-dist/pom.xml @@ -158,7 +158,7 @@ false -2.6.0 +2.7.2 -unstable @@ -171,7 +171,7 @@ org.apache.maven.plugins maven-javadoc-plugin - + javadoc @@ -239,6 +239,26 @@ org.apache.maven.plugins maven-antrun-plugin + + + +pre-site +prepare-package + + run + + + + + + + + + + + + site prepare-package @@ -247,7 +267,6 @@ - @@ -300,6 +319,25 @@ + + + +post-site +prepare-package + + run + + + + + + + + + + + + - To unsubscribe, e-mail: common-commits-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-commits-h...@hadoop.apache.org
[2/3] hadoop git commit: HADOOP-13428. Fixed hadoop-common build files to generate jdiff by working around a jdiff-bug. Contributed by Wangda Tan.
http://git-wip-us.apache.org/repos/asf/hadoop/blob/b2efb204/hadoop-common-project/hadoop-common/dev-support/jdiff/Apache_Hadoop_Common_2.7.2.xml -- diff --git a/hadoop-common-project/hadoop-common/dev-support/jdiff/Apache_Hadoop_Common_2.7.2.xml b/hadoop-common-project/hadoop-common/dev-support/jdiff/Apache_Hadoop_Common_2.7.2.xml new file mode 100644 index 000..5ef99b2 --- /dev/null +++ b/hadoop-common-project/hadoop-common/dev-support/jdiff/Apache_Hadoop_Common_2.7.2.xml @@ -0,0 +1,46648 @@ + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
[3/3] hadoop git commit: HADOOP-13428. Fixed hadoop-common build files to generate jdiff by working around a jdiff-bug. Contributed by Wangda Tan.
HADOOP-13428. Fixed hadoop-common build files to generate jdiff by working around a jdiff-bug. Contributed by Wangda Tan. (cherry picked from commit 99603e902244f17b04cfd55122f47355d070b588) Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/b2efb204 Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/b2efb204 Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/b2efb204 Branch: refs/heads/branch-2.8 Commit: b2efb2040cfd2313f767d028f7605bfdbae68381 Parents: 2e9d84d Author: Vinod Kumar Vavilapalli (I am also known as @tshooter.)Authored: Fri Aug 19 19:08:53 2016 -0700 Committer: Vinod Kumar Vavilapalli (I am also known as @tshooter.) Committed: Fri Aug 19 19:11:04 2016 -0700 -- .../dev-support/jdiff-workaround.patch |98 + .../jdiff/Apache_Hadoop_Common_2.7.2.xml| 46648 + hadoop-common-project/hadoop-common/pom.xml | 2 + hadoop-project-dist/pom.xml |44 +- 4 files changed, 46789 insertions(+), 3 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hadoop/blob/b2efb204/hadoop-common-project/hadoop-common/dev-support/jdiff-workaround.patch -- diff --git a/hadoop-common-project/hadoop-common/dev-support/jdiff-workaround.patch b/hadoop-common-project/hadoop-common/dev-support/jdiff-workaround.patch new file mode 100644 index 000..8f87d40 --- /dev/null +++ b/hadoop-common-project/hadoop-common/dev-support/jdiff-workaround.patch @@ -0,0 +1,98 @@ +diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/MetricsSystem.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/MetricsSystem.java +index a277abd..ed7c709 100644 +--- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/MetricsSystem.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/MetricsSystem.java +@@ -43,18 +43,6 @@ + public abstract MetricsSystem init(String prefix); + + /** +- * Register a metrics source +- * @paramthe actual type of the source object +- * @param source object to register +- * @param name of the source. Must be unique or null (then extracted from +- * the annotations of the source object.) +- * @param desc the description of the source (or null. See above.) +- * @return the source object +- * @exception MetricsException +- */ +- public abstract T register(String name, String desc, T source); +- +- /** +* Unregister a metrics source +* @param name of the source. This is the name you use to call register() +*/ +@@ -77,18 +65,19 @@ +*/ + @InterfaceAudience.Private + public abstract MetricsSource getSource(String name); ++ + + /** +- * Register a metrics sink +- * @paramthe type of the sink +- * @param sink to register +- * @param name of the sink. Must be unique. +- * @param desc the description of the sink +- * @return the sink ++ * Register a metrics source ++ * @paramthe actual type of the source object ++ * @param source object to register ++ * @param name of the source. Must be unique or null (then extracted from ++ * the annotations of the source object.) ++ * @param desc the description of the source (or null. See above.) ++ * @return the source object +* @exception MetricsException +*/ +- public abstract +- T register(String name, String desc, T sink); ++ public abstract T register(String name, String desc, T source); + + /** +* Register a callback interface for JMX events +diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/impl/MetricsSystemImpl.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/impl/MetricsSystemImpl.java +index 6986edb..eeea81f 100644 +--- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/impl/MetricsSystemImpl.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/impl/MetricsSystemImpl.java +@@ -270,27 +270,6 @@ void registerSource(String name, String desc, MetricsSource source) { + LOG.debug("Registered source "+ name); + } + +- @Override public synchronized +- T register(final String name, final String description, final T sink) { +-LOG.debug(name +", "+ description); +-if (allSinks.containsKey(name)) { +- LOG.warn("Sink "+ name +" already exists!"); +- return sink; +-} +-allSinks.put(name, sink); +-if (config != null) { +- registerSink(name, description, sink); +-} +-// We want to re-register the sink to pick up new config +-// when the
[2/3] hadoop git commit: HADOOP-13428. Fixed hadoop-common build files to generate jdiff by working around a jdiff-bug. Contributed by Wangda Tan.
http://git-wip-us.apache.org/repos/asf/hadoop/blob/a299146b/hadoop-common-project/hadoop-common/dev-support/jdiff/Apache_Hadoop_Common_2.7.2.xml -- diff --git a/hadoop-common-project/hadoop-common/dev-support/jdiff/Apache_Hadoop_Common_2.7.2.xml b/hadoop-common-project/hadoop-common/dev-support/jdiff/Apache_Hadoop_Common_2.7.2.xml new file mode 100644 index 000..5ef99b2 --- /dev/null +++ b/hadoop-common-project/hadoop-common/dev-support/jdiff/Apache_Hadoop_Common_2.7.2.xml @@ -0,0 +1,46648 @@ + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
[1/3] hadoop git commit: HADOOP-13428. Fixed hadoop-common build files to generate jdiff by working around a jdiff-bug. Contributed by Wangda Tan.
Repository: hadoop Updated Branches: refs/heads/branch-2 3e511f39c -> a299146b6 http://git-wip-us.apache.org/repos/asf/hadoop/blob/a299146b/hadoop-common-project/hadoop-common/pom.xml -- diff --git a/hadoop-common-project/hadoop-common/pom.xml b/hadoop-common-project/hadoop-common/pom.xml index c804fca..09f630c 100644 --- a/hadoop-common-project/hadoop-common/pom.xml +++ b/hadoop-common-project/hadoop-common/pom.xml @@ -33,6 +33,7 @@ common true +true ../etc/hadoop wsce-site.xml @@ -493,6 +494,7 @@ src/test/resources/test.har/_masterindex src/test/resources/test.har/part-0 src/test/resources/javakeystoreprovider.password +dev-support/jdiff-workaround.patch http://git-wip-us.apache.org/repos/asf/hadoop/blob/a299146b/hadoop-project-dist/pom.xml -- diff --git a/hadoop-project-dist/pom.xml b/hadoop-project-dist/pom.xml index b2e2c4f..c6cff92 100644 --- a/hadoop-project-dist/pom.xml +++ b/hadoop-project-dist/pom.xml @@ -158,7 +158,7 @@ false -2.6.0 +2.7.2 -unstable @@ -171,7 +171,7 @@ org.apache.maven.plugins maven-javadoc-plugin - + javadoc @@ -239,6 +239,26 @@ org.apache.maven.plugins maven-antrun-plugin + + + +pre-site +prepare-package + + run + + + + + + + + + + + + site prepare-package @@ -247,7 +267,6 @@ - @@ -300,6 +319,25 @@ + + + +post-site +prepare-package + + run + + + + + + + + + + + + - To unsubscribe, e-mail: common-commits-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-commits-h...@hadoop.apache.org
[3/3] hadoop git commit: HADOOP-13428. Fixed hadoop-common build files to generate jdiff by working around a jdiff-bug. Contributed by Wangda Tan.
HADOOP-13428. Fixed hadoop-common build files to generate jdiff by working around a jdiff-bug. Contributed by Wangda Tan. (cherry picked from commit 99603e902244f17b04cfd55122f47355d070b588) Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/a299146b Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/a299146b Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/a299146b Branch: refs/heads/branch-2 Commit: a299146b6d9ed3c558f7d4e5bc33d492362faffd Parents: 3e511f3 Author: Vinod Kumar Vavilapalli (I am also known as @tshooter.)Authored: Fri Aug 19 19:08:53 2016 -0700 Committer: Vinod Kumar Vavilapalli (I am also known as @tshooter.) Committed: Fri Aug 19 19:10:35 2016 -0700 -- .../dev-support/jdiff-workaround.patch |98 + .../jdiff/Apache_Hadoop_Common_2.7.2.xml| 46648 + hadoop-common-project/hadoop-common/pom.xml | 2 + hadoop-project-dist/pom.xml |44 +- 4 files changed, 46789 insertions(+), 3 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hadoop/blob/a299146b/hadoop-common-project/hadoop-common/dev-support/jdiff-workaround.patch -- diff --git a/hadoop-common-project/hadoop-common/dev-support/jdiff-workaround.patch b/hadoop-common-project/hadoop-common/dev-support/jdiff-workaround.patch new file mode 100644 index 000..8f87d40 --- /dev/null +++ b/hadoop-common-project/hadoop-common/dev-support/jdiff-workaround.patch @@ -0,0 +1,98 @@ +diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/MetricsSystem.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/MetricsSystem.java +index a277abd..ed7c709 100644 +--- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/MetricsSystem.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/MetricsSystem.java +@@ -43,18 +43,6 @@ + public abstract MetricsSystem init(String prefix); + + /** +- * Register a metrics source +- * @paramthe actual type of the source object +- * @param source object to register +- * @param name of the source. Must be unique or null (then extracted from +- * the annotations of the source object.) +- * @param desc the description of the source (or null. See above.) +- * @return the source object +- * @exception MetricsException +- */ +- public abstract T register(String name, String desc, T source); +- +- /** +* Unregister a metrics source +* @param name of the source. This is the name you use to call register() +*/ +@@ -77,18 +65,19 @@ +*/ + @InterfaceAudience.Private + public abstract MetricsSource getSource(String name); ++ + + /** +- * Register a metrics sink +- * @paramthe type of the sink +- * @param sink to register +- * @param name of the sink. Must be unique. +- * @param desc the description of the sink +- * @return the sink ++ * Register a metrics source ++ * @paramthe actual type of the source object ++ * @param source object to register ++ * @param name of the source. Must be unique or null (then extracted from ++ * the annotations of the source object.) ++ * @param desc the description of the source (or null. See above.) ++ * @return the source object +* @exception MetricsException +*/ +- public abstract +- T register(String name, String desc, T sink); ++ public abstract T register(String name, String desc, T source); + + /** +* Register a callback interface for JMX events +diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/impl/MetricsSystemImpl.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/impl/MetricsSystemImpl.java +index 6986edb..eeea81f 100644 +--- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/impl/MetricsSystemImpl.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/impl/MetricsSystemImpl.java +@@ -270,27 +270,6 @@ void registerSource(String name, String desc, MetricsSource source) { + LOG.debug("Registered source "+ name); + } + +- @Override public synchronized +- T register(final String name, final String description, final T sink) { +-LOG.debug(name +", "+ description); +-if (allSinks.containsKey(name)) { +- LOG.warn("Sink "+ name +" already exists!"); +- return sink; +-} +-allSinks.put(name, sink); +-if (config != null) { +- registerSink(name, description, sink); +-} +-// We want to re-register the sink to pick up new config +-// when the
[3/3] hadoop git commit: HADOOP-13428. Fixed hadoop-common build files to generate jdiff by working around a jdiff-bug. Contributed by Wangda Tan.
HADOOP-13428. Fixed hadoop-common build files to generate jdiff by working around a jdiff-bug. Contributed by Wangda Tan. Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/99603e90 Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/99603e90 Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/99603e90 Branch: refs/heads/trunk Commit: 99603e902244f17b04cfd55122f47355d070b588 Parents: 2da32a6 Author: Vinod Kumar Vavilapalli (I am also known as @tshooter.)Authored: Fri Aug 19 19:08:53 2016 -0700 Committer: Vinod Kumar Vavilapalli (I am also known as @tshooter.) Committed: Fri Aug 19 19:08:53 2016 -0700 -- .../dev-support/jdiff-workaround.patch |98 + .../jdiff/Apache_Hadoop_Common_2.7.2.xml| 46648 + hadoop-common-project/hadoop-common/pom.xml | 2 + hadoop-project-dist/pom.xml |44 +- 4 files changed, 46789 insertions(+), 3 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hadoop/blob/99603e90/hadoop-common-project/hadoop-common/dev-support/jdiff-workaround.patch -- diff --git a/hadoop-common-project/hadoop-common/dev-support/jdiff-workaround.patch b/hadoop-common-project/hadoop-common/dev-support/jdiff-workaround.patch new file mode 100644 index 000..8f87d40 --- /dev/null +++ b/hadoop-common-project/hadoop-common/dev-support/jdiff-workaround.patch @@ -0,0 +1,98 @@ +diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/MetricsSystem.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/MetricsSystem.java +index a277abd..ed7c709 100644 +--- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/MetricsSystem.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/MetricsSystem.java +@@ -43,18 +43,6 @@ + public abstract MetricsSystem init(String prefix); + + /** +- * Register a metrics source +- * @paramthe actual type of the source object +- * @param source object to register +- * @param name of the source. Must be unique or null (then extracted from +- * the annotations of the source object.) +- * @param desc the description of the source (or null. See above.) +- * @return the source object +- * @exception MetricsException +- */ +- public abstract T register(String name, String desc, T source); +- +- /** +* Unregister a metrics source +* @param name of the source. This is the name you use to call register() +*/ +@@ -77,18 +65,19 @@ +*/ + @InterfaceAudience.Private + public abstract MetricsSource getSource(String name); ++ + + /** +- * Register a metrics sink +- * @paramthe type of the sink +- * @param sink to register +- * @param name of the sink. Must be unique. +- * @param desc the description of the sink +- * @return the sink ++ * Register a metrics source ++ * @paramthe actual type of the source object ++ * @param source object to register ++ * @param name of the source. Must be unique or null (then extracted from ++ * the annotations of the source object.) ++ * @param desc the description of the source (or null. See above.) ++ * @return the source object +* @exception MetricsException +*/ +- public abstract +- T register(String name, String desc, T sink); ++ public abstract T register(String name, String desc, T source); + + /** +* Register a callback interface for JMX events +diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/impl/MetricsSystemImpl.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/impl/MetricsSystemImpl.java +index 6986edb..eeea81f 100644 +--- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/impl/MetricsSystemImpl.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/impl/MetricsSystemImpl.java +@@ -270,27 +270,6 @@ void registerSource(String name, String desc, MetricsSource source) { + LOG.debug("Registered source "+ name); + } + +- @Override public synchronized +- T register(final String name, final String description, final T sink) { +-LOG.debug(name +", "+ description); +-if (allSinks.containsKey(name)) { +- LOG.warn("Sink "+ name +" already exists!"); +- return sink; +-} +-allSinks.put(name, sink); +-if (config != null) { +- registerSink(name, description, sink); +-} +-// We want to re-register the sink to pick up new config +-// when the metrics system restarts. +-register(name, new AbstractCallback() { +-
[1/3] hadoop git commit: HADOOP-13428. Fixed hadoop-common build files to generate jdiff by working around a jdiff-bug. Contributed by Wangda Tan.
Repository: hadoop Updated Branches: refs/heads/trunk 2da32a6ef -> 99603e902 http://git-wip-us.apache.org/repos/asf/hadoop/blob/99603e90/hadoop-common-project/hadoop-common/pom.xml -- diff --git a/hadoop-common-project/hadoop-common/pom.xml b/hadoop-common-project/hadoop-common/pom.xml index c28a05c..54d1cdd 100644 --- a/hadoop-common-project/hadoop-common/pom.xml +++ b/hadoop-common-project/hadoop-common/pom.xml @@ -34,6 +34,7 @@ src/test/resources/kdc common true +true ../etc/hadoop wsce-site.xml @@ -512,6 +513,7 @@ src/test/resources/test.har/_masterindex src/test/resources/test.har/part-0 src/test/resources/javakeystoreprovider.password +dev-support/jdiff-workaround.patch http://git-wip-us.apache.org/repos/asf/hadoop/blob/99603e90/hadoop-project-dist/pom.xml -- diff --git a/hadoop-project-dist/pom.xml b/hadoop-project-dist/pom.xml index 0357269..0ee9895 100644 --- a/hadoop-project-dist/pom.xml +++ b/hadoop-project-dist/pom.xml @@ -160,7 +160,7 @@ false -2.6.0 +2.7.2 -unstable @@ -173,7 +173,7 @@ org.apache.maven.plugins maven-javadoc-plugin - + javadoc @@ -241,6 +241,26 @@ org.apache.maven.plugins maven-antrun-plugin + + + +pre-site +prepare-package + + run + + + + + + + + + + + + site prepare-package @@ -249,7 +269,6 @@ - @@ -302,6 +321,25 @@ + + + +post-site +prepare-package + + run + + + + + + + + + + + + - To unsubscribe, e-mail: common-commits-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-commits-h...@hadoop.apache.org
[2/3] hadoop git commit: HADOOP-13428. Fixed hadoop-common build files to generate jdiff by working around a jdiff-bug. Contributed by Wangda Tan.
http://git-wip-us.apache.org/repos/asf/hadoop/blob/99603e90/hadoop-common-project/hadoop-common/dev-support/jdiff/Apache_Hadoop_Common_2.7.2.xml -- diff --git a/hadoop-common-project/hadoop-common/dev-support/jdiff/Apache_Hadoop_Common_2.7.2.xml b/hadoop-common-project/hadoop-common/dev-support/jdiff/Apache_Hadoop_Common_2.7.2.xml new file mode 100644 index 000..5ef99b2 --- /dev/null +++ b/hadoop-common-project/hadoop-common/dev-support/jdiff/Apache_Hadoop_Common_2.7.2.xml @@ -0,0 +1,46648 @@ + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
[2/2] hadoop git commit: YARN-4307. Display blacklisted nodes for AM container in the RM web UI. Contributed by Naganarasimha G R.
YARN-4307. Display blacklisted nodes for AM container in the RM web UI. Contributed by Naganarasimha G R. (cherry picked from commit 308d63f382e1992ea2b8ccf3130edaaa751c644d) (cherry picked from commit 99966b1912320326fb63be1ce6d4222d736ca365) Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/5b49bcae Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/5b49bcae Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/5b49bcae Branch: refs/heads/branch-2.8 Commit: 5b49bcaee326a499d530f4ae8fb577ab010ccd51 Parents: 42a88f9 Author: Varun VasudevAuthored: Thu Feb 4 13:32:54 2016 +0530 Committer: Vinod Kumar Vavilapalli (I am also known as @tshooter.) Committed: Thu Aug 18 12:33:52 2016 -0700 -- .../hadoop/yarn/webapp/hamlet/Hamlet.java | 11 - .../blacklist/SimpleBlacklistManager.java | 9 ++-- .../rmapp/attempt/RMAppAttempt.java | 6 +++ .../rmapp/attempt/RMAppAttemptImpl.java | 17 .../webapp/RMAppAttemptBlock.java | 46 .../resourcemanager/webapp/RMAppBlock.java | 32 +++--- .../resourcemanager/webapp/RMAppsBlock.java | 5 ++- .../webapp/dao/AppAttemptInfo.java | 6 ++- .../webapp/TestRMWebServicesApps.java | 2 +- 9 files changed, 80 insertions(+), 54 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hadoop/blob/5b49bcae/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/webapp/hamlet/Hamlet.java -- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/webapp/hamlet/Hamlet.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/webapp/hamlet/Hamlet.java index 83fa5da..58d9066 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/webapp/hamlet/Hamlet.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/webapp/hamlet/Hamlet.java @@ -18,10 +18,13 @@ // Generated by HamletGen. Do NOT edit! package org.apache.hadoop.yarn.webapp.hamlet; +import static java.util.EnumSet.of; +import static org.apache.hadoop.yarn.webapp.hamlet.HamletImpl.EOpt.ENDTAG; +import static org.apache.hadoop.yarn.webapp.hamlet.HamletImpl.EOpt.INLINE; +import static org.apache.hadoop.yarn.webapp.hamlet.HamletImpl.EOpt.PRE; + import java.io.PrintWriter; import java.util.EnumSet; -import static java.util.EnumSet.*; -import static org.apache.hadoop.yarn.webapp.hamlet.HamletImpl.EOpt.*; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.yarn.webapp.SubView; @@ -2405,6 +2408,10 @@ public class Hamlet extends HamletImpl implements HamletSpec._Html { return setSelector(th(), selector)._(cdata)._(); } +public TR th(String selector, String title, String cdata) { + return setSelector(th(), selector).$title(title)._(cdata)._(); +} + @Override public TD td() { closeAttrs(); http://git-wip-us.apache.org/repos/asf/hadoop/blob/5b49bcae/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/blacklist/SimpleBlacklistManager.java -- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/blacklist/SimpleBlacklistManager.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/blacklist/SimpleBlacklistManager.java index a544ab8..c7bd0f8 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/blacklist/SimpleBlacklistManager.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/blacklist/SimpleBlacklistManager.java @@ -72,11 +72,10 @@ public class SimpleBlacklistManager implements BlacklistManager { } ret = new BlacklistUpdates(blacklist, EMPTY_LIST); } else { - if (LOG.isDebugEnabled()) { -LOG.debug("blacklist size " + currentBlacklistSize + " is more than " + -"failure threshold ratio " + blacklistDisableFailureThreshold + -" out of total usable nodes " + numberOfNodeManagerHosts); - } + LOG.warn("Ignoring Blacklists, blacklist size " + currentBlacklistSize + + "
[1/2] hadoop git commit: YARN-3446. FairScheduler headroom calculation should exclude nodes in the blacklist. (Zhihai Xu via kasha)
Repository: hadoop Updated Branches: refs/heads/branch-2.8 1c4e039ba -> 5b49bcaee YARN-3446. FairScheduler headroom calculation should exclude nodes in the blacklist. (Zhihai Xu via kasha) (cherry picked from commit 9d04f26d4c42170ee3dab2f6fb09a94bbf72fc65) (cherry picked from commit f0923819c35b8c499a9bcd0c6c78511f3a400fef) Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/42a88f9c Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/42a88f9c Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/42a88f9c Branch: refs/heads/branch-2.8 Commit: 42a88f9c5894265a9bd5584ea12b35a415d25307 Parents: 1c4e039 Author: Karthik KambatlaAuthored: Thu Jan 14 08:33:23 2016 -0800 Committer: Vinod Kumar Vavilapalli (I am also known as @tshooter.) Committed: Thu Aug 18 12:30:39 2016 -0700 -- .../scheduler/AbstractYarnScheduler.java| 15 .../scheduler/AppSchedulingInfo.java| 51 +++- .../scheduler/fair/FSAppAttempt.java| 25 ++ .../scheduler/TestAppSchedulingInfo.java| 73 .../scheduler/fair/TestFSAppAttempt.java| 88 ++-- 5 files changed, 225 insertions(+), 27 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hadoop/blob/42a88f9c/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/AbstractYarnScheduler.java -- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/AbstractYarnScheduler.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/AbstractYarnScheduler.java index cfe2a7a..5245bb6 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/AbstractYarnScheduler.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/AbstractYarnScheduler.java @@ -180,6 +180,21 @@ public abstract class AbstractYarnScheduler return applications; } + /** + * Add blacklisted NodeIds to the list that is passed. + * + * @param app application attempt. + * @param blacklistNodeIdList the list to store blacklisted NodeIds. + */ + public void addBlacklistedNodeIdsToList(SchedulerApplicationAttempt app, + List blacklistNodeIdList) { +for (Map.Entry nodeEntry : nodes.entrySet()) { + if (SchedulerAppUtils.isBlacklisted(app, nodeEntry.getValue(), LOG)) { +blacklistNodeIdList.add(nodeEntry.getKey()); + } +} + } + @Override public Resource getClusterResource() { return clusterResource; http://git-wip-us.apache.org/repos/asf/hadoop/blob/42a88f9c/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/AppSchedulingInfo.java -- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/AppSchedulingInfo.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/AppSchedulingInfo.java index 9b2ba14..5952cc2 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/AppSchedulingInfo.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/AppSchedulingInfo.java @@ -18,6 +18,20 @@ package org.apache.hadoop.yarn.server.resourcemanager.scheduler; +import java.util.ArrayList; +import java.util.Collection; +import java.util.Comparator; +import java.util.HashMap; +import java.util.HashSet; +import java.util.List; +import java.util.Map; +import java.util.Set; +import java.util.TreeMap; +import java.util.TreeSet; +import java.util.concurrent.ConcurrentHashMap; +import java.util.concurrent.atomic.AtomicBoolean; +import java.util.concurrent.atomic.AtomicLong; + import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.classification.InterfaceAudience.Private;
[hadoop] Git Push Summary
Repository: hadoop Updated Tags: refs/tags/release-2.7.3-RC2 [created] e05a76e64 - To unsubscribe, e-mail: common-commits-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-commits-h...@hadoop.apache.org
hadoop git commit: Set the release date for 2.7.3-RC2
Repository: hadoop Updated Branches: refs/heads/branch-2.7.3 51187688b -> baa91f7c6 Set the release date for 2.7.3-RC2 Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/baa91f7c Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/baa91f7c Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/baa91f7c Branch: refs/heads/branch-2.7.3 Commit: baa91f7c6bc9cb92be5982de4719c1c8af91ccff Parents: 5118768 Author: Vinod Kumar Vavilapalli (I am also known as @tshooter.)Authored: Wed Aug 17 17:55:21 2016 -0700 Committer: Vinod Kumar Vavilapalli (I am also known as @tshooter.) Committed: Wed Aug 17 17:55:21 2016 -0700 -- hadoop-common-project/hadoop-common/CHANGES.txt | 2 +- hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt | 2 +- hadoop-mapreduce-project/CHANGES.txt| 2 +- hadoop-yarn-project/CHANGES.txt | 2 +- 4 files changed, 4 insertions(+), 4 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hadoop/blob/baa91f7c/hadoop-common-project/hadoop-common/CHANGES.txt -- diff --git a/hadoop-common-project/hadoop-common/CHANGES.txt b/hadoop-common-project/hadoop-common/CHANGES.txt index 3460ccae..6d97302 100644 --- a/hadoop-common-project/hadoop-common/CHANGES.txt +++ b/hadoop-common-project/hadoop-common/CHANGES.txt @@ -1,6 +1,6 @@ Hadoop Change Log -Release 2.7.3 - 2016-08-19 +Release 2.7.3 - 2016-08-25 INCOMPATIBLE CHANGES http://git-wip-us.apache.org/repos/asf/hadoop/blob/baa91f7c/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt -- diff --git a/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt b/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt index a243834..db64d3c 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt +++ b/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt @@ -1,6 +1,6 @@ Hadoop HDFS Change Log -Release 2.7.3 - 2016-08-19 +Release 2.7.3 - 2016-08-25 INCOMPATIBLE CHANGES http://git-wip-us.apache.org/repos/asf/hadoop/blob/baa91f7c/hadoop-mapreduce-project/CHANGES.txt -- diff --git a/hadoop-mapreduce-project/CHANGES.txt b/hadoop-mapreduce-project/CHANGES.txt index 26b71a5..389d36f 100644 --- a/hadoop-mapreduce-project/CHANGES.txt +++ b/hadoop-mapreduce-project/CHANGES.txt @@ -1,6 +1,6 @@ Hadoop MapReduce Change Log -Release 2.7.3 - 2016-08-19 +Release 2.7.3 - 2016-08-25 INCOMPATIBLE CHANGES http://git-wip-us.apache.org/repos/asf/hadoop/blob/baa91f7c/hadoop-yarn-project/CHANGES.txt -- diff --git a/hadoop-yarn-project/CHANGES.txt b/hadoop-yarn-project/CHANGES.txt index 9f8b014..5f88bc1 100644 --- a/hadoop-yarn-project/CHANGES.txt +++ b/hadoop-yarn-project/CHANGES.txt @@ -1,6 +1,6 @@ Hadoop YARN Change Log -Release 2.7.3 - 2016-08-19 +Release 2.7.3 - 2016-08-25 INCOMPATIBLE CHANGES - To unsubscribe, e-mail: common-commits-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-commits-h...@hadoop.apache.org
[2/2] hadoop git commit: Preparing for release 2.7.3: Updating release notes.
Preparing for release 2.7.3: Updating release notes. Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/51187688 Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/51187688 Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/51187688 Branch: refs/heads/branch-2.7.3 Commit: 51187688bf92731444d5e05e038c3706adec8c16 Parents: 2f1a438 Author: Vinod Kumar Vavilapalli (I am also known as @tshooter.)Authored: Wed Aug 17 14:45:05 2016 -0700 Committer: Vinod Kumar Vavilapalli (I am also known as @tshooter.) Committed: Wed Aug 17 17:50:29 2016 -0700 -- .../src/main/docs/releasenotes.html | 20 ++-- 1 file changed, 14 insertions(+), 6 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hadoop/blob/51187688/hadoop-common-project/hadoop-common/src/main/docs/releasenotes.html -- diff --git a/hadoop-common-project/hadoop-common/src/main/docs/releasenotes.html b/hadoop-common-project/hadoop-common/src/main/docs/releasenotes.html index 052da7c..8251cd9 100644 --- a/hadoop-common-project/hadoop-common/src/main/docs/releasenotes.html +++ b/hadoop-common-project/hadoop-common/src/main/docs/releasenotes.html @@ -220,6 +220,10 @@ These release notes include new developer and user-facing incompatibilities, fea Minor test reported by zhihai xu and fixed by zhihai xu (test) TestResourceLocalizationService.testPublicResourceInitializesLocalDir fails Intermittently due to IOException from cleanup + https://issues.apache.org/jira/browse/YARN-3426;>YARN-3426. + Blocker sub-task reported by Li Lu and fixed by Li Lu + Add jdiff support to YARN + https://issues.apache.org/jira/browse/YARN-3404;>YARN-3404. Minor improvement reported by Ryu Kobayashi and fixed by Ryu Kobayashi View the queue name to YARN Application page @@ -556,7 +560,7 @@ In this improvement, we add a new API isCommitJobRepeatable() to OutputCommitter seen_txid in the shared edits directory is modified during bootstrapping https://issues.apache.org/jira/browse/HDFS-9530;>HDFS-9530. - Critical bug reported by Fei Hui and fixed by Brahma Reddy Battula + Critical bug reported by Fei Hui and fixed by Brahma Reddy Battula (datanode) ReservedSpace is not cleared for abandoned Blocks https://issues.apache.org/jira/browse/HDFS-9516;>HDFS-9516. @@ -583,17 +587,13 @@ In this improvement, we add a new API isCommitJobRepeatable() to OutputCommitter Major bug reported by Stanislav Antic and fixed by Yongjun Zhang (namenode) FSImage may get corrupted after deleting snapshot - https://issues.apache.org/jira/browse/HDFS-9395;>HDFS-9395. - Major bug reported by Kihwal Lee and fixed by Kuhu Shukla - Make HDFS audit logging consistant - https://issues.apache.org/jira/browse/HDFS-9383;>HDFS-9383. Major bug reported by Kihwal Lee and fixed by Tsz Wo Nicholas Sze TestByteArrayManager#testByteArrayManager fails https://issues.apache.org/jira/browse/HDFS-9365;>HDFS-9365. Major bug reported by Tsz Wo Nicholas Sze and fixed by Tsz Wo Nicholas Sze (balancer mover) - Balaner does not work with the HDFS-6376 HA setup + Balancer does not work with the HDFS-6376 HA setup https://issues.apache.org/jira/browse/HDFS-9347;>HDFS-9347. Major bug reported by Wei-Chiu Chuang and fixed by Wei-Chiu Chuang (test) @@ -671,6 +671,10 @@ In this improvement, we add a new API isCommitJobRepeatable() to OutputCommitter Major bug reported by Aaron T. Myers and fixed by Yiqun Lin (test) TestHFlush failing intermittently + https://issues.apache.org/jira/browse/HADOOP-13434;>HADOOP-13434. + Major bug reported by Owen O'Malley and fixed by Owen O'Malley + Add quoting to Shell class + https://issues.apache.org/jira/browse/HADOOP-13350;>HADOOP-13350. Blocker bug reported by Xiao Chen and fixed by Xiao Chen (build) Additional fix to LICENSE and NOTICE @@ -875,6 +879,10 @@ In this improvement, we add a new API isCommitJobRepeatable() to OutputCommitter Critical bug reported by Sangjin Lee and fixed by Sangjin Lee (fs) long running apps may have a huge number of StatisticsData instances under FileSystem + https://issues.apache.org/jira/browse/HADOOP-11814;>HADOOP-11814. + Minor task reported by Li Lu and fixed by Li Lu + Reformat hadoop-annotations, o.a.h.classification.tools + https://issues.apache.org/jira/browse/HADOOP-11252;>HADOOP-11252. Critical bug reported by Wilfred Spiegelenburg and fixed by Masatake Iwasaki (ipc) RPC client does not time out by default
[1/2] hadoop git commit: HADOOP-13434. Add bash quoting to Shell class. (Owen O'Malley) Added the missing CHANGES.txt entry.
Repository: hadoop Updated Branches: refs/heads/branch-2.7.3 22fd150e9 -> 51187688b HADOOP-13434. Add bash quoting to Shell class. (Owen O'Malley) Added the missing CHANGES.txt entry. Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/2f1a4387 Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/2f1a4387 Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/2f1a4387 Branch: refs/heads/branch-2.7.3 Commit: 2f1a4387fdc7c48dfd117d86555bccac118d9e24 Parents: 22fd150 Author: Vinod Kumar Vavilapalli (I am also known as @tshooter.)Authored: Wed Aug 17 14:38:59 2016 -0700 Committer: Vinod Kumar Vavilapalli (I am also known as @tshooter.) Committed: Wed Aug 17 14:38:59 2016 -0700 -- hadoop-common-project/hadoop-common/CHANGES.txt | 2 ++ 1 file changed, 2 insertions(+) -- http://git-wip-us.apache.org/repos/asf/hadoop/blob/2f1a4387/hadoop-common-project/hadoop-common/CHANGES.txt -- diff --git a/hadoop-common-project/hadoop-common/CHANGES.txt b/hadoop-common-project/hadoop-common/CHANGES.txt index 3777e32..3460ccae 100644 --- a/hadoop-common-project/hadoop-common/CHANGES.txt +++ b/hadoop-common-project/hadoop-common/CHANGES.txt @@ -26,6 +26,8 @@ Release 2.7.3 - 2016-08-19 HADOOP-13298. Fix the leftover L files in hadoop-build-tools/src/main/resources/META-INF/. (ozawa) +HADOOP-13434. Added quoting to Shell class. (Owen O'Malley via Arpit Agarwal) + OPTIMIZATIONS HADOOP-12810. FileSystem#listLocatedStatus causes unnecessary RPC calls - To unsubscribe, e-mail: common-commits-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-commits-h...@hadoop.apache.org
hadoop git commit: Revert "HDFS-9395. Make HDFS audit logging consistant. Contributed by Kuhu Shukla."
Repository: hadoop Updated Branches: refs/heads/branch-2.7.3 7a4746cbb -> 22fd150e9 Revert "HDFS-9395. Make HDFS audit logging consistant. Contributed by Kuhu Shukla." Reverting this on branch-2.* as it's an incompatible change. This reverts commit 25b3531eb40a32a602574e5cfc1ffe028044bcc9. Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/22fd150e Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/22fd150e Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/22fd150e Branch: refs/heads/branch-2.7.3 Commit: 22fd150e99c0148d7e2c1cc77428c6181cc931ba Parents: 7a4746c Author: Vinod Kumar Vavilapalli (I am also known as @tshooter.)Authored: Wed Aug 17 13:41:15 2016 -0700 Committer: Vinod Kumar Vavilapalli (I am also known as @tshooter.) Committed: Wed Aug 17 13:41:15 2016 -0700 -- hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt | 2 - .../hdfs/server/namenode/FSNamesystem.java | 181 ++ .../namenode/TestAuditLoggerWithCommands.java | 562 --- 3 files changed, 59 insertions(+), 686 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hadoop/blob/22fd150e/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt -- diff --git a/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt b/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt index 748489d..a243834 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt +++ b/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt @@ -36,8 +36,6 @@ Release 2.7.3 - 2016-08-19 HDFS-8578. On upgrade, Datanode should process all storage/data dirs in parallel. (vinayakumarb and szetszwo via szetszwo) -HDFS-9395. Make HDFS audit logging consistant (Kuhu Shukla via kihwal) - HDFS-9048. DistCp documentation is out-of-dated (Daisuke Kobayashi via iwasakims) http://git-wip-us.apache.org/repos/asf/hadoop/blob/22fd150e/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java -- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java index 20824cf..81a4ea4 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java @@ -1883,16 +1883,13 @@ public class FSNamesystem implements Namesystem, FSNamesystemMBean, checkNameNodeSafeMode("Cannot concat " + target); stat = FSDirConcatOp.concat(dir, target, srcs, logRetryCache); success = true; -} catch (AccessControlException ace) { - logAuditEvent(success, "concat", Arrays.toString(srcs), target, stat); - throw ace; } finally { writeUnlock(); if (success) { getEditLog().logSync(); } + logAuditEvent(success, "concat", Arrays.toString(srcs), target, stat); } -logAuditEvent(success, "concat", Arrays.toString(srcs), target, stat); } /** @@ -3662,8 +3659,9 @@ public class FSNamesystem implements Namesystem, FSNamesystemMBean, boolean success = ret != null && ret.success; if (success) { getEditLog().logSync(); - logAuditEvent(success, "rename", src, dst, ret.auditStat); } +logAuditEvent(success, "rename", src, dst, +ret == null ? null : ret.auditStat); return success; } @@ -3925,19 +3923,16 @@ public class FSNamesystem implements Namesystem, FSNamesystemMBean, checkOperation(OperationCategory.READ); readLock(); boolean success = true; -ContentSummary cs; try { checkOperation(OperationCategory.READ); - cs = FSDirStatAndListingOp.getContentSummary(dir, src); + return FSDirStatAndListingOp.getContentSummary(dir, src); } catch (AccessControlException ace) { success = false; - logAuditEvent(success, "contentSummary", src); throw ace; } finally { readUnlock(); + logAuditEvent(success, "contentSummary", src); } -logAuditEvent(success, "contentSummary", src); -return cs; } /** @@ -3957,16 +3952,13 @@ public class FSNamesystem implements Namesystem, FSNamesystemMBean, checkNameNodeSafeMode("Cannot set quota on " + src); FSDirAttrOp.setQuota(dir, src, nsQuota, ssQuota, type); success = true; -} catch (AccessControlException ace) { - logAuditEvent(success, "setQuota", src); - throw ace; } finally { writeUnlock(); if (success) { getEditLog().logSync(); }
hadoop git commit: Revert "HDFS-9395. Make HDFS audit logging consistant. Contributed by Kuhu Shukla."
Repository: hadoop Updated Branches: refs/heads/branch-2.8 00eb79062 -> d65024edd Revert "HDFS-9395. Make HDFS audit logging consistant. Contributed by Kuhu Shukla." Reverting this on branch-2.* as it's an incompatible change. This reverts commit 83f7f62be379045ad6933689b21b76c7086f919d. (cherry picked from commit 2486c4c63a35fcef7338ea63f0d8aafa778cd05d) Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/d65024ed Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/d65024ed Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/d65024ed Branch: refs/heads/branch-2.8 Commit: d65024eddc5e6c0590e99e02a41a3845594ef69f Parents: 00eb790 Author: Vinod Kumar Vavilapalli (I am also known as @tshooter.)Authored: Wed Aug 17 13:28:00 2016 -0700 Committer: Vinod Kumar Vavilapalli (I am also known as @tshooter.) Committed: Wed Aug 17 13:34:35 2016 -0700 -- .../hdfs/server/namenode/FSNamesystem.java | 185 ++ .../namenode/TestAuditLoggerWithCommands.java | 585 --- 2 files changed, 60 insertions(+), 710 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hadoop/blob/d65024ed/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java -- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java index 6cba82e..a8dc8fa 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java @@ -1855,16 +1855,13 @@ public class FSNamesystem implements Namesystem, FSNamesystemMBean, checkNameNodeSafeMode("Cannot concat " + target); stat = FSDirConcatOp.concat(dir, target, srcs, logRetryCache); success = true; -} catch (AccessControlException ace) { - logAuditEvent(success, "concat", Arrays.toString(srcs), target, stat); - throw ace; } finally { writeUnlock(); if (success) { getEditLog().logSync(); } + logAuditEvent(success, "concat", Arrays.toString(srcs), target, stat); } -logAuditEvent(success, "concat", Arrays.toString(srcs), target, stat); } /** @@ -2740,8 +2737,9 @@ public class FSNamesystem implements Namesystem, FSNamesystemMBean, boolean success = ret != null && ret.success; if (success) { getEditLog().logSync(); - logAuditEvent(success, "rename", src, dst, ret.auditStat); } +logAuditEvent(success, "rename", src, dst, +ret == null ? null : ret.auditStat); return success; } @@ -3005,19 +3003,16 @@ public class FSNamesystem implements Namesystem, FSNamesystemMBean, checkOperation(OperationCategory.READ); readLock(); boolean success = true; -ContentSummary cs; try { checkOperation(OperationCategory.READ); - cs = FSDirStatAndListingOp.getContentSummary(dir, src); + return FSDirStatAndListingOp.getContentSummary(dir, src); } catch (AccessControlException ace) { success = false; - logAuditEvent(success, "contentSummary", src); throw ace; } finally { readUnlock(); + logAuditEvent(success, "contentSummary", src); } -logAuditEvent(success, "contentSummary", src); -return cs; } /** @@ -3036,21 +3031,18 @@ public class FSNamesystem implements Namesystem, FSNamesystemMBean, */ QuotaUsage getQuotaUsage(final String src) throws IOException { checkOperation(OperationCategory.READ); -QuotaUsage quotaUsage; readLock(); boolean success = true; try { checkOperation(OperationCategory.READ); - quotaUsage = FSDirStatAndListingOp.getQuotaUsage(dir, src); + return FSDirStatAndListingOp.getQuotaUsage(dir, src); } catch (AccessControlException ace) { success = false; - logAuditEvent(success, "quotaUsage", src); throw ace; } finally { readUnlock(); + logAuditEvent(success, "quotaUsage", src); } -logAuditEvent(success, "quotaUsage", src); -return quotaUsage; } /** @@ -3073,16 +3065,13 @@ public class FSNamesystem implements Namesystem, FSNamesystemMBean, checkNameNodeSafeMode("Cannot set quota on " + src); FSDirAttrOp.setQuota(dir, src, nsQuota, ssQuota, type); success = true; -} catch (AccessControlException ace) { - logAuditEvent(success, "setQuota", src); - throw ace; } finally { writeUnlock(); if (success) {
hadoop git commit: Revert "HDFS-9395. Make HDFS audit logging consistant. Contributed by Kuhu Shukla."
Repository: hadoop Updated Branches: refs/heads/branch-2 9dcb7197d -> 2486c4c63 Revert "HDFS-9395. Make HDFS audit logging consistant. Contributed by Kuhu Shukla." Reverting this on branch-2.* as it's an incompatible change. This reverts commit 83f7f62be379045ad6933689b21b76c7086f919d. Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/2486c4c6 Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/2486c4c6 Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/2486c4c6 Branch: refs/heads/branch-2 Commit: 2486c4c63a35fcef7338ea63f0d8aafa778cd05d Parents: 9dcb719 Author: Vinod Kumar Vavilapalli (I am also known as @tshooter.)Authored: Wed Aug 17 13:28:00 2016 -0700 Committer: Vinod Kumar Vavilapalli (I am also known as @tshooter.) Committed: Wed Aug 17 13:33:09 2016 -0700 -- .../hdfs/server/namenode/FSNamesystem.java | 185 ++ .../namenode/TestAuditLoggerWithCommands.java | 585 --- 2 files changed, 60 insertions(+), 710 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hadoop/blob/2486c4c6/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java -- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java index 2d9a069..73c8c8b 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java @@ -1855,16 +1855,13 @@ public class FSNamesystem implements Namesystem, FSNamesystemMBean, checkNameNodeSafeMode("Cannot concat " + target); stat = FSDirConcatOp.concat(dir, target, srcs, logRetryCache); success = true; -} catch (AccessControlException ace) { - logAuditEvent(success, "concat", Arrays.toString(srcs), target, stat); - throw ace; } finally { writeUnlock(); if (success) { getEditLog().logSync(); } + logAuditEvent(success, "concat", Arrays.toString(srcs), target, stat); } -logAuditEvent(success, "concat", Arrays.toString(srcs), target, stat); } /** @@ -2728,8 +2725,9 @@ public class FSNamesystem implements Namesystem, FSNamesystemMBean, boolean success = ret != null && ret.success; if (success) { getEditLog().logSync(); - logAuditEvent(success, "rename", src, dst, ret.auditStat); } +logAuditEvent(success, "rename", src, dst, +ret == null ? null : ret.auditStat); return success; } @@ -2943,19 +2941,16 @@ public class FSNamesystem implements Namesystem, FSNamesystemMBean, checkOperation(OperationCategory.READ); readLock(); boolean success = true; -ContentSummary cs; try { checkOperation(OperationCategory.READ); - cs = FSDirStatAndListingOp.getContentSummary(dir, src); + return FSDirStatAndListingOp.getContentSummary(dir, src); } catch (AccessControlException ace) { success = false; - logAuditEvent(success, "contentSummary", src); throw ace; } finally { readUnlock(); + logAuditEvent(success, "contentSummary", src); } -logAuditEvent(success, "contentSummary", src); -return cs; } /** @@ -2974,21 +2969,18 @@ public class FSNamesystem implements Namesystem, FSNamesystemMBean, */ QuotaUsage getQuotaUsage(final String src) throws IOException { checkOperation(OperationCategory.READ); -QuotaUsage quotaUsage; readLock(); boolean success = true; try { checkOperation(OperationCategory.READ); - quotaUsage = FSDirStatAndListingOp.getQuotaUsage(dir, src); + return FSDirStatAndListingOp.getQuotaUsage(dir, src); } catch (AccessControlException ace) { success = false; - logAuditEvent(success, "quotaUsage", src); throw ace; } finally { readUnlock(); + logAuditEvent(success, "quotaUsage", src); } -logAuditEvent(success, "quotaUsage", src); -return quotaUsage; } /** @@ -3011,16 +3003,13 @@ public class FSNamesystem implements Namesystem, FSNamesystemMBean, checkNameNodeSafeMode("Cannot set quota on " + src); FSDirAttrOp.setQuota(dir, src, nsQuota, ssQuota, type); success = true; -} catch (AccessControlException ace) { - logAuditEvent(success, "setQuota", src); - throw ace; } finally { writeUnlock(); if (success) { getEditLog().logSync(); } + logAuditEvent(success,
[hadoop] Git Push Summary
Repository: hadoop Updated Tags: refs/tags/release-2.7.3-RC1 2c543d156 -> 98cacdb50 - To unsubscribe, e-mail: common-commits-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-commits-h...@hadoop.apache.org
hadoop git commit: Set the release date for 2.7.3-RC1
Repository: hadoop Updated Branches: refs/heads/branch-2.7.3 5474c9e73 -> 7a4746cbb Set the release date for 2.7.3-RC1 Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/7a4746cb Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/7a4746cb Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/7a4746cb Branch: refs/heads/branch-2.7.3 Commit: 7a4746cbbd4eb019219ab316aa757613b318d16f Parents: 5474c9e Author: Vinod Kumar Vavilapalli (I am also known as @tshooter.)Authored: Mon Aug 15 12:52:39 2016 -0700 Committer: Vinod Kumar Vavilapalli (I am also known as @tshooter.) Committed: Mon Aug 15 12:52:39 2016 -0700 -- hadoop-common-project/hadoop-common/CHANGES.txt | 2 +- hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt | 2 +- hadoop-mapreduce-project/CHANGES.txt| 2 +- hadoop-yarn-project/CHANGES.txt | 2 +- 4 files changed, 4 insertions(+), 4 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hadoop/blob/7a4746cb/hadoop-common-project/hadoop-common/CHANGES.txt -- diff --git a/hadoop-common-project/hadoop-common/CHANGES.txt b/hadoop-common-project/hadoop-common/CHANGES.txt index 2272388..3777e32 100644 --- a/hadoop-common-project/hadoop-common/CHANGES.txt +++ b/hadoop-common-project/hadoop-common/CHANGES.txt @@ -1,6 +1,6 @@ Hadoop Change Log -Release 2.7.3 - 2016-09-19 +Release 2.7.3 - 2016-08-19 INCOMPATIBLE CHANGES http://git-wip-us.apache.org/repos/asf/hadoop/blob/7a4746cb/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt -- diff --git a/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt b/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt index e467646..748489d 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt +++ b/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt @@ -1,6 +1,6 @@ Hadoop HDFS Change Log -Release 2.7.3 - 2016-09-19 +Release 2.7.3 - 2016-08-19 INCOMPATIBLE CHANGES http://git-wip-us.apache.org/repos/asf/hadoop/blob/7a4746cb/hadoop-mapreduce-project/CHANGES.txt -- diff --git a/hadoop-mapreduce-project/CHANGES.txt b/hadoop-mapreduce-project/CHANGES.txt index fff3b1d..26b71a5 100644 --- a/hadoop-mapreduce-project/CHANGES.txt +++ b/hadoop-mapreduce-project/CHANGES.txt @@ -1,6 +1,6 @@ Hadoop MapReduce Change Log -Release 2.7.3 - 2016-09-19 +Release 2.7.3 - 2016-08-19 INCOMPATIBLE CHANGES http://git-wip-us.apache.org/repos/asf/hadoop/blob/7a4746cb/hadoop-yarn-project/CHANGES.txt -- diff --git a/hadoop-yarn-project/CHANGES.txt b/hadoop-yarn-project/CHANGES.txt index de33018..9f8b014 100644 --- a/hadoop-yarn-project/CHANGES.txt +++ b/hadoop-yarn-project/CHANGES.txt @@ -1,6 +1,6 @@ Hadoop YARN Change Log -Release 2.7.3 - 2016-09-19 +Release 2.7.3 - 2016-08-19 INCOMPATIBLE CHANGES - To unsubscribe, e-mail: common-commits-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-commits-h...@hadoop.apache.org
hadoop git commit: Set the release date for 2.7.3-RC1
Repository: hadoop Updated Branches: refs/heads/branch-2.7.3 49a957187 -> 5474c9e73 Set the release date for 2.7.3-RC1 Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/5474c9e7 Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/5474c9e7 Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/5474c9e7 Branch: refs/heads/branch-2.7.3 Commit: 5474c9e736d4c44a603a3f6749130b67cd4da52f Parents: 49a9571 Author: Vinod Kumar Vavilapalli (I am also known as @tshooter.)Authored: Thu Aug 11 18:35:17 2016 -0700 Committer: Vinod Kumar Vavilapalli (I am also known as @tshooter.) Committed: Thu Aug 11 18:35:51 2016 -0700 -- hadoop-common-project/hadoop-common/CHANGES.txt | 2 +- hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt | 2 +- hadoop-mapreduce-project/CHANGES.txt| 2 +- hadoop-yarn-project/CHANGES.txt | 2 +- 4 files changed, 4 insertions(+), 4 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hadoop/blob/5474c9e7/hadoop-common-project/hadoop-common/CHANGES.txt -- diff --git a/hadoop-common-project/hadoop-common/CHANGES.txt b/hadoop-common-project/hadoop-common/CHANGES.txt index 49f0cbd..2272388 100644 --- a/hadoop-common-project/hadoop-common/CHANGES.txt +++ b/hadoop-common-project/hadoop-common/CHANGES.txt @@ -1,6 +1,6 @@ Hadoop Change Log -Release 2.7.3 - 2016-08-01 +Release 2.7.3 - 2016-09-19 INCOMPATIBLE CHANGES http://git-wip-us.apache.org/repos/asf/hadoop/blob/5474c9e7/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt -- diff --git a/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt b/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt index f38eba3..e467646 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt +++ b/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt @@ -1,6 +1,6 @@ Hadoop HDFS Change Log -Release 2.7.3 - 2016-08-01 +Release 2.7.3 - 2016-09-19 INCOMPATIBLE CHANGES http://git-wip-us.apache.org/repos/asf/hadoop/blob/5474c9e7/hadoop-mapreduce-project/CHANGES.txt -- diff --git a/hadoop-mapreduce-project/CHANGES.txt b/hadoop-mapreduce-project/CHANGES.txt index f72ede0..fff3b1d 100644 --- a/hadoop-mapreduce-project/CHANGES.txt +++ b/hadoop-mapreduce-project/CHANGES.txt @@ -1,6 +1,6 @@ Hadoop MapReduce Change Log -Release 2.7.3 - 2016-08-01 +Release 2.7.3 - 2016-09-19 INCOMPATIBLE CHANGES http://git-wip-us.apache.org/repos/asf/hadoop/blob/5474c9e7/hadoop-yarn-project/CHANGES.txt -- diff --git a/hadoop-yarn-project/CHANGES.txt b/hadoop-yarn-project/CHANGES.txt index 4d1d469..de33018 100644 --- a/hadoop-yarn-project/CHANGES.txt +++ b/hadoop-yarn-project/CHANGES.txt @@ -1,6 +1,6 @@ Hadoop YARN Change Log -Release 2.7.3 - 2016-08-01 +Release 2.7.3 - 2016-09-19 INCOMPATIBLE CHANGES - To unsubscribe, e-mail: common-commits-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-commits-h...@hadoop.apache.org
[hadoop] Git Push Summary
Repository: hadoop Updated Tags: refs/tags/release-2.7.3-RC1 [created] 2c543d156 - To unsubscribe, e-mail: common-commits-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-commits-h...@hadoop.apache.org
hadoop git commit: YARN-5470. Addedum to differentiate exactly matching of log-files with regex in yarn log CLI. Contributed by Xuan Gong.
Repository: hadoop Updated Branches: refs/heads/branch-2 2dbe16c02 -> 31bb302c9 YARN-5470. Addedum to differentiate exactly matching of log-files with regex in yarn log CLI. Contributed by Xuan Gong. (cherry picked from commit 83a2ffec408fc3ebb89ab13ab4788d9f02fae33b) Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/31bb302c Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/31bb302c Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/31bb302c Branch: refs/heads/branch-2 Commit: 31bb302c91397a77c3ba70589d0ca45e0282d524 Parents: 2dbe16c Author: Vinod Kumar Vavilapalli (I am also known as @tshooter.)Authored: Mon Aug 8 14:28:07 2016 -0700 Committer: Vinod Kumar Vavilapalli (I am also known as @tshooter.) Committed: Mon Aug 8 14:29:24 2016 -0700 -- .../apache/hadoop/yarn/client/cli/LogsCLI.java | 62 ++--- .../hadoop/yarn/client/cli/TestLogsCLI.java | 129 +++ 2 files changed, 122 insertions(+), 69 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hadoop/blob/31bb302c/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/LogsCLI.java -- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/LogsCLI.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/LogsCLI.java index 985f878..c0d8795 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/LogsCLI.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/LogsCLI.java @@ -84,6 +84,8 @@ public class LogsCLI extends Configured implements Tool { private static final String APP_OWNER_OPTION = "appOwner"; private static final String AM_CONTAINER_OPTION = "am"; private static final String PER_CONTAINER_LOG_FILES_OPTION = "log_files"; + private static final String PER_CONTAINER_LOG_FILES_REGEX_OPTION + = "log_files_pattern"; private static final String LIST_NODES_OPTION = "list_nodes"; private static final String SHOW_APPLICATION_LOG_INFO = "show_application_log_info"; @@ -91,7 +93,6 @@ public class LogsCLI extends Configured implements Tool { = "show_container_log_info"; private static final String OUT_OPTION = "out"; private static final String SIZE_OPTION = "size"; - private static final String REGEX_OPTION = "regex"; public static final String HELP_CMD = "help"; private PrintStream outStream = System.out; private YarnClient yarnClient = null; @@ -130,6 +131,7 @@ public class LogsCLI extends Configured implements Tool { boolean showContainerLogInfo = false; boolean useRegex = false; String[] logFiles = null; +String[] logFilesRegex = null; List amContainersList = new ArrayList(); String localDir = null; long bytes = Long.MAX_VALUE; @@ -145,7 +147,6 @@ public class LogsCLI extends Configured implements Tool { showApplicationLogInfo = commandLine.hasOption( SHOW_APPLICATION_LOG_INFO); showContainerLogInfo = commandLine.hasOption(SHOW_CONTAINER_LOG_INFO); - useRegex = commandLine.hasOption(REGEX_OPTION); if (getAMContainerLogs) { try { amContainersList = parseAMContainer(commandLine, printOpts); @@ -157,6 +158,11 @@ public class LogsCLI extends Configured implements Tool { if (commandLine.hasOption(PER_CONTAINER_LOG_FILES_OPTION)) { logFiles = commandLine.getOptionValues(PER_CONTAINER_LOG_FILES_OPTION); } + if (commandLine.hasOption(PER_CONTAINER_LOG_FILES_REGEX_OPTION)) { +logFilesRegex = commandLine.getOptionValues( +PER_CONTAINER_LOG_FILES_REGEX_OPTION); +useRegex = true; + } if (commandLine.hasOption(SIZE_OPTION)) { bytes = Long.parseLong(commandLine.getOptionValue(SIZE_OPTION)); } @@ -206,6 +212,12 @@ public class LogsCLI extends Configured implements Tool { return -1; } +if (logFiles != null && logFiles.length > 0 && logFilesRegex != null +&& logFilesRegex.length > 0) { + System.err.println("Invalid options. Can only accept one of " + + "log_files/log_files_pattern."); + return -1; +} if (localDir != null) { File file = new File(localDir); if (file.exists() && file.isFile()) { @@ -248,10 +260,12 @@ public class LogsCLI extends Configured implements Tool { } Set logs = new HashSet(); -if (fetchAllLogFiles(logFiles, useRegex)) { +if (fetchAllLogFiles(logFiles, logFilesRegex)) { logs.add("ALL"); } else if
hadoop git commit: YARN-5470. Addedum to differentiate exactly matching of log-files with regex in yarn log CLI. Contributed by Xuan Gong.
Repository: hadoop Updated Branches: refs/heads/trunk 58e1523c8 -> 83a2ffec4 YARN-5470. Addedum to differentiate exactly matching of log-files with regex in yarn log CLI. Contributed by Xuan Gong. Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/83a2ffec Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/83a2ffec Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/83a2ffec Branch: refs/heads/trunk Commit: 83a2ffec408fc3ebb89ab13ab4788d9f02fae33b Parents: 58e1523 Author: Vinod Kumar Vavilapalli (I am also known as @tshooter.)Authored: Mon Aug 8 14:28:07 2016 -0700 Committer: Vinod Kumar Vavilapalli (I am also known as @tshooter.) Committed: Mon Aug 8 14:28:07 2016 -0700 -- .../apache/hadoop/yarn/client/cli/LogsCLI.java | 62 ++--- .../hadoop/yarn/client/cli/TestLogsCLI.java | 129 +++ 2 files changed, 122 insertions(+), 69 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hadoop/blob/83a2ffec/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/LogsCLI.java -- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/LogsCLI.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/LogsCLI.java index ec045b2..908d379 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/LogsCLI.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/LogsCLI.java @@ -84,6 +84,8 @@ public class LogsCLI extends Configured implements Tool { private static final String APP_OWNER_OPTION = "appOwner"; private static final String AM_CONTAINER_OPTION = "am"; private static final String PER_CONTAINER_LOG_FILES_OPTION = "log_files"; + private static final String PER_CONTAINER_LOG_FILES_REGEX_OPTION + = "log_files_pattern"; private static final String LIST_NODES_OPTION = "list_nodes"; private static final String SHOW_APPLICATION_LOG_INFO = "show_application_log_info"; @@ -91,7 +93,6 @@ public class LogsCLI extends Configured implements Tool { = "show_container_log_info"; private static final String OUT_OPTION = "out"; private static final String SIZE_OPTION = "size"; - private static final String REGEX_OPTION = "regex"; public static final String HELP_CMD = "help"; private PrintStream outStream = System.out; private YarnClient yarnClient = null; @@ -130,6 +131,7 @@ public class LogsCLI extends Configured implements Tool { boolean showContainerLogInfo = false; boolean useRegex = false; String[] logFiles = null; +String[] logFilesRegex = null; List amContainersList = new ArrayList(); String localDir = null; long bytes = Long.MAX_VALUE; @@ -145,7 +147,6 @@ public class LogsCLI extends Configured implements Tool { showApplicationLogInfo = commandLine.hasOption( SHOW_APPLICATION_LOG_INFO); showContainerLogInfo = commandLine.hasOption(SHOW_CONTAINER_LOG_INFO); - useRegex = commandLine.hasOption(REGEX_OPTION); if (getAMContainerLogs) { try { amContainersList = parseAMContainer(commandLine, printOpts); @@ -157,6 +158,11 @@ public class LogsCLI extends Configured implements Tool { if (commandLine.hasOption(PER_CONTAINER_LOG_FILES_OPTION)) { logFiles = commandLine.getOptionValues(PER_CONTAINER_LOG_FILES_OPTION); } + if (commandLine.hasOption(PER_CONTAINER_LOG_FILES_REGEX_OPTION)) { +logFilesRegex = commandLine.getOptionValues( +PER_CONTAINER_LOG_FILES_REGEX_OPTION); +useRegex = true; + } if (commandLine.hasOption(SIZE_OPTION)) { bytes = Long.parseLong(commandLine.getOptionValue(SIZE_OPTION)); } @@ -206,6 +212,12 @@ public class LogsCLI extends Configured implements Tool { return -1; } +if (logFiles != null && logFiles.length > 0 && logFilesRegex != null +&& logFilesRegex.length > 0) { + System.err.println("Invalid options. Can only accept one of " + + "log_files/log_files_pattern."); + return -1; +} if (localDir != null) { File file = new File(localDir); if (file.exists() && file.isFile()) { @@ -248,10 +260,12 @@ public class LogsCLI extends Configured implements Tool { } Set logs = new HashSet(); -if (fetchAllLogFiles(logFiles, useRegex)) { +if (fetchAllLogFiles(logFiles, logFilesRegex)) { logs.add("ALL"); } else if (logFiles != null && logFiles.length > 0) {
[hadoop] Git Push Summary
Repository: hadoop Updated Branches: refs/heads/yarn-native-services [created] 2d8d183b1 - To unsubscribe, e-mail: common-commits-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-commits-h...@hadoop.apache.org
[2/2] hadoop git commit: Preparing for release 2.7.3: Creating release notes.
Preparing for release 2.7.3: Creating release notes. Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/5d810bae Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/5d810bae Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/5d810bae Branch: refs/heads/branch-2.7.3 Commit: 5d810bae8df975517a478029303ad866193d1af8 Parents: 8c11fda Author: Vinod Kumar Vavilapalli (I am also known as @tshooter.)Authored: Fri Jul 22 15:54:02 2016 -0700 Committer: Vinod Kumar Vavilapalli (I am also known as @tshooter.) Committed: Fri Jul 22 15:54:02 2016 -0700 -- .../src/main/docs/releasenotes.html | 29777 + 1 file changed, 749 insertions(+), 29028 deletions(-) -- - To unsubscribe, e-mail: common-commits-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-commits-h...@hadoop.apache.org
hadoop git commit: Preparing for 2.7.3 release: Reset the tentative release date for 2.7.3.
Repository: hadoop Updated Branches: refs/heads/branch-2.7.3 f2b7ae5e2 -> 8c11fda9e Preparing for 2.7.3 release: Reset the tentative release date for 2.7.3. Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/8c11fda9 Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/8c11fda9 Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/8c11fda9 Branch: refs/heads/branch-2.7.3 Commit: 8c11fda9e688193b0d93c1dd8cec4dcd685b0dd7 Parents: f2b7ae5 Author: Vinod Kumar Vavilapalli (I am also known as @tshooter.)Authored: Fri Jul 22 11:32:58 2016 -0700 Committer: Vinod Kumar Vavilapalli (I am also known as @tshooter.) Committed: Fri Jul 22 11:32:58 2016 -0700 -- hadoop-common-project/hadoop-common/CHANGES.txt | 2 +- hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt | 2 +- hadoop-mapreduce-project/CHANGES.txt| 2 +- hadoop-yarn-project/CHANGES.txt | 2 +- 4 files changed, 4 insertions(+), 4 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hadoop/blob/8c11fda9/hadoop-common-project/hadoop-common/CHANGES.txt -- diff --git a/hadoop-common-project/hadoop-common/CHANGES.txt b/hadoop-common-project/hadoop-common/CHANGES.txt index 9070515..d5b8e47 100644 --- a/hadoop-common-project/hadoop-common/CHANGES.txt +++ b/hadoop-common-project/hadoop-common/CHANGES.txt @@ -1,6 +1,6 @@ Hadoop Change Log -Release 2.7.3 - 2016-07-27 +Release 2.7.3 - 2016-08-01 INCOMPATIBLE CHANGES http://git-wip-us.apache.org/repos/asf/hadoop/blob/8c11fda9/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt -- diff --git a/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt b/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt index 250af29..f38eba3 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt +++ b/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt @@ -1,6 +1,6 @@ Hadoop HDFS Change Log -Release 2.7.3 - 2016-07-27 +Release 2.7.3 - 2016-08-01 INCOMPATIBLE CHANGES http://git-wip-us.apache.org/repos/asf/hadoop/blob/8c11fda9/hadoop-mapreduce-project/CHANGES.txt -- diff --git a/hadoop-mapreduce-project/CHANGES.txt b/hadoop-mapreduce-project/CHANGES.txt index b18212a..f72ede0 100644 --- a/hadoop-mapreduce-project/CHANGES.txt +++ b/hadoop-mapreduce-project/CHANGES.txt @@ -1,6 +1,6 @@ Hadoop MapReduce Change Log -Release 2.7.3 - 2016-07-27 +Release 2.7.3 - 2016-08-01 INCOMPATIBLE CHANGES http://git-wip-us.apache.org/repos/asf/hadoop/blob/8c11fda9/hadoop-yarn-project/CHANGES.txt -- diff --git a/hadoop-yarn-project/CHANGES.txt b/hadoop-yarn-project/CHANGES.txt index 1e4256e..4d1d469 100644 --- a/hadoop-yarn-project/CHANGES.txt +++ b/hadoop-yarn-project/CHANGES.txt @@ -1,6 +1,6 @@ Hadoop YARN Change Log -Release 2.7.3 - 2016-07-27 +Release 2.7.3 - 2016-08-01 INCOMPATIBLE CHANGES - To unsubscribe, e-mail: common-commits-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-commits-h...@hadoop.apache.org
hadoop git commit: Preparing for 2.7.3 release: Set the release date for 2.7.3.
Repository: hadoop Updated Branches: refs/heads/branch-2.7.3 93a05a956 -> f2b7ae5e2 Updated Tags: refs/tags/release-2.7.3-RC0 [created] 8375f7015 Preparing for 2.7.3 release: Set the release date for 2.7.3. Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/f2b7ae5e Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/f2b7ae5e Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/f2b7ae5e Branch: refs/heads/branch-2.7.3 Commit: f2b7ae5e28dfa4da1ab9632caf0c2b1407d0c1f3 Parents: 93a05a9 Author: Vinod Kumar VavilapalliAuthored: Wed Jul 20 10:48:55 2016 -0700 Committer: Vinod Kumar Vavilapalli Committed: Wed Jul 20 10:57:35 2016 -0700 -- hadoop-common-project/hadoop-common/CHANGES.txt | 2 +- hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt | 2 +- hadoop-mapreduce-project/CHANGES.txt| 2 +- hadoop-yarn-project/CHANGES.txt | 5 - 4 files changed, 7 insertions(+), 4 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hadoop/blob/f2b7ae5e/hadoop-common-project/hadoop-common/CHANGES.txt -- diff --git a/hadoop-common-project/hadoop-common/CHANGES.txt b/hadoop-common-project/hadoop-common/CHANGES.txt index ae67722..9070515 100644 --- a/hadoop-common-project/hadoop-common/CHANGES.txt +++ b/hadoop-common-project/hadoop-common/CHANGES.txt @@ -1,6 +1,6 @@ Hadoop Change Log -Release 2.7.3 - UNRELEASED +Release 2.7.3 - 2016-07-27 INCOMPATIBLE CHANGES http://git-wip-us.apache.org/repos/asf/hadoop/blob/f2b7ae5e/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt -- diff --git a/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt b/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt index 4dc5e59..250af29 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt +++ b/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt @@ -1,6 +1,6 @@ Hadoop HDFS Change Log -Release 2.7.3 - UNRELEASED +Release 2.7.3 - 2016-07-27 INCOMPATIBLE CHANGES http://git-wip-us.apache.org/repos/asf/hadoop/blob/f2b7ae5e/hadoop-mapreduce-project/CHANGES.txt -- diff --git a/hadoop-mapreduce-project/CHANGES.txt b/hadoop-mapreduce-project/CHANGES.txt index 4993de9..b18212a 100644 --- a/hadoop-mapreduce-project/CHANGES.txt +++ b/hadoop-mapreduce-project/CHANGES.txt @@ -1,6 +1,6 @@ Hadoop MapReduce Change Log -Release 2.7.3 - UNRELEASED +Release 2.7.3 - 2016-07-27 INCOMPATIBLE CHANGES http://git-wip-us.apache.org/repos/asf/hadoop/blob/f2b7ae5e/hadoop-yarn-project/CHANGES.txt -- diff --git a/hadoop-yarn-project/CHANGES.txt b/hadoop-yarn-project/CHANGES.txt index 0ae1f94..1e4256e 100644 --- a/hadoop-yarn-project/CHANGES.txt +++ b/hadoop-yarn-project/CHANGES.txt @@ -1,6 +1,6 @@ Hadoop YARN Change Log -Release 2.7.3 - UNRELEASED +Release 2.7.3 - 2016-07-27 INCOMPATIBLE CHANGES @@ -176,6 +176,9 @@ Release 2.7.3 - UNRELEASED YARN-4686. MiniYARNCluster.start() returns before cluster is completely started. (Eric Badger via epayne) +YARN-5309. Fixed SSLFactory truststore reloader thread leak in +TimelineClientImpl. (Weiwei Yang via vvasudev) + Release 2.7.2 - 2016-01-25 INCOMPATIBLE CHANGES - To unsubscribe, e-mail: common-commits-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-commits-h...@hadoop.apache.org
hadoop git commit: YARN-5340. Fixed a race condition in RollingLevelDBTimelineStore that caused loss of Timeline events. Contributed by Li Lu.
Repository: hadoop Updated Branches: refs/heads/trunk 37362c2f9 -> 1c9d2ab50 YARN-5340. Fixed a race condition in RollingLevelDBTimelineStore that caused loss of Timeline events. Contributed by Li Lu. Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/1c9d2ab5 Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/1c9d2ab5 Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/1c9d2ab5 Branch: refs/heads/trunk Commit: 1c9d2ab503ea5a3f16757351af9603041059b390 Parents: 37362c2 Author: Vinod Kumar VavilapalliAuthored: Wed Jul 20 08:36:36 2016 -0700 Committer: Vinod Kumar Vavilapalli Committed: Wed Jul 20 08:36:36 2016 -0700 -- .../timeline/RollingLevelDBTimelineStore.java | 39 ++-- 1 file changed, 20 insertions(+), 19 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hadoop/blob/1c9d2ab5/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/src/main/java/org/apache/hadoop/yarn/server/timeline/RollingLevelDBTimelineStore.java -- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/src/main/java/org/apache/hadoop/yarn/server/timeline/RollingLevelDBTimelineStore.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/src/main/java/org/apache/hadoop/yarn/server/timeline/RollingLevelDBTimelineStore.java index 8b6a51b..4d38008 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/src/main/java/org/apache/hadoop/yarn/server/timeline/RollingLevelDBTimelineStore.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/src/main/java/org/apache/hadoop/yarn/server/timeline/RollingLevelDBTimelineStore.java @@ -1243,8 +1243,7 @@ public class RollingLevelDBTimelineStore extends AbstractService implements * Get the unique start time for a given entity as a byte array that sorts the * timestamps in reverse order (see * {@link GenericObjectMapper#writeReverseOrderedLong(long)}). If the start - * time doesn't exist, set it based on the information provided. Should only - * be called when a lock has been obtained on the entity. + * time doesn't exist, set it based on the information provided. * * @param entityId * The id of the entity @@ -1257,8 +1256,9 @@ public class RollingLevelDBTimelineStore extends AbstractService implements * @return A StartAndInsertTime * @throws IOException */ - private Long getAndSetStartTime(String entityId, String entityType, - Long startTime, List events) throws IOException { + private Long getAndSetStartTime(String entityId, + String entityType, Long startTime, List events) + throws IOException { EntityIdentifier entity = new EntityIdentifier(entityId, entityType); Long time = startTimeWriteCache.get(entity); if (time != null) { @@ -1282,28 +1282,29 @@ public class RollingLevelDBTimelineStore extends AbstractService implements * Checks db for start time and returns it if it exists. If it doesn't exist, * writes the suggested start time (if it is not null). This is only called * when the start time is not found in the cache, so it adds it back into the - * cache if it is found. Should only be called when a lock has been obtained - * on the entity. + * cache if it is found. */ private Long checkStartTimeInDb(EntityIdentifier entity, Long suggestedStartTime) throws IOException { Long startAndInsertTime = null; // create lookup key for start time byte[] b = createStartTimeLookupKey(entity.getId(), entity.getType()); -// retrieve value for key -byte[] v = starttimedb.get(b); -if (v == null) { - // start time doesn't exist in db - if (suggestedStartTime == null) { -return null; - } - startAndInsertTime = suggestedStartTime; +synchronized (this) { + // retrieve value for key + byte[] v = starttimedb.get(b); + if (v == null) { +// start time doesn't exist in db +if (suggestedStartTime == null) { + return null; +} +startAndInsertTime = suggestedStartTime; - // write suggested start time - starttimedb.put(b, writeReverseOrderedLong(suggestedStartTime)); -} else { - // found start time in db, so ignore suggested start time - startAndInsertTime = readReverseOrderedLong(v, 0); +// write suggested start time +starttimedb.put(b, writeReverseOrderedLong(suggestedStartTime)); + } else { +// found start time in db, so ignore
hadoop git commit: YARN-5340. Fixed a race condition in RollingLevelDBTimelineStore that caused loss of Timeline events. Contributed by Li Lu.
Repository: hadoop Updated Branches: refs/heads/branch-2.8 c177823eb -> 8c5101ea6 YARN-5340. Fixed a race condition in RollingLevelDBTimelineStore that caused loss of Timeline events. Contributed by Li Lu. (cherry picked from commit 1c9d2ab503ea5a3f16757351af9603041059b390) Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/8c5101ea Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/8c5101ea Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/8c5101ea Branch: refs/heads/branch-2.8 Commit: 8c5101ea6dd7d900802da6a6d2f8f239f92c4669 Parents: c177823 Author: Vinod Kumar VavilapalliAuthored: Wed Jul 20 08:36:36 2016 -0700 Committer: Vinod Kumar Vavilapalli Committed: Wed Jul 20 08:39:19 2016 -0700 -- .../timeline/RollingLevelDBTimelineStore.java | 39 ++-- 1 file changed, 20 insertions(+), 19 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hadoop/blob/8c5101ea/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/src/main/java/org/apache/hadoop/yarn/server/timeline/RollingLevelDBTimelineStore.java -- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/src/main/java/org/apache/hadoop/yarn/server/timeline/RollingLevelDBTimelineStore.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/src/main/java/org/apache/hadoop/yarn/server/timeline/RollingLevelDBTimelineStore.java index 8b6a51b..4d38008 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/src/main/java/org/apache/hadoop/yarn/server/timeline/RollingLevelDBTimelineStore.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/src/main/java/org/apache/hadoop/yarn/server/timeline/RollingLevelDBTimelineStore.java @@ -1243,8 +1243,7 @@ public class RollingLevelDBTimelineStore extends AbstractService implements * Get the unique start time for a given entity as a byte array that sorts the * timestamps in reverse order (see * {@link GenericObjectMapper#writeReverseOrderedLong(long)}). If the start - * time doesn't exist, set it based on the information provided. Should only - * be called when a lock has been obtained on the entity. + * time doesn't exist, set it based on the information provided. * * @param entityId * The id of the entity @@ -1257,8 +1256,9 @@ public class RollingLevelDBTimelineStore extends AbstractService implements * @return A StartAndInsertTime * @throws IOException */ - private Long getAndSetStartTime(String entityId, String entityType, - Long startTime, List events) throws IOException { + private Long getAndSetStartTime(String entityId, + String entityType, Long startTime, List events) + throws IOException { EntityIdentifier entity = new EntityIdentifier(entityId, entityType); Long time = startTimeWriteCache.get(entity); if (time != null) { @@ -1282,28 +1282,29 @@ public class RollingLevelDBTimelineStore extends AbstractService implements * Checks db for start time and returns it if it exists. If it doesn't exist, * writes the suggested start time (if it is not null). This is only called * when the start time is not found in the cache, so it adds it back into the - * cache if it is found. Should only be called when a lock has been obtained - * on the entity. + * cache if it is found. */ private Long checkStartTimeInDb(EntityIdentifier entity, Long suggestedStartTime) throws IOException { Long startAndInsertTime = null; // create lookup key for start time byte[] b = createStartTimeLookupKey(entity.getId(), entity.getType()); -// retrieve value for key -byte[] v = starttimedb.get(b); -if (v == null) { - // start time doesn't exist in db - if (suggestedStartTime == null) { -return null; - } - startAndInsertTime = suggestedStartTime; +synchronized (this) { + // retrieve value for key + byte[] v = starttimedb.get(b); + if (v == null) { +// start time doesn't exist in db +if (suggestedStartTime == null) { + return null; +} +startAndInsertTime = suggestedStartTime; - // write suggested start time - starttimedb.put(b, writeReverseOrderedLong(suggestedStartTime)); -} else { - // found start time in db, so ignore suggested start time - startAndInsertTime = readReverseOrderedLong(v, 0); +// write suggested start time +starttimedb.put(b,
hadoop git commit: YARN-5340. Fixed a race condition in RollingLevelDBTimelineStore that caused loss of Timeline events. Contributed by Li Lu.
Repository: hadoop Updated Branches: refs/heads/branch-2 d3df810dc -> 5641c28c1 YARN-5340. Fixed a race condition in RollingLevelDBTimelineStore that caused loss of Timeline events. Contributed by Li Lu. (cherry picked from commit 1c9d2ab503ea5a3f16757351af9603041059b390) Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/5641c28c Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/5641c28c Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/5641c28c Branch: refs/heads/branch-2 Commit: 5641c28c1c583c9607d4adde0773a6307d2a6d48 Parents: d3df810 Author: Vinod Kumar VavilapalliAuthored: Wed Jul 20 08:36:36 2016 -0700 Committer: Vinod Kumar Vavilapalli Committed: Wed Jul 20 08:38:56 2016 -0700 -- .../timeline/RollingLevelDBTimelineStore.java | 39 ++-- 1 file changed, 20 insertions(+), 19 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hadoop/blob/5641c28c/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/src/main/java/org/apache/hadoop/yarn/server/timeline/RollingLevelDBTimelineStore.java -- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/src/main/java/org/apache/hadoop/yarn/server/timeline/RollingLevelDBTimelineStore.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/src/main/java/org/apache/hadoop/yarn/server/timeline/RollingLevelDBTimelineStore.java index 8b6a51b..4d38008 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/src/main/java/org/apache/hadoop/yarn/server/timeline/RollingLevelDBTimelineStore.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/src/main/java/org/apache/hadoop/yarn/server/timeline/RollingLevelDBTimelineStore.java @@ -1243,8 +1243,7 @@ public class RollingLevelDBTimelineStore extends AbstractService implements * Get the unique start time for a given entity as a byte array that sorts the * timestamps in reverse order (see * {@link GenericObjectMapper#writeReverseOrderedLong(long)}). If the start - * time doesn't exist, set it based on the information provided. Should only - * be called when a lock has been obtained on the entity. + * time doesn't exist, set it based on the information provided. * * @param entityId * The id of the entity @@ -1257,8 +1256,9 @@ public class RollingLevelDBTimelineStore extends AbstractService implements * @return A StartAndInsertTime * @throws IOException */ - private Long getAndSetStartTime(String entityId, String entityType, - Long startTime, List events) throws IOException { + private Long getAndSetStartTime(String entityId, + String entityType, Long startTime, List events) + throws IOException { EntityIdentifier entity = new EntityIdentifier(entityId, entityType); Long time = startTimeWriteCache.get(entity); if (time != null) { @@ -1282,28 +1282,29 @@ public class RollingLevelDBTimelineStore extends AbstractService implements * Checks db for start time and returns it if it exists. If it doesn't exist, * writes the suggested start time (if it is not null). This is only called * when the start time is not found in the cache, so it adds it back into the - * cache if it is found. Should only be called when a lock has been obtained - * on the entity. + * cache if it is found. */ private Long checkStartTimeInDb(EntityIdentifier entity, Long suggestedStartTime) throws IOException { Long startAndInsertTime = null; // create lookup key for start time byte[] b = createStartTimeLookupKey(entity.getId(), entity.getType()); -// retrieve value for key -byte[] v = starttimedb.get(b); -if (v == null) { - // start time doesn't exist in db - if (suggestedStartTime == null) { -return null; - } - startAndInsertTime = suggestedStartTime; +synchronized (this) { + // retrieve value for key + byte[] v = starttimedb.get(b); + if (v == null) { +// start time doesn't exist in db +if (suggestedStartTime == null) { + return null; +} +startAndInsertTime = suggestedStartTime; - // write suggested start time - starttimedb.put(b, writeReverseOrderedLong(suggestedStartTime)); -} else { - // found start time in db, so ignore suggested start time - startAndInsertTime = readReverseOrderedLong(v, 0); +// write suggested start time +starttimedb.put(b,
hadoop git commit: YARN-5339. Fixed "yarn logs" to fail when a file is passed to -out option instead of a directory. Contributed by Xuan Gong.
Repository: hadoop Updated Branches: refs/heads/branch-2 ee1e45e4b -> cbd2970be YARN-5339. Fixed "yarn logs" to fail when a file is passed to -out option instead of a directory. Contributed by Xuan Gong. (cherry picked from commit d18050522c5c6bd9e32eb9a1be4ffe2288624c40) Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/cbd2970b Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/cbd2970b Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/cbd2970b Branch: refs/heads/branch-2 Commit: cbd2970be769a23f7a0ebf6011b45f546ee9 Parents: ee1e45e Author: Vinod Kumar VavilapalliAuthored: Wed Jul 13 14:52:30 2016 -0700 Committer: Vinod Kumar Vavilapalli Committed: Wed Jul 13 14:55:44 2016 -0700 -- .../apache/hadoop/yarn/client/cli/LogsCLI.java | 10 +++ .../hadoop/yarn/client/cli/TestLogsCLI.java | 65 +--- 2 files changed, 68 insertions(+), 7 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hadoop/blob/cbd2970b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/LogsCLI.java -- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/LogsCLI.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/LogsCLI.java index ce26bd0..c1ac792 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/LogsCLI.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/LogsCLI.java @@ -18,6 +18,7 @@ package org.apache.hadoop.yarn.client.cli; +import java.io.File; import java.io.IOException; import java.io.PrintStream; import java.util.ArrayList; @@ -197,6 +198,15 @@ public class LogsCLI extends Configured implements Tool { return -1; } +if (localDir != null) { + File file = new File(localDir); + if (file.exists() && file.isFile()) { +System.err.println("Invalid value for -out option. " ++ "Please provide a directory."); +return -1; + } +} + LogCLIHelpers logCliHelper = new LogCLIHelpers(); logCliHelper.setConf(getConf()); http://git-wip-us.apache.org/repos/asf/hadoop/blob/cbd2970b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/test/java/org/apache/hadoop/yarn/client/cli/TestLogsCLI.java -- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/test/java/org/apache/hadoop/yarn/client/cli/TestLogsCLI.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/test/java/org/apache/hadoop/yarn/client/cli/TestLogsCLI.java index cb9a663..3fe53d9 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/test/java/org/apache/hadoop/yarn/client/cli/TestLogsCLI.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/test/java/org/apache/hadoop/yarn/client/cli/TestLogsCLI.java @@ -711,6 +711,64 @@ public class TestLogsCLI { } } + @Test (timeout = 5000) + public void testLogsCLIWithInvalidArgs() throws Exception { +String localDir = "target/SaveLogs"; +Path localPath = new Path(localDir); +Configuration configuration = new Configuration(); +FileSystem fs = FileSystem.get(configuration); +ApplicationId appId = ApplicationId.newInstance(0, 1); +YarnClient mockYarnClient = +createMockYarnClient(YarnApplicationState.FINISHED, +UserGroupInformation.getCurrentUser().getShortUserName()); +LogsCLI cli = new LogsCLIForTest(mockYarnClient); +cli.setConf(configuration); + +// Specify an invalid applicationId +int exitCode = cli.run(new String[] {"-applicationId", +"123"}); +assertTrue(exitCode == -1); +assertTrue(sysErrStream.toString().contains( +"Invalid ApplicationId specified")); +sysErrStream.reset(); + +// Specify an invalid containerId +exitCode = cli.run(new String[] {"-containerId", +"123"}); +assertTrue(exitCode == -1); +assertTrue(sysErrStream.toString().contains( +"Invalid ContainerId specified")); +sysErrStream.reset(); + +// Specify show_container_log_info and show_application_log_info +// at the same time +exitCode = cli.run(new String[] {"-applicationId", appId.toString(), +"-show_container_log_info", "-show_application_log_info"}); +assertTrue(exitCode == -1); +assertTrue(sysErrStream.toString().contains("Invalid options. " ++ "Can only accept one of show_application_log_info/" ++ "show_container_log_info.")); +sysErrStream.reset(); + +
hadoop git commit: YARN-5339. Fixed "yarn logs" to fail when a file is passed to -out option instead of a directory. Contributed by Xuan Gong.
Repository: hadoop Updated Branches: refs/heads/trunk af8f480c2 -> d18050522 YARN-5339. Fixed "yarn logs" to fail when a file is passed to -out option instead of a directory. Contributed by Xuan Gong. Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/d1805052 Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/d1805052 Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/d1805052 Branch: refs/heads/trunk Commit: d18050522c5c6bd9e32eb9a1be4ffe2288624c40 Parents: af8f480 Author: Vinod Kumar VavilapalliAuthored: Wed Jul 13 14:52:30 2016 -0700 Committer: Vinod Kumar Vavilapalli Committed: Wed Jul 13 14:52:30 2016 -0700 -- .../apache/hadoop/yarn/client/cli/LogsCLI.java | 10 +++ .../hadoop/yarn/client/cli/TestLogsCLI.java | 65 +--- 2 files changed, 68 insertions(+), 7 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hadoop/blob/d1805052/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/LogsCLI.java -- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/LogsCLI.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/LogsCLI.java index 31395be..1560557 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/LogsCLI.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/LogsCLI.java @@ -18,6 +18,7 @@ package org.apache.hadoop.yarn.client.cli; +import java.io.File; import java.io.IOException; import java.io.PrintStream; import java.util.ArrayList; @@ -197,6 +198,15 @@ public class LogsCLI extends Configured implements Tool { return -1; } +if (localDir != null) { + File file = new File(localDir); + if (file.exists() && file.isFile()) { +System.err.println("Invalid value for -out option. " ++ "Please provide a directory."); +return -1; + } +} + LogCLIHelpers logCliHelper = new LogCLIHelpers(); logCliHelper.setConf(getConf()); http://git-wip-us.apache.org/repos/asf/hadoop/blob/d1805052/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/test/java/org/apache/hadoop/yarn/client/cli/TestLogsCLI.java -- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/test/java/org/apache/hadoop/yarn/client/cli/TestLogsCLI.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/test/java/org/apache/hadoop/yarn/client/cli/TestLogsCLI.java index cb9a663..3fe53d9 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/test/java/org/apache/hadoop/yarn/client/cli/TestLogsCLI.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/test/java/org/apache/hadoop/yarn/client/cli/TestLogsCLI.java @@ -711,6 +711,64 @@ public class TestLogsCLI { } } + @Test (timeout = 5000) + public void testLogsCLIWithInvalidArgs() throws Exception { +String localDir = "target/SaveLogs"; +Path localPath = new Path(localDir); +Configuration configuration = new Configuration(); +FileSystem fs = FileSystem.get(configuration); +ApplicationId appId = ApplicationId.newInstance(0, 1); +YarnClient mockYarnClient = +createMockYarnClient(YarnApplicationState.FINISHED, +UserGroupInformation.getCurrentUser().getShortUserName()); +LogsCLI cli = new LogsCLIForTest(mockYarnClient); +cli.setConf(configuration); + +// Specify an invalid applicationId +int exitCode = cli.run(new String[] {"-applicationId", +"123"}); +assertTrue(exitCode == -1); +assertTrue(sysErrStream.toString().contains( +"Invalid ApplicationId specified")); +sysErrStream.reset(); + +// Specify an invalid containerId +exitCode = cli.run(new String[] {"-containerId", +"123"}); +assertTrue(exitCode == -1); +assertTrue(sysErrStream.toString().contains( +"Invalid ContainerId specified")); +sysErrStream.reset(); + +// Specify show_container_log_info and show_application_log_info +// at the same time +exitCode = cli.run(new String[] {"-applicationId", appId.toString(), +"-show_container_log_info", "-show_application_log_info"}); +assertTrue(exitCode == -1); +assertTrue(sysErrStream.toString().contains("Invalid options. " ++ "Can only accept one of show_application_log_info/" ++ "show_container_log_info.")); +sysErrStream.reset(); + +// Specify a file name to the option -out +try { +
hadoop git commit: YARN-5200. Enhanced "yarn logs" to be able to get a list of containers whose logs are aggregated via a "show_container_log_info" option. Contributed by Xuan Gong.
Repository: hadoop Updated Branches: refs/heads/branch-2 694b4abbc -> 4a6cef8c9 YARN-5200. Enhanced "yarn logs" to be able to get a list of containers whose logs are aggregated via a "show_container_log_info" option. Contributed by Xuan Gong. (cherry picked from commit eb471632349deac4b62f8dec853c8ceb64c9617a) Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/4a6cef8c Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/4a6cef8c Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/4a6cef8c Branch: refs/heads/branch-2 Commit: 4a6cef8c910548251bec336147ee4be65736102a Parents: 694b4ab Author: Vinod Kumar VavilapalliAuthored: Wed Jul 13 10:47:43 2016 -0700 Committer: Vinod Kumar Vavilapalli Committed: Wed Jul 13 10:55:18 2016 -0700 -- .../apache/hadoop/yarn/client/cli/LogsCLI.java | 321 ++- .../hadoop/yarn/client/cli/TestLogsCLI.java | 85 +++-- .../logaggregation/AggregatedLogFormat.java | 13 +- .../yarn/logaggregation/LogCLIHelpers.java | 97 +- 4 files changed, 392 insertions(+), 124 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hadoop/blob/4a6cef8c/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/LogsCLI.java -- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/LogsCLI.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/LogsCLI.java index 7194cd0..ce26bd0 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/LogsCLI.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/LogsCLI.java @@ -78,14 +78,30 @@ public class LogsCLI extends Configured implements Tool { private static final String APP_OWNER_OPTION = "appOwner"; private static final String AM_CONTAINER_OPTION = "am"; private static final String CONTAINER_LOG_FILES = "logFiles"; - private static final String SHOW_META_INFO = "show_meta_info"; private static final String LIST_NODES_OPTION = "list_nodes"; + private static final String SHOW_APPLICATION_LOG_INFO + = "show_application_log_info"; + private static final String SHOW_CONTAINER_LOG_INFO + = "show_container_log_info"; private static final String OUT_OPTION = "out"; private static final String SIZE_OPTION = "size"; public static final String HELP_CMD = "help"; + private PrintStream outStream = System.out; + private YarnClient yarnClient = null; @Override public int run(String[] args) throws Exception { +try { + yarnClient = createYarnClient(); + return runCommand(args); +} finally { + if (yarnClient != null) { +yarnClient.close(); + } +} + } + + private int runCommand(String[] args) throws Exception { Options opts = createCommandOpts(); Options printOpts = createPrintOpts(opts); if (args.length < 1) { @@ -102,8 +118,9 @@ public class LogsCLI extends Configured implements Tool { String nodeAddress = null; String appOwner = null; boolean getAMContainerLogs = false; -boolean showMetaInfo = false; boolean nodesList = false; +boolean showApplicationLogInfo = false; +boolean showContainerLogInfo = false; String[] logFiles = null; List amContainersList = new ArrayList(); String localDir = null; @@ -115,9 +132,11 @@ public class LogsCLI extends Configured implements Tool { nodeAddress = commandLine.getOptionValue(NODE_ADDRESS_OPTION); appOwner = commandLine.getOptionValue(APP_OWNER_OPTION); getAMContainerLogs = commandLine.hasOption(AM_CONTAINER_OPTION); - showMetaInfo = commandLine.hasOption(SHOW_META_INFO); nodesList = commandLine.hasOption(LIST_NODES_OPTION); localDir = commandLine.getOptionValue(OUT_OPTION); + showApplicationLogInfo = commandLine.hasOption( + SHOW_APPLICATION_LOG_INFO); + showContainerLogInfo = commandLine.hasOption(SHOW_CONTAINER_LOG_INFO); if (getAMContainerLogs) { try { amContainersList = parseAMContainer(commandLine, printOpts); @@ -172,6 +191,12 @@ public class LogsCLI extends Configured implements Tool { } } +if (showApplicationLogInfo && showContainerLogInfo) { + System.err.println("Invalid options. Can only accept one of " + + "show_application_log_info/show_container_log_info."); + return -1; +} + LogCLIHelpers logCliHelper = new LogCLIHelpers(); logCliHelper.setConf(getConf()); @@ -215,14 +240,17 @@ public class LogsCLI extends
hadoop git commit: YARN-5200. Enhanced "yarn logs" to be able to get a list of containers whose logs are aggregated via a "show_container_log_info" option. Contributed by Xuan Gong.
Repository: hadoop Updated Branches: refs/heads/trunk 56142171b -> eb4716323 YARN-5200. Enhanced "yarn logs" to be able to get a list of containers whose logs are aggregated via a "show_container_log_info" option. Contributed by Xuan Gong. Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/eb471632 Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/eb471632 Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/eb471632 Branch: refs/heads/trunk Commit: eb471632349deac4b62f8dec853c8ceb64c9617a Parents: 5614217 Author: Vinod Kumar VavilapalliAuthored: Wed Jul 13 10:47:43 2016 -0700 Committer: Vinod Kumar Vavilapalli Committed: Wed Jul 13 10:54:15 2016 -0700 -- .../apache/hadoop/yarn/client/cli/LogsCLI.java | 321 ++- .../hadoop/yarn/client/cli/TestLogsCLI.java | 85 +++-- .../logaggregation/AggregatedLogFormat.java | 13 +- .../yarn/logaggregation/LogCLIHelpers.java | 97 +- 4 files changed, 392 insertions(+), 124 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hadoop/blob/eb471632/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/LogsCLI.java -- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/LogsCLI.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/LogsCLI.java index 79401b1..31395be 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/LogsCLI.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/LogsCLI.java @@ -78,14 +78,30 @@ public class LogsCLI extends Configured implements Tool { private static final String APP_OWNER_OPTION = "appOwner"; private static final String AM_CONTAINER_OPTION = "am"; private static final String CONTAINER_LOG_FILES = "logFiles"; - private static final String SHOW_META_INFO = "show_meta_info"; private static final String LIST_NODES_OPTION = "list_nodes"; + private static final String SHOW_APPLICATION_LOG_INFO + = "show_application_log_info"; + private static final String SHOW_CONTAINER_LOG_INFO + = "show_container_log_info"; private static final String OUT_OPTION = "out"; private static final String SIZE_OPTION = "size"; public static final String HELP_CMD = "help"; + private PrintStream outStream = System.out; + private YarnClient yarnClient = null; @Override public int run(String[] args) throws Exception { +try { + yarnClient = createYarnClient(); + return runCommand(args); +} finally { + if (yarnClient != null) { +yarnClient.close(); + } +} + } + + private int runCommand(String[] args) throws Exception { Options opts = createCommandOpts(); Options printOpts = createPrintOpts(opts); if (args.length < 1) { @@ -102,8 +118,9 @@ public class LogsCLI extends Configured implements Tool { String nodeAddress = null; String appOwner = null; boolean getAMContainerLogs = false; -boolean showMetaInfo = false; boolean nodesList = false; +boolean showApplicationLogInfo = false; +boolean showContainerLogInfo = false; String[] logFiles = null; List amContainersList = new ArrayList(); String localDir = null; @@ -115,9 +132,11 @@ public class LogsCLI extends Configured implements Tool { nodeAddress = commandLine.getOptionValue(NODE_ADDRESS_OPTION); appOwner = commandLine.getOptionValue(APP_OWNER_OPTION); getAMContainerLogs = commandLine.hasOption(AM_CONTAINER_OPTION); - showMetaInfo = commandLine.hasOption(SHOW_META_INFO); nodesList = commandLine.hasOption(LIST_NODES_OPTION); localDir = commandLine.getOptionValue(OUT_OPTION); + showApplicationLogInfo = commandLine.hasOption( + SHOW_APPLICATION_LOG_INFO); + showContainerLogInfo = commandLine.hasOption(SHOW_CONTAINER_LOG_INFO); if (getAMContainerLogs) { try { amContainersList = parseAMContainer(commandLine, printOpts); @@ -172,6 +191,12 @@ public class LogsCLI extends Configured implements Tool { } } +if (showApplicationLogInfo && showContainerLogInfo) { + System.err.println("Invalid options. Can only accept one of " + + "show_application_log_info/show_container_log_info."); + return -1; +} + LogCLIHelpers logCliHelper = new LogCLIHelpers(); logCliHelper.setConf(getConf()); @@ -215,14 +240,17 @@ public class LogsCLI extends Configured implements Tool { isApplicationFinished(appState), appOwner,
hadoop git commit: HADOOP-13312. Updated CHANGES.txt to reflect all the changes in branch-2.7. Contributed by Akira Ajisaka.
Repository: hadoop Updated Branches: refs/heads/branch-2.7.3 31898c0a9 -> 7a492f411 HADOOP-13312. Updated CHANGES.txt to reflect all the changes in branch-2.7. Contributed by Akira Ajisaka. (cherry picked from commit c454639abf2a49b3c84c221d3a541b6c6c6abbc8) Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/7a492f41 Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/7a492f41 Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/7a492f41 Branch: refs/heads/branch-2.7.3 Commit: 7a492f4111d1abc85479085a61ae1609f033282c Parents: 31898c0 Author: Vinod Kumar Vavilapalli <vino...@apache.org> Authored: Tue Jul 12 15:41:22 2016 -0700 Committer: Vinod Kumar Vavilapalli <vino...@apache.org> Committed: Tue Jul 12 15:43:48 2016 -0700 -- hadoop-common-project/hadoop-common/CHANGES.txt | 31 +++ hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt | 40 hadoop-mapreduce-project/CHANGES.txt| 6 +++ hadoop-yarn-project/CHANGES.txt | 10 + 4 files changed, 87 insertions(+) -- http://git-wip-us.apache.org/repos/asf/hadoop/blob/7a492f41/hadoop-common-project/hadoop-common/CHANGES.txt -- diff --git a/hadoop-common-project/hadoop-common/CHANGES.txt b/hadoop-common-project/hadoop-common/CHANGES.txt index 3919f2e..a752588 100644 --- a/hadoop-common-project/hadoop-common/CHANGES.txt +++ b/hadoop-common-project/hadoop-common/CHANGES.txt @@ -14,6 +14,13 @@ Release 2.7.3 - UNRELEASED HADOOP-12794. Support additional compression levels for GzipCodec. (Ravi Mutyala via junping_du) +HADOOP-13103. Group resolution from LDAP may fail on +javax.naming.ServiceUnavailableException (szetszwo) + +HADOOP-13039. Add documentation for configuration property +ipc.maximum.data.length for controlling maximum RPC message size. +(Mingliang Liu via arp) + OPTIMIZATIONS HADOOP-12810. FileSystem#listLocatedStatus causes unnecessary RPC calls @@ -111,6 +118,30 @@ Release 2.7.3 - UNRELEASED HADOOP-13297. Add missing dependency in setting maven-remote-resource-plugin to fix builds. (Sean Busbey via aajisaka) +HADOOP-13255. KMSClientProvider should check and renew tgt when doing +delegation token operations. (xyao) + +HADOOP-12559. KMS connection failures should trigger TGT renewal. +(Zhe Zhang via xyao) + +HADOOP-13043. Add LICENSE.txt entries for bundled javascript dependencies. +(wang) + +HADOOP-13042. Restore lost leveldbjni LICENSE and NOTICE changes. (wang) + +HADOOP-12406. AbstractMapWritable.readFields throws ClassNotFoundException +with custom writables. (Nadeem Douba via vinodkv) + +HADOOP-12636. Fix to prevent ServiceLoader failure init for unused FileSystems. +(Inigo Goiri via cdouglas) + +HADOOP-13350. Additional fixes to LICENSE and NOTICE. (Xiao Chen via wang) + +HADOOP-12682. Fixed TestKMS#testKMSRestart* failure. (Wei-Chiu Chuang via xyao) + +HADOOP-13312. Updated CHANGES.txt to reflect all the changes in branch-2.7. +(Akira Ajisaka via vinodkv) + Release 2.7.2 - 2016-01-25 INCOMPATIBLE CHANGES http://git-wip-us.apache.org/repos/asf/hadoop/blob/7a492f41/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt -- diff --git a/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt b/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt index 0e0db3c..c9d7aad 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt +++ b/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt @@ -49,6 +49,20 @@ Release 2.7.3 - UNRELEASED HDFS-8772. Fix TestStandbyIsHot#testDatanodeRestarts which occasionally fails. (Walter Su via wang). +HDFS-9906. Remove spammy log spew when a datanode is restarted. +(Brahma Reddy Battula via arp) + +HDFS-2580. NameNode#main(...) can make use of GenericOptionsParser. (harsh) + +HDFS-10264. Logging improvements in FSImageFormatProtobuf.Saver. +(Xiaobing Zhou via arp) + +HDFS-10186. DirectoryScanner: Improve logs by adding full path of both +actual and expected block directories. (Rakesh R via szetszwo) + +HDFS-10488. Updated WebHDFS documentation regarding CREATE and MKDIR default +permissions. (Wellington Chevreuil via wang) + OPTIMIZATIONS HDFS-8845. DiskChecker should not traverse the entire tree (Chang Li via @@ -192,6 +206,32 @@ Release 2.7.3 - UNRELEASED HDFS-9530. ReservedSpace is not cleared for abandoned Blocks (Brahma Reddy Battula) +HDFS-10458. getFileEncryptionInfo should return quickly for non-encrypted +cluster. (zhz) + +HDFS-9365. Balaner does not work with the HDFS-6376 HA setup. (szetszwo) + +HDFS-10335.
hadoop git commit: HADOOP-13312. Updated CHANGES.txt to reflect all the changes in branch-2.7. Contributed by Akira Ajisaka.
Repository: hadoop Updated Branches: refs/heads/branch-2.7 f8f995629 -> c454639ab HADOOP-13312. Updated CHANGES.txt to reflect all the changes in branch-2.7. Contributed by Akira Ajisaka. Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/c454639a Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/c454639a Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/c454639a Branch: refs/heads/branch-2.7 Commit: c454639abf2a49b3c84c221d3a541b6c6c6abbc8 Parents: f8f9956 Author: Vinod Kumar Vavilapalli <vino...@apache.org> Authored: Tue Jul 12 15:41:22 2016 -0700 Committer: Vinod Kumar Vavilapalli <vino...@apache.org> Committed: Tue Jul 12 15:41:22 2016 -0700 -- hadoop-common-project/hadoop-common/CHANGES.txt | 31 +++ hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt | 40 hadoop-mapreduce-project/CHANGES.txt| 6 +++ hadoop-yarn-project/CHANGES.txt | 10 + 4 files changed, 87 insertions(+) -- http://git-wip-us.apache.org/repos/asf/hadoop/blob/c454639a/hadoop-common-project/hadoop-common/CHANGES.txt -- diff --git a/hadoop-common-project/hadoop-common/CHANGES.txt b/hadoop-common-project/hadoop-common/CHANGES.txt index 46667cc..e8535cc 100644 --- a/hadoop-common-project/hadoop-common/CHANGES.txt +++ b/hadoop-common-project/hadoop-common/CHANGES.txt @@ -29,6 +29,13 @@ Release 2.7.3 - UNRELEASED HADOOP-12794. Support additional compression levels for GzipCodec. (Ravi Mutyala via junping_du) +HADOOP-13103. Group resolution from LDAP may fail on +javax.naming.ServiceUnavailableException (szetszwo) + +HADOOP-13039. Add documentation for configuration property +ipc.maximum.data.length for controlling maximum RPC message size. +(Mingliang Liu via arp) + OPTIMIZATIONS HADOOP-12810. FileSystem#listLocatedStatus causes unnecessary RPC calls @@ -126,6 +133,30 @@ Release 2.7.3 - UNRELEASED HADOOP-13297. Add missing dependency in setting maven-remote-resource-plugin to fix builds. (Sean Busbey via aajisaka) +HADOOP-13255. KMSClientProvider should check and renew tgt when doing +delegation token operations. (xyao) + +HADOOP-12559. KMS connection failures should trigger TGT renewal. +(Zhe Zhang via xyao) + +HADOOP-13043. Add LICENSE.txt entries for bundled javascript dependencies. +(wang) + +HADOOP-13042. Restore lost leveldbjni LICENSE and NOTICE changes. (wang) + +HADOOP-12406. AbstractMapWritable.readFields throws ClassNotFoundException +with custom writables. (Nadeem Douba via vinodkv) + +HADOOP-12636. Fix to prevent ServiceLoader failure init for unused FileSystems. +(Inigo Goiri via cdouglas) + +HADOOP-13350. Additional fixes to LICENSE and NOTICE. (Xiao Chen via wang) + +HADOOP-12682. Fixed TestKMS#testKMSRestart* failure. (Wei-Chiu Chuang via xyao) + +HADOOP-13312. Updated CHANGES.txt to reflect all the changes in branch-2.7. +(Akira Ajisaka via vinodkv) + Release 2.7.2 - 2016-01-25 INCOMPATIBLE CHANGES http://git-wip-us.apache.org/repos/asf/hadoop/blob/c454639a/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt -- diff --git a/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt b/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt index 6746542..c5625aa 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt +++ b/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt @@ -61,6 +61,20 @@ Release 2.7.3 - UNRELEASED HDFS-8772. Fix TestStandbyIsHot#testDatanodeRestarts which occasionally fails. (Walter Su via wang). +HDFS-9906. Remove spammy log spew when a datanode is restarted. +(Brahma Reddy Battula via arp) + +HDFS-2580. NameNode#main(...) can make use of GenericOptionsParser. (harsh) + +HDFS-10264. Logging improvements in FSImageFormatProtobuf.Saver. +(Xiaobing Zhou via arp) + +HDFS-10186. DirectoryScanner: Improve logs by adding full path of both +actual and expected block directories. (Rakesh R via szetszwo) + +HDFS-10488. Updated WebHDFS documentation regarding CREATE and MKDIR default +permissions. (Wellington Chevreuil via wang) + OPTIMIZATIONS HDFS-8845. DiskChecker should not traverse the entire tree (Chang Li via @@ -204,6 +218,32 @@ Release 2.7.3 - UNRELEASED HDFS-9530. ReservedSpace is not cleared for abandoned Blocks (Brahma Reddy Battula) +HDFS-10458. getFileEncryptionInfo should return quickly for non-encrypted +cluster. (zhz) + +HDFS-9365. Balaner does not work with the HDFS-6376 HA setup. (szetszwo) + +HDFS-10335. Mover$Processor#chooseTarget() always chooses the first +matching targe
[2/2] hadoop git commit: HDFS-10488. Update WebHDFS documentation regarding CREATE and MKDIR default permissions. Contributed by Wellington Chevreuil.
HDFS-10488. Update WebHDFS documentation regarding CREATE and MKDIR default permissions. Contributed by Wellington Chevreuil. (cherry picked from commit 728623242f47c6bffb46cfcc8f3425b58c1fde3a) (cherry picked from commit 5ebebc652454fd5541d1a8231ba7ac2656e36cba) Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/490bcda7 Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/490bcda7 Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/490bcda7 Branch: refs/heads/branch-2.7.3 Commit: 490bcda71fc607dbc9d2785695e43f251571bb24 Parents: fcd058e Author: Andrew WangAuthored: Thu Jul 7 10:36:18 2016 -0700 Committer: Vinod Kumar Vavilapalli Committed: Mon Jul 11 15:28:37 2016 -0700 -- hadoop-hdfs-project/hadoop-hdfs/src/site/markdown/WebHDFS.md | 4 1 file changed, 4 insertions(+) -- http://git-wip-us.apache.org/repos/asf/hadoop/blob/490bcda7/hadoop-hdfs-project/hadoop-hdfs/src/site/markdown/WebHDFS.md -- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/site/markdown/WebHDFS.md b/hadoop-hdfs-project/hadoop-hdfs/src/site/markdown/WebHDFS.md index efab9eb..62d3237 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/site/markdown/WebHDFS.md +++ b/hadoop-hdfs-project/hadoop-hdfs/src/site/markdown/WebHDFS.md @@ -265,6 +265,8 @@ File and Directory Operations Location: webhdfs://:/ Content-Length: 0 +If no permissions are specified, the newly created file will be assigned with default 755 permission. This also applies for new directories. No umask mode will be applied from server side (so "fs.permissions.umask-mode" value configuration set on Namenode side will have no effect). + **Note** that the reason of having two-step create/append is for preventing clients to send out data before the redirect. This issue is addressed by the "`Expect: 100-continue`" header in HTTP/1.1; see [RFC 2616, Section 8.2.3](http://www.w3.org/Protocols/rfc2616/rfc2616-sec8.html#sec8.2.3). Unfortunately, there are software library bugs (e.g. Jetty 6 HTTP server and Java 6 HTTP client), which do not correctly implement "`Expect: 100-continue`". The two-step create/append is a temporary workaround for the software library bugs. See also: [`overwrite`](#Overwrite), [`blocksize`](#Block_Size), [`replication`](#Replication), [`permission`](#Permission), [`buffersize`](#Buffer_Size), [FileSystem](../../api/org/apache/hadoop/fs/FileSystem.html).create @@ -344,6 +346,8 @@ See also: [`offset`](#Offset), [`length`](#Length), [`buffersize`](#Buffer_Size) {"boolean": true} +If no permissions are specified, the newly created directory will be assigned with default 755 permission. This also applies to new files created. No umask mode will be applied from server side (so "fs.permissions.umask-mode" value configuration set on Namenode side will have no effect). + See also: [`permission`](#Permission), [FileSystem](../../api/org/apache/hadoop/fs/FileSystem.html).mkdirs ### Create a Symbolic Link - To unsubscribe, e-mail: common-commits-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-commits-h...@hadoop.apache.org
[1/2] hadoop git commit: HADOOP-12636. Prevent ServiceLoader failure init for unused FileSystems. Contributed by Inigo Goiri. Cherry picked from commit #8652cce5b2
Repository: hadoop Updated Branches: refs/heads/branch-2.7.3 01aa7d8f2 -> 490bcda71 HADOOP-12636. Prevent ServiceLoader failure init for unused FileSystems. Contributed by Inigo Goiri. Cherry picked from commit #8652cce5b2 (cherry picked from commit 7595d89612f4916ab29a6ac4e07374771fcbc94f) Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/fcd058ed Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/fcd058ed Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/fcd058ed Branch: refs/heads/branch-2.7.3 Commit: fcd058edb60c7a02ad07d0fb8ccf85befe212308 Parents: 01aa7d8 Author: Chris DouglasAuthored: Fri Dec 18 18:21:52 2015 -0800 Committer: Vinod Kumar Vavilapalli Committed: Mon Jul 11 15:28:01 2016 -0700 -- .../java/org/apache/hadoop/fs/FileSystem.java | 18 -- .../hadoop/fs/TestFileSystemInitialization.java | 16 2 files changed, 32 insertions(+), 2 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hadoop/blob/fcd058ed/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FileSystem.java -- diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FileSystem.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FileSystem.java index fac3c40..2e684f5 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FileSystem.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FileSystem.java @@ -35,6 +35,7 @@ import java.util.Iterator; import java.util.List; import java.util.Map; import java.util.NoSuchElementException; +import java.util.ServiceConfigurationError; import java.util.ServiceLoader; import java.util.Set; import java.util.Stack; @@ -61,6 +62,7 @@ import org.apache.hadoop.security.Credentials; import org.apache.hadoop.security.SecurityUtil; import org.apache.hadoop.security.UserGroupInformation; import org.apache.hadoop.security.token.Token; +import org.apache.hadoop.util.ClassUtil; import org.apache.hadoop.util.DataChecksum; import org.apache.hadoop.util.Progressable; import org.apache.hadoop.util.ReflectionUtils; @@ -2622,8 +2624,20 @@ public abstract class FileSystem extends Configured implements Closeable { synchronized (FileSystem.class) { if (!FILE_SYSTEMS_LOADED) { ServiceLoader serviceLoader = ServiceLoader.load(FileSystem.class); -for (FileSystem fs : serviceLoader) { - SERVICE_FILE_SYSTEMS.put(fs.getScheme(), fs.getClass()); +Iterator it = serviceLoader.iterator(); +while (it.hasNext()) { + FileSystem fs = null; + try { +fs = it.next(); +try { + SERVICE_FILE_SYSTEMS.put(fs.getScheme(), fs.getClass()); +} catch (Exception e) { + LOG.warn("Cannot load: " + fs + " from " + + ClassUtil.findContainingJar(fs.getClass()), e); +} + } catch (ServiceConfigurationError ee) { +LOG.warn("Cannot load filesystem", ee); + } } FILE_SYSTEMS_LOADED = true; } http://git-wip-us.apache.org/repos/asf/hadoop/blob/fcd058ed/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/TestFileSystemInitialization.java -- diff --git a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/TestFileSystemInitialization.java b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/TestFileSystemInitialization.java index d3fceec..18e8b01 100644 --- a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/TestFileSystemInitialization.java +++ b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/TestFileSystemInitialization.java @@ -21,6 +21,7 @@ import org.apache.hadoop.conf.Configuration; import java.io.IOException; import java.net.URL; +import java.util.ServiceConfigurationError; import org.junit.Test; import static org.junit.Assert.*; @@ -43,4 +44,19 @@ public class TestFileSystemInitialization { assertFalse(false); } } + + @Test + public void testMissingLibraries() { +boolean catched = false; +try { + Configuration conf = new Configuration(); + FileSystem.getFileSystemClass("s3a", conf); +} catch (Exception e) { + catched = true; +} catch (ServiceConfigurationError e) { + // S3A shouldn't find AWS SDK and fail + catched = true; +} +assertTrue(catched); + } } - To unsubscribe, e-mail:
hadoop git commit: YARN-5314. Fixed a ConcurrentModificationException in ATS v1.5 EntityGroupFSTimelineStore. Contributed by Li Lu.
Repository: hadoop Updated Branches: refs/heads/branch-2 b3b81b4f0 -> 687185feb YARN-5314. Fixed a ConcurrentModificationException in ATS v1.5 EntityGroupFSTimelineStore. Contributed by Li Lu. (cherry picked from commit 673e5e02feba9171498a518c06ae70639c5f8854) Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/687185fe Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/687185fe Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/687185fe Branch: refs/heads/branch-2 Commit: 687185feb24e759691b891d1a60eb1410ff0567d Parents: b3b81b4 Author: Vinod Kumar VavilapalliAuthored: Thu Jul 7 22:37:29 2016 -0700 Committer: Vinod Kumar Vavilapalli Committed: Thu Jul 7 22:38:41 2016 -0700 -- .../timeline/webapp/TimelineWebServices.java| 2 -- .../yarn/server/timeline/EntityCacheItem.java | 35 ++-- .../timeline/EntityGroupFSTimelineStore.java| 34 +++ .../TestEntityGroupFSTimelineStore.java | 6 ++-- 4 files changed, 34 insertions(+), 43 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hadoop/blob/687185fe/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/src/main/java/org/apache/hadoop/yarn/server/timeline/webapp/TimelineWebServices.java -- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/src/main/java/org/apache/hadoop/yarn/server/timeline/webapp/TimelineWebServices.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/src/main/java/org/apache/hadoop/yarn/server/timeline/webapp/TimelineWebServices.java index e1e684b..5c80132 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/src/main/java/org/apache/hadoop/yarn/server/timeline/webapp/TimelineWebServices.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/src/main/java/org/apache/hadoop/yarn/server/timeline/webapp/TimelineWebServices.java @@ -47,7 +47,6 @@ import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.security.UserGroupInformation; import org.apache.hadoop.util.StringUtils; -import org.apache.hadoop.util.VersionInfo; import org.apache.hadoop.yarn.api.records.timeline.TimelineDomain; import org.apache.hadoop.yarn.api.records.timeline.TimelineDomains; import org.apache.hadoop.yarn.api.records.timeline.TimelineEntities; @@ -61,7 +60,6 @@ import org.apache.hadoop.yarn.server.timeline.NameValuePair; import org.apache.hadoop.yarn.server.timeline.TimelineDataManager; import org.apache.hadoop.yarn.server.timeline.TimelineReader.Field; import org.apache.hadoop.yarn.api.records.timeline.TimelineAbout; -import org.apache.hadoop.yarn.util.YarnVersionInfo; import org.apache.hadoop.yarn.util.timeline.TimelineUtils; import org.apache.hadoop.yarn.webapp.BadRequestException; import org.apache.hadoop.yarn.webapp.ForbiddenException; http://git-wip-us.apache.org/repos/asf/hadoop/blob/687185fe/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timeline-pluginstorage/src/main/java/org/apache/hadoop/yarn/server/timeline/EntityCacheItem.java -- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timeline-pluginstorage/src/main/java/org/apache/hadoop/yarn/server/timeline/EntityCacheItem.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timeline-pluginstorage/src/main/java/org/apache/hadoop/yarn/server/timeline/EntityCacheItem.java index 1566ae2..2b6e023 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timeline-pluginstorage/src/main/java/org/apache/hadoop/yarn/server/timeline/EntityCacheItem.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timeline-pluginstorage/src/main/java/org/apache/hadoop/yarn/server/timeline/EntityCacheItem.java @@ -19,19 +19,13 @@ package org.apache.hadoop.yarn.server.timeline; import com.google.common.annotations.VisibleForTesting; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; -import org.apache.hadoop.fs.FileSystem; -import org.apache.hadoop.fs.Path; import org.apache.hadoop.util.Time; import org.apache.hadoop.yarn.api.records.timeline.TimelineEntityGroupId; import org.apache.hadoop.yarn.server.timeline.security.TimelineACLsManager; -import org.codehaus.jackson.JsonFactory; -import org.codehaus.jackson.map.ObjectMapper; import org.slf4j.Logger; import
hadoop git commit: YARN-5314. Fixed a ConcurrentModificationException in ATS v1.5 EntityGroupFSTimelineStore. Contributed by Li Lu.
Repository: hadoop Updated Branches: refs/heads/trunk 9d46a49c7 -> 673e5e02f YARN-5314. Fixed a ConcurrentModificationException in ATS v1.5 EntityGroupFSTimelineStore. Contributed by Li Lu. Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/673e5e02 Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/673e5e02 Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/673e5e02 Branch: refs/heads/trunk Commit: 673e5e02feba9171498a518c06ae70639c5f8854 Parents: 9d46a49 Author: Vinod Kumar VavilapalliAuthored: Thu Jul 7 22:37:29 2016 -0700 Committer: Vinod Kumar Vavilapalli Committed: Thu Jul 7 22:37:29 2016 -0700 -- .../timeline/webapp/TimelineWebServices.java| 2 -- .../yarn/server/timeline/EntityCacheItem.java | 35 ++-- .../timeline/EntityGroupFSTimelineStore.java| 34 +++ .../TestEntityGroupFSTimelineStore.java | 6 ++-- 4 files changed, 34 insertions(+), 43 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hadoop/blob/673e5e02/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/src/main/java/org/apache/hadoop/yarn/server/timeline/webapp/TimelineWebServices.java -- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/src/main/java/org/apache/hadoop/yarn/server/timeline/webapp/TimelineWebServices.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/src/main/java/org/apache/hadoop/yarn/server/timeline/webapp/TimelineWebServices.java index e1e684b..5c80132 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/src/main/java/org/apache/hadoop/yarn/server/timeline/webapp/TimelineWebServices.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/src/main/java/org/apache/hadoop/yarn/server/timeline/webapp/TimelineWebServices.java @@ -47,7 +47,6 @@ import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.security.UserGroupInformation; import org.apache.hadoop.util.StringUtils; -import org.apache.hadoop.util.VersionInfo; import org.apache.hadoop.yarn.api.records.timeline.TimelineDomain; import org.apache.hadoop.yarn.api.records.timeline.TimelineDomains; import org.apache.hadoop.yarn.api.records.timeline.TimelineEntities; @@ -61,7 +60,6 @@ import org.apache.hadoop.yarn.server.timeline.NameValuePair; import org.apache.hadoop.yarn.server.timeline.TimelineDataManager; import org.apache.hadoop.yarn.server.timeline.TimelineReader.Field; import org.apache.hadoop.yarn.api.records.timeline.TimelineAbout; -import org.apache.hadoop.yarn.util.YarnVersionInfo; import org.apache.hadoop.yarn.util.timeline.TimelineUtils; import org.apache.hadoop.yarn.webapp.BadRequestException; import org.apache.hadoop.yarn.webapp.ForbiddenException; http://git-wip-us.apache.org/repos/asf/hadoop/blob/673e5e02/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timeline-pluginstorage/src/main/java/org/apache/hadoop/yarn/server/timeline/EntityCacheItem.java -- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timeline-pluginstorage/src/main/java/org/apache/hadoop/yarn/server/timeline/EntityCacheItem.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timeline-pluginstorage/src/main/java/org/apache/hadoop/yarn/server/timeline/EntityCacheItem.java index 1566ae2..2b6e023 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timeline-pluginstorage/src/main/java/org/apache/hadoop/yarn/server/timeline/EntityCacheItem.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timeline-pluginstorage/src/main/java/org/apache/hadoop/yarn/server/timeline/EntityCacheItem.java @@ -19,19 +19,13 @@ package org.apache.hadoop.yarn.server.timeline; import com.google.common.annotations.VisibleForTesting; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; -import org.apache.hadoop.fs.FileSystem; -import org.apache.hadoop.fs.Path; import org.apache.hadoop.util.Time; import org.apache.hadoop.yarn.api.records.timeline.TimelineEntityGroupId; import org.apache.hadoop.yarn.server.timeline.security.TimelineACLsManager; -import org.codehaus.jackson.JsonFactory; -import org.codehaus.jackson.map.ObjectMapper; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import java.io.IOException; -import
hadoop git commit: YARN-5224. Added new web-services /containers/{containerid}/logs & /containers/{containerid}/logs/{filename} and using them in "yarn logs" CLI to get logs of finished containers of
Repository: hadoop Updated Branches: refs/heads/branch-2 ccb079190 -> c3d9ac82a YARN-5224. Added new web-services /containers/{containerid}/logs & /containers/{containerid}/logs/{filename} and using them in "yarn logs" CLI to get logs of finished containers of a running application. Contributed by Xuan Gong. Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/c3d9ac82 Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/c3d9ac82 Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/c3d9ac82 Branch: refs/heads/branch-2 Commit: c3d9ac82af8ee425971858efac569a52731fba81 Parents: ccb0791 Author: Vinod Kumar VavilapalliAuthored: Wed Jul 6 14:19:16 2016 -0700 Committer: Vinod Kumar Vavilapalli Committed: Wed Jul 6 14:19:16 2016 -0700 -- .../apache/hadoop/yarn/client/cli/LogsCLI.java | 32 ++ .../webapp/AHSWebServices.java | 4 +- .../webapp/TestAHSWebServices.java | 3 +- .../nodemanager/webapp/NMWebServices.java | 66 ++- .../webapp/dao/ContainerLogsInfo.java | 112 +++ .../nodemanager/webapp/TestNMWebServices.java | 71 +++- 6 files changed, 236 insertions(+), 52 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hadoop/blob/c3d9ac82/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/LogsCLI.java -- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/LogsCLI.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/LogsCLI.java index ad4964a..4a2dac0 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/LogsCLI.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/LogsCLI.java @@ -20,7 +20,6 @@ package org.apache.hadoop.yarn.client.cli; import java.io.IOException; import java.io.PrintStream; -import java.io.StringReader; import java.util.ArrayList; import java.util.Arrays; import java.util.Collection; @@ -29,9 +28,6 @@ import java.util.List; import java.util.Set; import java.util.regex.Pattern; import javax.ws.rs.core.MediaType; -import javax.xml.parsers.DocumentBuilder; -import javax.xml.parsers.DocumentBuilderFactory; - import org.apache.commons.cli.CommandLine; import org.apache.commons.cli.CommandLineParser; import org.apache.commons.cli.GnuParser; @@ -71,9 +67,6 @@ import com.sun.jersey.api.client.ClientHandlerException; import com.sun.jersey.api.client.ClientResponse; import com.sun.jersey.api.client.UniformInterfaceException; import com.sun.jersey.api.client.WebResource; -import org.w3c.dom.Document; -import org.w3c.dom.NodeList; -import org.xml.sax.InputSource; @Public @Evolving @@ -353,22 +346,20 @@ public class LogsCLI extends Configured implements Tool { .resource(WebAppUtils.getHttpSchemePrefix(conf) + nodeHttpAddress); ClientResponse response = webResource.path("ws").path("v1").path("node").path("containers") - .path(containerIdStr).accept(MediaType.APPLICATION_XML) + .path(containerIdStr).path("logs") + .accept(MediaType.APPLICATION_JSON) .get(ClientResponse.class); - if (response.getClientResponseStatus().equals(ClientResponse.Status.OK)) { + if (response.getClientResponseStatus().equals( + ClientResponse.Status.OK)) { try { - String xml = response.getEntity(String.class); - DocumentBuilderFactory dbf = DocumentBuilderFactory.newInstance(); - DocumentBuilder db = dbf.newDocumentBuilder(); - InputSource is = new InputSource(); - is.setCharacterStream(new StringReader(xml)); - Document dom = db.parse(is); - NodeList elements = dom.getElementsByTagName("containerLogFiles"); - for (int i = 0; i < elements.getLength(); i++) { -logFiles.add(elements.item(i).getTextContent()); + JSONObject json = + response.getEntity(JSONObject.class); + JSONArray array = json.getJSONArray("containerLogInfo"); + for (int i = 0; i < array.length(); i++) { +logFiles.add(array.getJSONObject(i).getString("fileName")); } } catch (Exception e) { - System.err.println("Unable to parse xml from webservice. Error:"); + System.err.println("Unable to parse json from webservice. Error:"); System.err.println(e.getMessage()); throw new IOException(e); } @@ -424,7 +415,8 @@ public class LogsCLI extends
hadoop git commit: YARN-5224. Added new web-services /containers/{containerid}/logs & /containers/{containerid}/logs/{filename} and using them in "yarn logs" CLI to get logs of finished containers of
Repository: hadoop Updated Branches: refs/heads/trunk d169f5052 -> 4c9e1aeb9 YARN-5224. Added new web-services /containers/{containerid}/logs & /containers/{containerid}/logs/{filename} and using them in "yarn logs" CLI to get logs of finished containers of a running application. Contributed by Xuan Gong. Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/4c9e1aeb Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/4c9e1aeb Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/4c9e1aeb Branch: refs/heads/trunk Commit: 4c9e1aeb94247a6e97215e902bdc71a325244243 Parents: d169f50 Author: Vinod Kumar VavilapalliAuthored: Wed Jul 6 14:07:54 2016 -0700 Committer: Vinod Kumar Vavilapalli Committed: Wed Jul 6 14:10:54 2016 -0700 -- .../apache/hadoop/yarn/client/cli/LogsCLI.java | 28 ++--- .../webapp/AHSWebServices.java | 4 +- .../webapp/TestAHSWebServices.java | 3 +- .../nodemanager/webapp/NMWebServices.java | 66 ++- .../webapp/dao/ContainerLogsInfo.java | 112 +++ .../nodemanager/webapp/TestNMWebServices.java | 71 +++- 6 files changed, 234 insertions(+), 50 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hadoop/blob/4c9e1aeb/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/LogsCLI.java -- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/LogsCLI.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/LogsCLI.java index 76d1fb3..86187ca 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/LogsCLI.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/LogsCLI.java @@ -20,7 +20,6 @@ package org.apache.hadoop.yarn.client.cli; import java.io.IOException; import java.io.PrintStream; -import java.io.StringReader; import java.util.ArrayList; import java.util.Arrays; import java.util.Collection; @@ -29,9 +28,6 @@ import java.util.List; import java.util.Set; import java.util.regex.Pattern; import javax.ws.rs.core.MediaType; -import javax.xml.parsers.DocumentBuilder; -import javax.xml.parsers.DocumentBuilderFactory; - import org.apache.commons.cli.CommandLine; import org.apache.commons.cli.CommandLineParser; import org.apache.commons.cli.GnuParser; @@ -71,9 +67,6 @@ import com.sun.jersey.api.client.ClientHandlerException; import com.sun.jersey.api.client.ClientResponse; import com.sun.jersey.api.client.UniformInterfaceException; import com.sun.jersey.api.client.WebResource; -import org.w3c.dom.Document; -import org.w3c.dom.NodeList; -import org.xml.sax.InputSource; @Public @Evolving @@ -353,23 +346,19 @@ public class LogsCLI extends Configured implements Tool { .resource(WebAppUtils.getHttpSchemePrefix(conf) + nodeHttpAddress); ClientResponse response = webResource.path("ws").path("v1").path("node").path("containers") - .path(containerIdStr).accept(MediaType.APPLICATION_XML) + .path(containerIdStr).path("logs") + .accept(MediaType.APPLICATION_JSON) .get(ClientResponse.class); if (response.getStatusInfo().getStatusCode() == ClientResponse.Status.OK.getStatusCode()) { try { - String xml = response.getEntity(String.class); - DocumentBuilderFactory dbf = DocumentBuilderFactory.newInstance(); - DocumentBuilder db = dbf.newDocumentBuilder(); - InputSource is = new InputSource(); - is.setCharacterStream(new StringReader(xml)); - Document dom = db.parse(is); - NodeList elements = dom.getElementsByTagName("containerLogFiles"); - for (int i = 0; i < elements.getLength(); i++) { -logFiles.add(elements.item(i).getTextContent()); + JSONObject json = response.getEntity(JSONObject.class); + JSONArray array = json.getJSONArray("containerLogInfo"); + for (int i = 0; i < array.length(); i++) { +logFiles.add(array.getJSONObject(i).getString("fileName")); } } catch (Exception e) { - System.err.println("Unable to parse xml from webservice. Error:"); + System.err.println("Unable to parse json from webservice. Error:"); System.err.println(e.getMessage()); throw new IOException(e); } @@ -425,7 +414,8 @@ public class LogsCLI extends Configured implements Tool { + nodeHttpAddress); ClientResponse
hadoop git commit: YARN-5214. Fixed locking in DirectoryCollection to avoid hanging NMs when various code-paths hit slow disks. Contributed by Junping Du.
Repository: hadoop Updated Branches: refs/heads/branch-2.8 4b263b30c -> 4ea87cb38 YARN-5214. Fixed locking in DirectoryCollection to avoid hanging NMs when various code-paths hit slow disks. Contributed by Junping Du. (cherry picked from commit ce9c006430d13a28bc1ca57c5c70cc1b7cba1692) Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/4ea87cb3 Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/4ea87cb3 Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/4ea87cb3 Branch: refs/heads/branch-2.8 Commit: 4ea87cb38c5d1f673e1bc82cdf157a1e2d1de9d4 Parents: 4b263b3 Author: Vinod Kumar VavilapalliAuthored: Tue Jul 5 16:07:28 2016 -0700 Committer: Vinod Kumar Vavilapalli Committed: Tue Jul 5 17:12:37 2016 -0700 -- .../server/nodemanager/DirectoryCollection.java | 221 --- 1 file changed, 145 insertions(+), 76 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hadoop/blob/4ea87cb3/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/DirectoryCollection.java -- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/DirectoryCollection.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/DirectoryCollection.java index a2bfd20..417b207 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/DirectoryCollection.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/DirectoryCollection.java @@ -28,12 +28,17 @@ import java.util.HashSet; import java.util.List; import java.util.Map; import java.util.Set; +import java.util.concurrent.ConcurrentHashMap; import java.util.concurrent.CopyOnWriteArrayList; +import java.util.concurrent.locks.ReentrantReadWriteLock; +import java.util.concurrent.locks.ReentrantReadWriteLock.ReadLock; +import java.util.concurrent.locks.ReentrantReadWriteLock.WriteLock; import org.apache.commons.io.FileUtils; import org.apache.commons.lang.RandomStringUtils; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.fs.FileAlreadyExistsException; import org.apache.hadoop.fs.FileContext; import org.apache.hadoop.fs.Path; import org.apache.hadoop.fs.permission.FsPermission; @@ -89,6 +94,10 @@ public class DirectoryCollection { private List errorDirs; private List fullDirs; + // read/write lock for accessing above directories. + private final ReadLock readLock; + private final WriteLock writeLock; + private int numFailures; private float diskUtilizationPercentageCutoffHigh; @@ -163,9 +172,13 @@ public class DirectoryCollection { float utilizationPercentageCutOffHigh, float utilizationPercentageCutOffLow, long utilizationSpaceCutOff) { -localDirs = new CopyOnWriteArrayList(dirs); -errorDirs = new CopyOnWriteArrayList(); -fullDirs = new CopyOnWriteArrayList(); +localDirs = new CopyOnWriteArrayList<>(dirs); +errorDirs = new CopyOnWriteArrayList<>(); +fullDirs = new CopyOnWriteArrayList<>(); + +ReentrantReadWriteLock lock = new ReentrantReadWriteLock(); +this.readLock = lock.readLock(); +this.writeLock = lock.writeLock(); diskUtilizationPercentageCutoffHigh = Math.max(0.0F, Math.min(100.0F, utilizationPercentageCutOffHigh)); @@ -174,17 +187,18 @@ public class DirectoryCollection { diskUtilizationSpaceCutoff = utilizationSpaceCutOff < 0 ? 0 : utilizationSpaceCutOff; -dirsChangeListeners = new HashSet(); +dirsChangeListeners = Collections.newSetFromMap( +new ConcurrentHashMap ()); } - synchronized void registerDirsChangeListener( + void registerDirsChangeListener( DirsChangeListener listener) { if (dirsChangeListeners.add(listener)) { listener.onDirsChanged(); } } - synchronized void deregisterDirsChangeListener( + void deregisterDirsChangeListener( DirsChangeListener listener) { dirsChangeListeners.remove(listener); } @@ -192,31 +206,51 @@ public class DirectoryCollection { /** * @return the current valid directories */ - synchronized List getGoodDirs() { -return Collections.unmodifiableList(localDirs); + List getGoodDirs() { +this.readLock.lock(); +try { + return
hadoop git commit: YARN-5214. Fixed locking in DirectoryCollection to avoid hanging NMs when various code-paths hit slow disks. Contributed by Junping Du.
Repository: hadoop Updated Branches: refs/heads/branch-2 516aa58c1 -> 56a08983e YARN-5214. Fixed locking in DirectoryCollection to avoid hanging NMs when various code-paths hit slow disks. Contributed by Junping Du. (cherry picked from commit ce9c006430d13a28bc1ca57c5c70cc1b7cba1692) Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/56a08983 Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/56a08983 Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/56a08983 Branch: refs/heads/branch-2 Commit: 56a08983e26a0f7af9f9a5e8ba71ef12d560d7c7 Parents: 516aa58 Author: Vinod Kumar VavilapalliAuthored: Tue Jul 5 16:07:28 2016 -0700 Committer: Vinod Kumar Vavilapalli Committed: Tue Jul 5 16:08:57 2016 -0700 -- .../server/nodemanager/DirectoryCollection.java | 221 --- 1 file changed, 145 insertions(+), 76 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hadoop/blob/56a08983/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/DirectoryCollection.java -- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/DirectoryCollection.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/DirectoryCollection.java index a2bfd20..417b207 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/DirectoryCollection.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/DirectoryCollection.java @@ -28,12 +28,17 @@ import java.util.HashSet; import java.util.List; import java.util.Map; import java.util.Set; +import java.util.concurrent.ConcurrentHashMap; import java.util.concurrent.CopyOnWriteArrayList; +import java.util.concurrent.locks.ReentrantReadWriteLock; +import java.util.concurrent.locks.ReentrantReadWriteLock.ReadLock; +import java.util.concurrent.locks.ReentrantReadWriteLock.WriteLock; import org.apache.commons.io.FileUtils; import org.apache.commons.lang.RandomStringUtils; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.fs.FileAlreadyExistsException; import org.apache.hadoop.fs.FileContext; import org.apache.hadoop.fs.Path; import org.apache.hadoop.fs.permission.FsPermission; @@ -89,6 +94,10 @@ public class DirectoryCollection { private List errorDirs; private List fullDirs; + // read/write lock for accessing above directories. + private final ReadLock readLock; + private final WriteLock writeLock; + private int numFailures; private float diskUtilizationPercentageCutoffHigh; @@ -163,9 +172,13 @@ public class DirectoryCollection { float utilizationPercentageCutOffHigh, float utilizationPercentageCutOffLow, long utilizationSpaceCutOff) { -localDirs = new CopyOnWriteArrayList(dirs); -errorDirs = new CopyOnWriteArrayList(); -fullDirs = new CopyOnWriteArrayList(); +localDirs = new CopyOnWriteArrayList<>(dirs); +errorDirs = new CopyOnWriteArrayList<>(); +fullDirs = new CopyOnWriteArrayList<>(); + +ReentrantReadWriteLock lock = new ReentrantReadWriteLock(); +this.readLock = lock.readLock(); +this.writeLock = lock.writeLock(); diskUtilizationPercentageCutoffHigh = Math.max(0.0F, Math.min(100.0F, utilizationPercentageCutOffHigh)); @@ -174,17 +187,18 @@ public class DirectoryCollection { diskUtilizationSpaceCutoff = utilizationSpaceCutOff < 0 ? 0 : utilizationSpaceCutOff; -dirsChangeListeners = new HashSet(); +dirsChangeListeners = Collections.newSetFromMap( +new ConcurrentHashMap ()); } - synchronized void registerDirsChangeListener( + void registerDirsChangeListener( DirsChangeListener listener) { if (dirsChangeListeners.add(listener)) { listener.onDirsChanged(); } } - synchronized void deregisterDirsChangeListener( + void deregisterDirsChangeListener( DirsChangeListener listener) { dirsChangeListeners.remove(listener); } @@ -192,31 +206,51 @@ public class DirectoryCollection { /** * @return the current valid directories */ - synchronized List getGoodDirs() { -return Collections.unmodifiableList(localDirs); + List getGoodDirs() { +this.readLock.lock(); +try { + return Collections.unmodifiableList(localDirs); +
hadoop git commit: YARN-5214. Fixed locking in DirectoryCollection to avoid hanging NMs when various code-paths hit slow disks. Contributed by Junping Du.
Repository: hadoop Updated Branches: refs/heads/trunk 9560f252c -> ce9c00643 YARN-5214. Fixed locking in DirectoryCollection to avoid hanging NMs when various code-paths hit slow disks. Contributed by Junping Du. Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/ce9c0064 Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/ce9c0064 Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/ce9c0064 Branch: refs/heads/trunk Commit: ce9c006430d13a28bc1ca57c5c70cc1b7cba1692 Parents: 9560f25 Author: Vinod Kumar VavilapalliAuthored: Tue Jul 5 16:07:28 2016 -0700 Committer: Vinod Kumar Vavilapalli Committed: Tue Jul 5 16:07:28 2016 -0700 -- .../server/nodemanager/DirectoryCollection.java | 221 --- 1 file changed, 145 insertions(+), 76 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hadoop/blob/ce9c0064/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/DirectoryCollection.java -- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/DirectoryCollection.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/DirectoryCollection.java index a2bfd20..417b207 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/DirectoryCollection.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/DirectoryCollection.java @@ -28,12 +28,17 @@ import java.util.HashSet; import java.util.List; import java.util.Map; import java.util.Set; +import java.util.concurrent.ConcurrentHashMap; import java.util.concurrent.CopyOnWriteArrayList; +import java.util.concurrent.locks.ReentrantReadWriteLock; +import java.util.concurrent.locks.ReentrantReadWriteLock.ReadLock; +import java.util.concurrent.locks.ReentrantReadWriteLock.WriteLock; import org.apache.commons.io.FileUtils; import org.apache.commons.lang.RandomStringUtils; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.fs.FileAlreadyExistsException; import org.apache.hadoop.fs.FileContext; import org.apache.hadoop.fs.Path; import org.apache.hadoop.fs.permission.FsPermission; @@ -89,6 +94,10 @@ public class DirectoryCollection { private List errorDirs; private List fullDirs; + // read/write lock for accessing above directories. + private final ReadLock readLock; + private final WriteLock writeLock; + private int numFailures; private float diskUtilizationPercentageCutoffHigh; @@ -163,9 +172,13 @@ public class DirectoryCollection { float utilizationPercentageCutOffHigh, float utilizationPercentageCutOffLow, long utilizationSpaceCutOff) { -localDirs = new CopyOnWriteArrayList(dirs); -errorDirs = new CopyOnWriteArrayList(); -fullDirs = new CopyOnWriteArrayList(); +localDirs = new CopyOnWriteArrayList<>(dirs); +errorDirs = new CopyOnWriteArrayList<>(); +fullDirs = new CopyOnWriteArrayList<>(); + +ReentrantReadWriteLock lock = new ReentrantReadWriteLock(); +this.readLock = lock.readLock(); +this.writeLock = lock.writeLock(); diskUtilizationPercentageCutoffHigh = Math.max(0.0F, Math.min(100.0F, utilizationPercentageCutOffHigh)); @@ -174,17 +187,18 @@ public class DirectoryCollection { diskUtilizationSpaceCutoff = utilizationSpaceCutOff < 0 ? 0 : utilizationSpaceCutOff; -dirsChangeListeners = new HashSet(); +dirsChangeListeners = Collections.newSetFromMap( +new ConcurrentHashMap ()); } - synchronized void registerDirsChangeListener( + void registerDirsChangeListener( DirsChangeListener listener) { if (dirsChangeListeners.add(listener)) { listener.onDirsChanged(); } } - synchronized void deregisterDirsChangeListener( + void deregisterDirsChangeListener( DirsChangeListener listener) { dirsChangeListeners.remove(listener); } @@ -192,31 +206,51 @@ public class DirectoryCollection { /** * @return the current valid directories */ - synchronized List getGoodDirs() { -return Collections.unmodifiableList(localDirs); + List getGoodDirs() { +this.readLock.lock(); +try { + return Collections.unmodifiableList(localDirs); +} finally { + this.readLock.unlock(); +} } /** *
hadoop git commit: Preparing for 2.7.3 release: Updating version to 2.7.3 on branch-2.7.3.
Repository: hadoop Updated Branches: refs/heads/branch-2.7.3 [created] ff327bc7f Preparing for 2.7.3 release: Updating version to 2.7.3 on branch-2.7.3. Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/ff327bc7 Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/ff327bc7 Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/ff327bc7 Branch: refs/heads/branch-2.7.3 Commit: ff327bc7fc990bff26e8af68077fc31ce95e7b8e Parents: 8854cdd Author: Vinod Kumar VavilapalliAuthored: Tue Jun 14 19:25:45 2016 -0700 Committer: Vinod Kumar Vavilapalli Committed: Tue Jun 14 19:25:45 2016 -0700 -- hadoop-assemblies/pom.xml| 4 ++-- hadoop-client/pom.xml| 4 ++-- hadoop-common-project/hadoop-annotations/pom.xml | 4 ++-- hadoop-common-project/hadoop-auth-examples/pom.xml | 4 ++-- hadoop-common-project/hadoop-auth/pom.xml| 4 ++-- hadoop-common-project/hadoop-common/pom.xml | 4 ++-- hadoop-common-project/hadoop-kms/pom.xml | 4 ++-- hadoop-common-project/hadoop-minikdc/pom.xml | 4 ++-- hadoop-common-project/hadoop-nfs/pom.xml | 4 ++-- hadoop-common-project/pom.xml| 4 ++-- hadoop-dist/pom.xml | 4 ++-- hadoop-hdfs-project/hadoop-hdfs-httpfs/pom.xml | 4 ++-- hadoop-hdfs-project/hadoop-hdfs-nfs/pom.xml | 4 ++-- hadoop-hdfs-project/hadoop-hdfs/pom.xml | 4 ++-- hadoop-hdfs-project/hadoop-hdfs/src/contrib/bkjournal/pom.xml| 4 ++-- hadoop-hdfs-project/pom.xml | 4 ++-- .../hadoop-mapreduce-client/hadoop-mapreduce-client-app/pom.xml | 4 ++-- .../hadoop-mapreduce-client-common/pom.xml | 4 ++-- .../hadoop-mapreduce-client/hadoop-mapreduce-client-core/pom.xml | 4 ++-- .../hadoop-mapreduce-client-hs-plugins/pom.xml | 4 ++-- .../hadoop-mapreduce-client/hadoop-mapreduce-client-hs/pom.xml | 4 ++-- .../hadoop-mapreduce-client-jobclient/pom.xml| 4 ++-- .../hadoop-mapreduce-client-shuffle/pom.xml | 4 ++-- hadoop-mapreduce-project/hadoop-mapreduce-client/pom.xml | 4 ++-- hadoop-mapreduce-project/hadoop-mapreduce-examples/pom.xml | 4 ++-- hadoop-mapreduce-project/pom.xml | 4 ++-- hadoop-maven-plugins/pom.xml | 2 +- hadoop-minicluster/pom.xml | 4 ++-- hadoop-project-dist/pom.xml | 4 ++-- hadoop-project/pom.xml | 4 ++-- hadoop-project/src/site/markdown/index.md.vm | 2 +- hadoop-tools/hadoop-ant/pom.xml | 4 ++-- hadoop-tools/hadoop-archives/pom.xml | 4 ++-- hadoop-tools/hadoop-aws/pom.xml | 4 ++-- hadoop-tools/hadoop-azure/pom.xml| 2 +- hadoop-tools/hadoop-datajoin/pom.xml | 4 ++-- hadoop-tools/hadoop-distcp/pom.xml | 4 ++-- hadoop-tools/hadoop-extras/pom.xml | 4 ++-- hadoop-tools/hadoop-gridmix/pom.xml | 4 ++-- hadoop-tools/hadoop-openstack/pom.xml| 4 ++-- hadoop-tools/hadoop-pipes/pom.xml| 4 ++-- hadoop-tools/hadoop-rumen/pom.xml| 4 ++-- hadoop-tools/hadoop-sls/pom.xml | 4 ++-- hadoop-tools/hadoop-streaming/pom.xml| 4 ++-- hadoop-tools/hadoop-tools-dist/pom.xml | 4 ++-- hadoop-tools/pom.xml | 4 ++-- hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/pom.xml | 4 ++-- .../hadoop-yarn-applications-distributedshell/pom.xml| 4 ++-- .../hadoop-yarn-applications-unmanaged-am-launcher/pom.xml | 4 ++-- hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/pom.xml | 4 ++-- hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/pom.xml | 4 ++-- hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/pom.xml | 4 ++-- hadoop-yarn-project/hadoop-yarn/hadoop-yarn-registry/pom.xml | 4 ++-- .../hadoop-yarn-server-applicationhistoryservice/pom.xml | 4 ++-- .../hadoop-yarn-server/hadoop-yarn-server-common/pom.xml | 4 ++-- .../hadoop-yarn-server/hadoop-yarn-server-nodemanager/pom.xml| 4 ++--
hadoop git commit: Preparing for 2.7.4 development: Updating snapshot version to 2.7.4 on branch-2.7.
Repository: hadoop Updated Branches: refs/heads/branch-2.7 8854cdd9e -> f4b58eae9 Preparing for 2.7.4 development: Updating snapshot version to 2.7.4 on branch-2.7. Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/f4b58eae Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/f4b58eae Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/f4b58eae Branch: refs/heads/branch-2.7 Commit: f4b58eae94d40c8a8acc9cf704630849a84e4a44 Parents: 8854cdd Author: Vinod Kumar VavilapalliAuthored: Tue Jun 14 19:13:43 2016 -0700 Committer: Vinod Kumar Vavilapalli Committed: Tue Jun 14 19:14:02 2016 -0700 -- hadoop-assemblies/pom.xml| 4 ++-- hadoop-client/pom.xml| 4 ++-- hadoop-common-project/hadoop-annotations/pom.xml | 4 ++-- hadoop-common-project/hadoop-auth-examples/pom.xml | 4 ++-- hadoop-common-project/hadoop-auth/pom.xml| 4 ++-- hadoop-common-project/hadoop-common/pom.xml | 4 ++-- hadoop-common-project/hadoop-kms/pom.xml | 4 ++-- hadoop-common-project/hadoop-minikdc/pom.xml | 4 ++-- hadoop-common-project/hadoop-nfs/pom.xml | 4 ++-- hadoop-common-project/pom.xml| 4 ++-- hadoop-dist/pom.xml | 4 ++-- hadoop-hdfs-project/hadoop-hdfs-httpfs/pom.xml | 4 ++-- hadoop-hdfs-project/hadoop-hdfs-nfs/pom.xml | 4 ++-- hadoop-hdfs-project/hadoop-hdfs/pom.xml | 4 ++-- hadoop-hdfs-project/hadoop-hdfs/src/contrib/bkjournal/pom.xml| 4 ++-- hadoop-hdfs-project/pom.xml | 4 ++-- .../hadoop-mapreduce-client/hadoop-mapreduce-client-app/pom.xml | 4 ++-- .../hadoop-mapreduce-client-common/pom.xml | 4 ++-- .../hadoop-mapreduce-client/hadoop-mapreduce-client-core/pom.xml | 4 ++-- .../hadoop-mapreduce-client-hs-plugins/pom.xml | 4 ++-- .../hadoop-mapreduce-client/hadoop-mapreduce-client-hs/pom.xml | 4 ++-- .../hadoop-mapreduce-client-jobclient/pom.xml| 4 ++-- .../hadoop-mapreduce-client-shuffle/pom.xml | 4 ++-- hadoop-mapreduce-project/hadoop-mapreduce-client/pom.xml | 4 ++-- hadoop-mapreduce-project/hadoop-mapreduce-examples/pom.xml | 4 ++-- hadoop-mapreduce-project/pom.xml | 4 ++-- hadoop-maven-plugins/pom.xml | 2 +- hadoop-minicluster/pom.xml | 4 ++-- hadoop-project-dist/pom.xml | 4 ++-- hadoop-project/pom.xml | 4 ++-- hadoop-tools/hadoop-ant/pom.xml | 4 ++-- hadoop-tools/hadoop-archives/pom.xml | 4 ++-- hadoop-tools/hadoop-aws/pom.xml | 4 ++-- hadoop-tools/hadoop-azure/pom.xml| 2 +- hadoop-tools/hadoop-datajoin/pom.xml | 4 ++-- hadoop-tools/hadoop-distcp/pom.xml | 4 ++-- hadoop-tools/hadoop-extras/pom.xml | 4 ++-- hadoop-tools/hadoop-gridmix/pom.xml | 4 ++-- hadoop-tools/hadoop-openstack/pom.xml| 4 ++-- hadoop-tools/hadoop-pipes/pom.xml| 4 ++-- hadoop-tools/hadoop-rumen/pom.xml| 4 ++-- hadoop-tools/hadoop-sls/pom.xml | 4 ++-- hadoop-tools/hadoop-streaming/pom.xml| 4 ++-- hadoop-tools/hadoop-tools-dist/pom.xml | 4 ++-- hadoop-tools/pom.xml | 4 ++-- hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/pom.xml | 4 ++-- .../hadoop-yarn-applications-distributedshell/pom.xml| 4 ++-- .../hadoop-yarn-applications-unmanaged-am-launcher/pom.xml | 4 ++-- hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/pom.xml | 4 ++-- hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/pom.xml | 4 ++-- hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/pom.xml | 4 ++-- hadoop-yarn-project/hadoop-yarn/hadoop-yarn-registry/pom.xml | 4 ++-- .../hadoop-yarn-server-applicationhistoryservice/pom.xml | 4 ++-- .../hadoop-yarn-server/hadoop-yarn-server-common/pom.xml | 4 ++-- .../hadoop-yarn-server/hadoop-yarn-server-nodemanager/pom.xml| 4 ++-- .../hadoop-yarn-server-resourcemanager/pom.xml | 4 ++--
hadoop git commit: YARN-5191. Renamed the newly added “download=true” option for getting logs via NMWebServices and AHSWebServices to be a better "format" option. (Xuan Gong via vinodkv)
Repository: hadoop Updated Branches: refs/heads/branch-2 aef6e455b -> f6cd1bcf8 YARN-5191. Renamed the newly added âdownload=trueâ option for getting logs via NMWebServices and AHSWebServices to be a better "format" option. (Xuan Gong via vinodkv) (cherry picked from commit 9378d9428f127eff7acd6c13544016cdbf2d65fb) Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/f6cd1bcf Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/f6cd1bcf Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/f6cd1bcf Branch: refs/heads/branch-2 Commit: f6cd1bcf896ac91157dd02b0ae988c06eb70706b Parents: aef6e45 Author: Vinod Kumar Vavilapalli <vino...@apache.org> Authored: Thu Jun 9 12:30:58 2016 -0700 Committer: Vinod Kumar Vavilapalli <vino...@apache.org> Committed: Thu Jun 9 12:32:53 2016 -0700 -- .../hadoop/yarn/webapp/util/WebAppUtils.java| 18 ++ .../webapp/AHSWebServices.java | 36 .../nodemanager/webapp/NMWebServices.java | 33 +++--- .../nodemanager/webapp/TestNMWebServices.java | 15 +++- 4 files changed, 74 insertions(+), 28 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hadoop/blob/f6cd1bcf/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/webapp/util/WebAppUtils.java -- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/webapp/util/WebAppUtils.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/webapp/util/WebAppUtils.java index faf4a77..3aa773a 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/webapp/util/WebAppUtils.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/webapp/util/WebAppUtils.java @@ -24,6 +24,7 @@ import java.net.InetAddress; import java.net.InetSocketAddress; import java.net.UnknownHostException; import java.util.ArrayList; +import java.util.Arrays; import java.util.List; import org.apache.hadoop.classification.InterfaceAudience.Private; @@ -400,4 +401,21 @@ public class WebAppUtils { } return aid; } + + public static String getSupportedLogContentType(String format) { +if (format.equalsIgnoreCase("text")) { + return "text/plain"; +} else if (format.equalsIgnoreCase("octet-stream")) { + return "application/octet-stream"; +} +return null; + } + + public static String getDefaultLogContentType() { +return "text/plain"; + } + + public static List listSupportedLogContentType() { +return Arrays.asList("text", "octet-stream"); + } } http://git-wip-us.apache.org/repos/asf/hadoop/blob/f6cd1bcf/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/src/main/java/org/apache/hadoop/yarn/server/applicationhistoryservice/webapp/AHSWebServices.java -- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/src/main/java/org/apache/hadoop/yarn/server/applicationhistoryservice/webapp/AHSWebServices.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/src/main/java/org/apache/hadoop/yarn/server/applicationhistoryservice/webapp/AHSWebServices.java index 59dbd44..692b172 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/src/main/java/org/apache/hadoop/yarn/server/applicationhistoryservice/webapp/AHSWebServices.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/src/main/java/org/apache/hadoop/yarn/server/applicationhistoryservice/webapp/AHSWebServices.java @@ -66,6 +66,7 @@ import org.apache.hadoop.yarn.server.webapp.dao.ContainersInfo; import org.apache.hadoop.yarn.util.Times; import org.apache.hadoop.yarn.util.timeline.TimelineUtils; import org.apache.hadoop.yarn.webapp.BadRequestException; +import org.apache.hadoop.yarn.webapp.util.WebAppUtils; import com.google.common.base.Joiner; import com.google.inject.Inject; import com.google.inject.Singleton; @@ -212,7 +213,7 @@ public class AHSWebServices extends WebServices { @Context HttpServletResponse res, @PathParam("containerid") String containerIdStr, @PathParam("filename") String filename, - @QueryParam("download") String download, + @QueryParam("format") String format, @QueryParam(&
hadoop git commit: YARN-5191. Renamed the newly added “download=true” option for getting logs via NMWebServices and AHSWebServices to be a better "format" option. (Xuan Gong via vinodkv)
Repository: hadoop Updated Branches: refs/heads/trunk 656c460c0 -> 9378d9428 YARN-5191. Renamed the newly added âdownload=trueâ option for getting logs via NMWebServices and AHSWebServices to be a better "format" option. (Xuan Gong via vinodkv) Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/9378d942 Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/9378d942 Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/9378d942 Branch: refs/heads/trunk Commit: 9378d9428f127eff7acd6c13544016cdbf2d65fb Parents: 656c460 Author: Vinod Kumar Vavilapalli <vino...@apache.org> Authored: Thu Jun 9 12:30:58 2016 -0700 Committer: Vinod Kumar Vavilapalli <vino...@apache.org> Committed: Thu Jun 9 12:30:58 2016 -0700 -- .../hadoop/yarn/webapp/util/WebAppUtils.java| 18 ++ .../webapp/AHSWebServices.java | 36 .../nodemanager/webapp/NMWebServices.java | 33 +++--- .../nodemanager/webapp/TestNMWebServices.java | 15 +++- 4 files changed, 74 insertions(+), 28 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hadoop/blob/9378d942/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/webapp/util/WebAppUtils.java -- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/webapp/util/WebAppUtils.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/webapp/util/WebAppUtils.java index faf4a77..3aa773a 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/webapp/util/WebAppUtils.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/webapp/util/WebAppUtils.java @@ -24,6 +24,7 @@ import java.net.InetAddress; import java.net.InetSocketAddress; import java.net.UnknownHostException; import java.util.ArrayList; +import java.util.Arrays; import java.util.List; import org.apache.hadoop.classification.InterfaceAudience.Private; @@ -400,4 +401,21 @@ public class WebAppUtils { } return aid; } + + public static String getSupportedLogContentType(String format) { +if (format.equalsIgnoreCase("text")) { + return "text/plain"; +} else if (format.equalsIgnoreCase("octet-stream")) { + return "application/octet-stream"; +} +return null; + } + + public static String getDefaultLogContentType() { +return "text/plain"; + } + + public static List listSupportedLogContentType() { +return Arrays.asList("text", "octet-stream"); + } } http://git-wip-us.apache.org/repos/asf/hadoop/blob/9378d942/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/src/main/java/org/apache/hadoop/yarn/server/applicationhistoryservice/webapp/AHSWebServices.java -- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/src/main/java/org/apache/hadoop/yarn/server/applicationhistoryservice/webapp/AHSWebServices.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/src/main/java/org/apache/hadoop/yarn/server/applicationhistoryservice/webapp/AHSWebServices.java index 59dbd44..692b172 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/src/main/java/org/apache/hadoop/yarn/server/applicationhistoryservice/webapp/AHSWebServices.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/src/main/java/org/apache/hadoop/yarn/server/applicationhistoryservice/webapp/AHSWebServices.java @@ -66,6 +66,7 @@ import org.apache.hadoop.yarn.server.webapp.dao.ContainersInfo; import org.apache.hadoop.yarn.util.Times; import org.apache.hadoop.yarn.util.timeline.TimelineUtils; import org.apache.hadoop.yarn.webapp.BadRequestException; +import org.apache.hadoop.yarn.webapp.util.WebAppUtils; import com.google.common.base.Joiner; import com.google.inject.Inject; import com.google.inject.Singleton; @@ -212,7 +213,7 @@ public class AHSWebServices extends WebServices { @Context HttpServletResponse res, @PathParam("containerid") String containerIdStr, @PathParam("filename") String filename, - @QueryParam("download") String download, + @QueryParam("format") String format, @QueryParam("size") String size) { init(res); ContainerId containerId; @@ -223,9 +22
hadoop git commit: YARN-5080. Addendum fix to the original patch to fix YARN logs CLI. Contributed by Xuan Gong
Repository: hadoop Updated Branches: refs/heads/branch-2.8 fb31aaef6 -> 593d2c66c YARN-5080. Addendum fix to the original patch to fix YARN logs CLI. Contributed by Xuan Gong (cherry picked from commit 5a43583c0bbb9650ea6a9f48d9544ec3ec24b580) Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/593d2c66 Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/593d2c66 Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/593d2c66 Branch: refs/heads/branch-2.8 Commit: 593d2c66c44bc2fc74cfaa07750932f9cebc3151 Parents: fb31aae Author: Vinod Kumar VavilapalliAuthored: Wed Jun 8 09:49:55 2016 -0700 Committer: Vinod Kumar Vavilapalli Committed: Wed Jun 8 09:52:37 2016 -0700 -- .../src/main/java/org/apache/hadoop/yarn/client/cli/LogsCLI.java | 4 +++- 1 file changed, 3 insertions(+), 1 deletion(-) -- http://git-wip-us.apache.org/repos/asf/hadoop/blob/593d2c66/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/LogsCLI.java -- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/LogsCLI.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/LogsCLI.java index 89e6bff..5a6a8a5 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/LogsCLI.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/LogsCLI.java @@ -379,7 +379,9 @@ public class LogsCLI extends Configured implements Tool { Configuration conf, String appId) throws ClientHandlerException, UniformInterfaceException, JSONException { Client webServiceClient = Client.create(); -String webAppAddress = WebAppUtils.getRMWebAppURLWithScheme(conf); +String webAppAddress = WebAppUtils.getHttpSchemePrefix(conf) + +WebAppUtils.getWebAppBindURL(conf, YarnConfiguration.RM_BIND_HOST, +WebAppUtils.getRMWebAppURLWithoutScheme(conf)); WebResource webResource = webServiceClient.resource(webAppAddress); ClientResponse response = - To unsubscribe, e-mail: common-commits-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-commits-h...@hadoop.apache.org
hadoop git commit: YARN-5080. Addendum fix to the original patch to fix YARN logs CLI. Contributed by Xuan Gong
Repository: hadoop Updated Branches: refs/heads/branch-2 c29fbdd8a -> 67dbd4c46 YARN-5080. Addendum fix to the original patch to fix YARN logs CLI. Contributed by Xuan Gong (cherry picked from commit 5a43583c0bbb9650ea6a9f48d9544ec3ec24b580) Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/67dbd4c4 Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/67dbd4c4 Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/67dbd4c4 Branch: refs/heads/branch-2 Commit: 67dbd4c466c8e71181dc9c681ab9b95b1f56039b Parents: c29fbdd Author: Vinod Kumar VavilapalliAuthored: Wed Jun 8 09:49:55 2016 -0700 Committer: Vinod Kumar Vavilapalli Committed: Wed Jun 8 09:51:36 2016 -0700 -- .../src/main/java/org/apache/hadoop/yarn/client/cli/LogsCLI.java | 4 +++- 1 file changed, 3 insertions(+), 1 deletion(-) -- http://git-wip-us.apache.org/repos/asf/hadoop/blob/67dbd4c4/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/LogsCLI.java -- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/LogsCLI.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/LogsCLI.java index bbe636f..d62ee5e 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/LogsCLI.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/LogsCLI.java @@ -278,7 +278,9 @@ public class LogsCLI extends Configured implements Tool { Configuration conf, String appId) throws ClientHandlerException, UniformInterfaceException, JSONException { Client webServiceClient = Client.create(); -String webAppAddress = WebAppUtils.getRMWebAppURLWithScheme(conf); +String webAppAddress = WebAppUtils.getHttpSchemePrefix(conf) + +WebAppUtils.getWebAppBindURL(conf, YarnConfiguration.RM_BIND_HOST, +WebAppUtils.getRMWebAppURLWithoutScheme(conf)); WebResource webResource = webServiceClient.resource(webAppAddress); ClientResponse response = - To unsubscribe, e-mail: common-commits-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-commits-h...@hadoop.apache.org
hadoop git commit: YARN-5080. Addendum fix to the original patch to fix YARN logs CLI. Contributed by Xuan Gong
Repository: hadoop Updated Branches: refs/heads/trunk 3344ba70e -> 5a43583c0 YARN-5080. Addendum fix to the original patch to fix YARN logs CLI. Contributed by Xuan Gong Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/5a43583c Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/5a43583c Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/5a43583c Branch: refs/heads/trunk Commit: 5a43583c0bbb9650ea6a9f48d9544ec3ec24b580 Parents: 3344ba7 Author: Vinod Kumar VavilapalliAuthored: Wed Jun 8 09:49:55 2016 -0700 Committer: Vinod Kumar Vavilapalli Committed: Wed Jun 8 09:49:55 2016 -0700 -- .../src/main/java/org/apache/hadoop/yarn/client/cli/LogsCLI.java | 4 +++- 1 file changed, 3 insertions(+), 1 deletion(-) -- http://git-wip-us.apache.org/repos/asf/hadoop/blob/5a43583c/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/LogsCLI.java -- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/LogsCLI.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/LogsCLI.java index bbe636f..d62ee5e 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/LogsCLI.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/LogsCLI.java @@ -278,7 +278,9 @@ public class LogsCLI extends Configured implements Tool { Configuration conf, String appId) throws ClientHandlerException, UniformInterfaceException, JSONException { Client webServiceClient = Client.create(); -String webAppAddress = WebAppUtils.getRMWebAppURLWithScheme(conf); +String webAppAddress = WebAppUtils.getHttpSchemePrefix(conf) + +WebAppUtils.getWebAppBindURL(conf, YarnConfiguration.RM_BIND_HOST, +WebAppUtils.getRMWebAppURLWithoutScheme(conf)); WebResource webResource = webServiceClient.resource(webAppAddress); ClientResponse response = - To unsubscribe, e-mail: common-commits-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-commits-h...@hadoop.apache.org
1 2 3 4 5 6 7 8 9 10 >