[jira] [Commented] (HADOOP-8549) Allow other implementation's of java.util.Map in MapWritable
[ https://issues.apache.org/jira/browse/HADOOP-8549?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13414867#comment-13414867 ] madhukara phatak commented on HADOOP-8549: -- bq. Why can we hence not have a MapWritable(Class ? extends MapWritable, Writable mapType) that then instantiates the class for the field member instance, rather than a dev hacking like that? AFAIK in M/R writable's are instantiated through default constructor, which means the above constructor will not be called. So if I want to have a custom Map instance implementation it has to be in read/write methods. Please correct me if I am wrong :) Allow other implementation's of java.util.Map in MapWritable Key: HADOOP-8549 URL: https://issues.apache.org/jira/browse/HADOOP-8549 Project: Hadoop Common Issue Type: Improvement Components: io Affects Versions: 2.0.0-alpha Reporter: madhukara phatak Assignee: madhukara phatak Priority: Minor Attachments: HADOOP-8549.patch Current implementation of MapWritable uses HashMap as Map implementation. But in some of the use cases we need other implementations of Map like LinkedHashMap,SortedMap.This jira changes visibility of 'instance' in MapWritable from private to protected which allows us to inject custom Map implementation through sub classing MapWritable . -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HADOOP-8531) SequenceFile Writer can throw out a better error if a serializer or deserializer isn't available
[ https://issues.apache.org/jira/browse/HADOOP-8531?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Harsh J updated HADOOP-8531: Hadoop Flags: Reviewed SequenceFile Writer can throw out a better error if a serializer or deserializer isn't available Key: HADOOP-8531 URL: https://issues.apache.org/jira/browse/HADOOP-8531 Project: Hadoop Common Issue Type: Improvement Components: io Affects Versions: 2.0.0-alpha Reporter: Harsh J Assignee: madhukara phatak Priority: Trivial Labels: newbie Attachments: HADOOP-8531-1.patch, HADOOP-8531-2.patch, HADOOP-8531-2.patch, HADOOP-8531.patch Currently, if the provided Key/Value class lacks a proper serializer in the loaded config for the SequenceFile.Writer, we get an NPE as the null return goes unchecked. Hence we get: {code} java.lang.NullPointerException at org.apache.hadoop.io.SequenceFile$Writer.init(SequenceFile.java:1163) at org.apache.hadoop.io.SequenceFile$Writer.init(SequenceFile.java:1079) at org.apache.hadoop.io.SequenceFile$RecordCompressWriter.init(SequenceFile.java:1331) at org.apache.hadoop.io.SequenceFile.createWriter(SequenceFile.java:271) {code} We can provide a better message + exception in such cases. This is slightly related to MAPREDUCE-2584. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HADOOP-8531) SequenceFile Writer can throw out a better error if a serializer or deserializer isn't available
[ https://issues.apache.org/jira/browse/HADOOP-8531?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Harsh J updated HADOOP-8531: Component/s: io Target Version/s: 2.1.0-alpha Affects Version/s: 2.0.0-alpha Summary: SequenceFile Writer can throw out a better error if a serializer or deserializer isn't available (was: SequenceFile Writer can throw out a better error if a serializer isn't available) SequenceFile Writer can throw out a better error if a serializer or deserializer isn't available Key: HADOOP-8531 URL: https://issues.apache.org/jira/browse/HADOOP-8531 Project: Hadoop Common Issue Type: Improvement Components: io Affects Versions: 2.0.0-alpha Reporter: Harsh J Assignee: madhukara phatak Priority: Trivial Labels: newbie Attachments: HADOOP-8531-1.patch, HADOOP-8531-2.patch, HADOOP-8531-2.patch, HADOOP-8531.patch Currently, if the provided Key/Value class lacks a proper serializer in the loaded config for the SequenceFile.Writer, we get an NPE as the null return goes unchecked. Hence we get: {code} java.lang.NullPointerException at org.apache.hadoop.io.SequenceFile$Writer.init(SequenceFile.java:1163) at org.apache.hadoop.io.SequenceFile$Writer.init(SequenceFile.java:1079) at org.apache.hadoop.io.SequenceFile$RecordCompressWriter.init(SequenceFile.java:1331) at org.apache.hadoop.io.SequenceFile.createWriter(SequenceFile.java:271) {code} We can provide a better message + exception in such cases. This is slightly related to MAPREDUCE-2584. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HADOOP-8531) SequenceFile Writer can throw out a better error if a serializer or deserializer isn't available
[ https://issues.apache.org/jira/browse/HADOOP-8531?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13414931#comment-13414931 ] Hudson commented on HADOOP-8531: Integrated in Hadoop-Common-trunk-Commit #2477 (See [https://builds.apache.org/job/Hadoop-Common-trunk-Commit/2477/]) HADOOP-8531. SequenceFile Writer can throw out a better error if a serializer or deserializer isn't available. Contributed by Madhukara Phatak. (harsh) (Revision 1361933) Result = SUCCESS harsh : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1361933 Files : * /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/SequenceFile.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/io/TestSequenceFile.java SequenceFile Writer can throw out a better error if a serializer or deserializer isn't available Key: HADOOP-8531 URL: https://issues.apache.org/jira/browse/HADOOP-8531 Project: Hadoop Common Issue Type: Improvement Components: io Affects Versions: 2.0.0-alpha Reporter: Harsh J Assignee: madhukara phatak Priority: Trivial Labels: newbie Attachments: HADOOP-8531-1.patch, HADOOP-8531-2.patch, HADOOP-8531-2.patch, HADOOP-8531.patch Currently, if the provided Key/Value class lacks a proper serializer in the loaded config for the SequenceFile.Writer, we get an NPE as the null return goes unchecked. Hence we get: {code} java.lang.NullPointerException at org.apache.hadoop.io.SequenceFile$Writer.init(SequenceFile.java:1163) at org.apache.hadoop.io.SequenceFile$Writer.init(SequenceFile.java:1079) at org.apache.hadoop.io.SequenceFile$RecordCompressWriter.init(SequenceFile.java:1331) at org.apache.hadoop.io.SequenceFile.createWriter(SequenceFile.java:271) {code} We can provide a better message + exception in such cases. This is slightly related to MAPREDUCE-2584. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HADOOP-8531) SequenceFile Writer can throw out a better error if a serializer or deserializer isn't available
[ https://issues.apache.org/jira/browse/HADOOP-8531?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13414932#comment-13414932 ] Hudson commented on HADOOP-8531: Integrated in Hadoop-Hdfs-trunk-Commit #2542 (See [https://builds.apache.org/job/Hadoop-Hdfs-trunk-Commit/2542/]) HADOOP-8531. SequenceFile Writer can throw out a better error if a serializer or deserializer isn't available. Contributed by Madhukara Phatak. (harsh) (Revision 1361933) Result = SUCCESS harsh : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1361933 Files : * /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/SequenceFile.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/io/TestSequenceFile.java SequenceFile Writer can throw out a better error if a serializer or deserializer isn't available Key: HADOOP-8531 URL: https://issues.apache.org/jira/browse/HADOOP-8531 Project: Hadoop Common Issue Type: Improvement Components: io Affects Versions: 2.0.0-alpha Reporter: Harsh J Assignee: madhukara phatak Priority: Trivial Labels: newbie Attachments: HADOOP-8531-1.patch, HADOOP-8531-2.patch, HADOOP-8531-2.patch, HADOOP-8531.patch Currently, if the provided Key/Value class lacks a proper serializer in the loaded config for the SequenceFile.Writer, we get an NPE as the null return goes unchecked. Hence we get: {code} java.lang.NullPointerException at org.apache.hadoop.io.SequenceFile$Writer.init(SequenceFile.java:1163) at org.apache.hadoop.io.SequenceFile$Writer.init(SequenceFile.java:1079) at org.apache.hadoop.io.SequenceFile$RecordCompressWriter.init(SequenceFile.java:1331) at org.apache.hadoop.io.SequenceFile.createWriter(SequenceFile.java:271) {code} We can provide a better message + exception in such cases. This is slightly related to MAPREDUCE-2584. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HADOOP-8531) SequenceFile Writer can throw out a better error if a serializer or deserializer isn't available
[ https://issues.apache.org/jira/browse/HADOOP-8531?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13414934#comment-13414934 ] Hudson commented on HADOOP-8531: Integrated in Hadoop-Mapreduce-trunk-Commit #2497 (See [https://builds.apache.org/job/Hadoop-Mapreduce-trunk-Commit/2497/]) HADOOP-8531. SequenceFile Writer can throw out a better error if a serializer or deserializer isn't available. Contributed by Madhukara Phatak. (harsh) (Revision 1361933) Result = FAILURE harsh : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1361933 Files : * /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/SequenceFile.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/io/TestSequenceFile.java SequenceFile Writer can throw out a better error if a serializer or deserializer isn't available Key: HADOOP-8531 URL: https://issues.apache.org/jira/browse/HADOOP-8531 Project: Hadoop Common Issue Type: Improvement Components: io Affects Versions: 2.0.0-alpha Reporter: Harsh J Assignee: madhukara phatak Priority: Trivial Labels: newbie Attachments: HADOOP-8531-1.patch, HADOOP-8531-2.patch, HADOOP-8531-2.patch, HADOOP-8531.patch Currently, if the provided Key/Value class lacks a proper serializer in the loaded config for the SequenceFile.Writer, we get an NPE as the null return goes unchecked. Hence we get: {code} java.lang.NullPointerException at org.apache.hadoop.io.SequenceFile$Writer.init(SequenceFile.java:1163) at org.apache.hadoop.io.SequenceFile$Writer.init(SequenceFile.java:1079) at org.apache.hadoop.io.SequenceFile$RecordCompressWriter.init(SequenceFile.java:1331) at org.apache.hadoop.io.SequenceFile.createWriter(SequenceFile.java:271) {code} We can provide a better message + exception in such cases. This is slightly related to MAPREDUCE-2584. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HADOOP-8531) SequenceFile Writer can throw out a better error if a serializer or deserializer isn't available
[ https://issues.apache.org/jira/browse/HADOOP-8531?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Harsh J updated HADOOP-8531: Resolution: Fixed Fix Version/s: 2.1.0-alpha Status: Resolved (was: Patch Available) Committed to branch-2 and trunk. Thanks for the patch Madhukara! :) SequenceFile Writer can throw out a better error if a serializer or deserializer isn't available Key: HADOOP-8531 URL: https://issues.apache.org/jira/browse/HADOOP-8531 Project: Hadoop Common Issue Type: Improvement Components: io Affects Versions: 2.0.0-alpha Reporter: Harsh J Assignee: madhukara phatak Priority: Trivial Labels: newbie Fix For: 2.1.0-alpha Attachments: HADOOP-8531-1.patch, HADOOP-8531-2.patch, HADOOP-8531-2.patch, HADOOP-8531.patch Currently, if the provided Key/Value class lacks a proper serializer in the loaded config for the SequenceFile.Writer, we get an NPE as the null return goes unchecked. Hence we get: {code} java.lang.NullPointerException at org.apache.hadoop.io.SequenceFile$Writer.init(SequenceFile.java:1163) at org.apache.hadoop.io.SequenceFile$Writer.init(SequenceFile.java:1079) at org.apache.hadoop.io.SequenceFile$RecordCompressWriter.init(SequenceFile.java:1331) at org.apache.hadoop.io.SequenceFile.createWriter(SequenceFile.java:271) {code} We can provide a better message + exception in such cases. This is slightly related to MAPREDUCE-2584. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HADOOP-8531) SequenceFile Writer can throw out a better error if a serializer or deserializer isn't available
[ https://issues.apache.org/jira/browse/HADOOP-8531?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Harsh J updated HADOOP-8531: Target Version/s: (was: 2.1.0-alpha) SequenceFile Writer can throw out a better error if a serializer or deserializer isn't available Key: HADOOP-8531 URL: https://issues.apache.org/jira/browse/HADOOP-8531 Project: Hadoop Common Issue Type: Improvement Components: io Affects Versions: 2.0.0-alpha Reporter: Harsh J Assignee: madhukara phatak Priority: Trivial Labels: newbie Fix For: 2.1.0-alpha Attachments: HADOOP-8531-1.patch, HADOOP-8531-2.patch, HADOOP-8531-2.patch, HADOOP-8531.patch Currently, if the provided Key/Value class lacks a proper serializer in the loaded config for the SequenceFile.Writer, we get an NPE as the null return goes unchecked. Hence we get: {code} java.lang.NullPointerException at org.apache.hadoop.io.SequenceFile$Writer.init(SequenceFile.java:1163) at org.apache.hadoop.io.SequenceFile$Writer.init(SequenceFile.java:1079) at org.apache.hadoop.io.SequenceFile$RecordCompressWriter.init(SequenceFile.java:1331) at org.apache.hadoop.io.SequenceFile.createWriter(SequenceFile.java:271) {code} We can provide a better message + exception in such cases. This is slightly related to MAPREDUCE-2584. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HADOOP-7818) DiskChecker#checkDir should fail if the directory is not executable
[ https://issues.apache.org/jira/browse/HADOOP-7818?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13415024#comment-13415024 ] Hudson commented on HADOOP-7818: Integrated in Hadoop-Hdfs-trunk #1105 (See [https://builds.apache.org/job/Hadoop-Hdfs-trunk/1105/]) HADOOP-7818. DiskChecker#checkDir should fail if the directory is not executable. Contributed by Madhukara Phatak. (harsh) (Revision 1361717) Result = FAILURE harsh : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1361717 Files : * /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/DiskChecker.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/util/TestDiskChecker.java DiskChecker#checkDir should fail if the directory is not executable --- Key: HADOOP-7818 URL: https://issues.apache.org/jira/browse/HADOOP-7818 Project: Hadoop Common Issue Type: Bug Components: util Affects Versions: 2.0.0-alpha Reporter: Eli Collins Assignee: madhukara phatak Priority: Minor Fix For: 2.1.0-alpha Attachments: HADOOP-7818-1.patch, HADOOP-7818-2.patch, HADOOP-7818-3.patch, HADOOP-7818-4.patch, HADOOP-7818.patch DiskChecker#checkDir fails if a directory can't be created, read, or written but does not fail if the directory exists and is not executable. This causes subsequent code to think the directory is OK but later fail due to an inability to access the directory (eg see MAPREDUCE-2921). I propose checkDir fails if the directory is not executable. Looking at the uses, this should be fine, I think it was ignored because checkDir is often used to create directories and it creates executable directories. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HADOOP-8362) Improve exception message when Configuration.set() is called with a null key or value
[ https://issues.apache.org/jira/browse/HADOOP-8362?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13415027#comment-13415027 ] Hudson commented on HADOOP-8362: Integrated in Hadoop-Hdfs-trunk #1105 (See [https://builds.apache.org/job/Hadoop-Hdfs-trunk/1105/]) HADOOP-8362. Improve exception message when Configuration.set() is called with a null key or value. Contributed by Madhukara Phatak and Suresh Srinivas (harsh) (Revision 1361712) Result = FAILURE harsh : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1361712 Files : * /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/conf/Configuration.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/conf/TestConfiguration.java Improve exception message when Configuration.set() is called with a null key or value - Key: HADOOP-8362 URL: https://issues.apache.org/jira/browse/HADOOP-8362 Project: Hadoop Common Issue Type: Improvement Components: conf Affects Versions: 2.0.0-alpha Reporter: Todd Lipcon Assignee: madhukara phatak Priority: Trivial Labels: newbie Fix For: 2.1.0-alpha Attachments: HADOOP-8362-1.patch, HADOOP-8362-2.patch, HADOOP-8362-3.patch, HADOOP-8362-4.patch, HADOOP-8362-5.patch, HADOOP-8362-6.patch, HADOOP-8362-7.patch, HADOOP-8362-8.patch, HADOOP-8362.10.patch, HADOOP-8362.9.patch, HADOOP-8362.patch Currently, calling Configuration.set(...) with a null value results in a NullPointerException within Properties.setProperty. We should check for null key/value and throw a better exception. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HADOOP-7818) DiskChecker#checkDir should fail if the directory is not executable
[ https://issues.apache.org/jira/browse/HADOOP-7818?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13415095#comment-13415095 ] Hudson commented on HADOOP-7818: Integrated in Hadoop-Mapreduce-trunk #1138 (See [https://builds.apache.org/job/Hadoop-Mapreduce-trunk/1138/]) HADOOP-7818. DiskChecker#checkDir should fail if the directory is not executable. Contributed by Madhukara Phatak. (harsh) (Revision 1361717) Result = FAILURE harsh : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1361717 Files : * /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/DiskChecker.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/util/TestDiskChecker.java DiskChecker#checkDir should fail if the directory is not executable --- Key: HADOOP-7818 URL: https://issues.apache.org/jira/browse/HADOOP-7818 Project: Hadoop Common Issue Type: Bug Components: util Affects Versions: 2.0.0-alpha Reporter: Eli Collins Assignee: madhukara phatak Priority: Minor Fix For: 2.1.0-alpha Attachments: HADOOP-7818-1.patch, HADOOP-7818-2.patch, HADOOP-7818-3.patch, HADOOP-7818-4.patch, HADOOP-7818.patch DiskChecker#checkDir fails if a directory can't be created, read, or written but does not fail if the directory exists and is not executable. This causes subsequent code to think the directory is OK but later fail due to an inability to access the directory (eg see MAPREDUCE-2921). I propose checkDir fails if the directory is not executable. Looking at the uses, this should be fine, I think it was ignored because checkDir is often used to create directories and it creates executable directories. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HADOOP-8531) SequenceFile Writer can throw out a better error if a serializer or deserializer isn't available
[ https://issues.apache.org/jira/browse/HADOOP-8531?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13415096#comment-13415096 ] Hudson commented on HADOOP-8531: Integrated in Hadoop-Mapreduce-trunk #1138 (See [https://builds.apache.org/job/Hadoop-Mapreduce-trunk/1138/]) HADOOP-8531. SequenceFile Writer can throw out a better error if a serializer or deserializer isn't available. Contributed by Madhukara Phatak. (harsh) (Revision 1361933) Result = FAILURE harsh : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1361933 Files : * /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/SequenceFile.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/io/TestSequenceFile.java SequenceFile Writer can throw out a better error if a serializer or deserializer isn't available Key: HADOOP-8531 URL: https://issues.apache.org/jira/browse/HADOOP-8531 Project: Hadoop Common Issue Type: Improvement Components: io Affects Versions: 2.0.0-alpha Reporter: Harsh J Assignee: madhukara phatak Priority: Trivial Labels: newbie Fix For: 2.1.0-alpha Attachments: HADOOP-8531-1.patch, HADOOP-8531-2.patch, HADOOP-8531-2.patch, HADOOP-8531.patch Currently, if the provided Key/Value class lacks a proper serializer in the loaded config for the SequenceFile.Writer, we get an NPE as the null return goes unchecked. Hence we get: {code} java.lang.NullPointerException at org.apache.hadoop.io.SequenceFile$Writer.init(SequenceFile.java:1163) at org.apache.hadoop.io.SequenceFile$Writer.init(SequenceFile.java:1079) at org.apache.hadoop.io.SequenceFile$RecordCompressWriter.init(SequenceFile.java:1331) at org.apache.hadoop.io.SequenceFile.createWriter(SequenceFile.java:271) {code} We can provide a better message + exception in such cases. This is slightly related to MAPREDUCE-2584. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HADOOP-8362) Improve exception message when Configuration.set() is called with a null key or value
[ https://issues.apache.org/jira/browse/HADOOP-8362?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13415098#comment-13415098 ] Hudson commented on HADOOP-8362: Integrated in Hadoop-Mapreduce-trunk #1138 (See [https://builds.apache.org/job/Hadoop-Mapreduce-trunk/1138/]) HADOOP-8362. Improve exception message when Configuration.set() is called with a null key or value. Contributed by Madhukara Phatak and Suresh Srinivas (harsh) (Revision 1361712) Result = FAILURE harsh : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1361712 Files : * /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/conf/Configuration.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/conf/TestConfiguration.java Improve exception message when Configuration.set() is called with a null key or value - Key: HADOOP-8362 URL: https://issues.apache.org/jira/browse/HADOOP-8362 Project: Hadoop Common Issue Type: Improvement Components: conf Affects Versions: 2.0.0-alpha Reporter: Todd Lipcon Assignee: madhukara phatak Priority: Trivial Labels: newbie Fix For: 2.1.0-alpha Attachments: HADOOP-8362-1.patch, HADOOP-8362-2.patch, HADOOP-8362-3.patch, HADOOP-8362-4.patch, HADOOP-8362-5.patch, HADOOP-8362-6.patch, HADOOP-8362-7.patch, HADOOP-8362-8.patch, HADOOP-8362.10.patch, HADOOP-8362.9.patch, HADOOP-8362.patch Currently, calling Configuration.set(...) with a null value results in a NullPointerException within Properties.setProperty. We should check for null key/value and throw a better exception. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HADOOP-8594) Upgrade to findbugs 2
[ https://issues.apache.org/jira/browse/HADOOP-8594?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13415125#comment-13415125 ] Robert Joseph Evans commented on HADOOP-8594: - I agree with Eli on this. Switching to a new findbugs is a great thing, but I don't want to break the pre-commit builds in the process. If we do that new findbugs issues will get past it simply because 54 looks a lot like 53 and it is simpler to say I didn't break it then to look at all 54 of the errors to see what the new ones might be. Upgrade to findbugs 2 - Key: HADOOP-8594 URL: https://issues.apache.org/jira/browse/HADOOP-8594 Project: Hadoop Common Issue Type: Bug Components: build Affects Versions: 2.0.0-alpha Reporter: Eli Collins Attachments: HADOOP-8594.patch, findbugs-2-html-reports.tar.gz, findbugs.html Harsh recently ran findbugs 2 (instead of 1.3.9 which is what jenkins runs) and it showed thousands of warnings (they've made a lot of progress in findbugs releases). We should upgrade to findbugs 2 and fix these. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Assigned] (HADOOP-7823) port HADOOP-4012 to branch-1 (splitting support for bzip2)
[ https://issues.apache.org/jira/browse/HADOOP-7823?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Andrew Purtell reassigned HADOOP-7823: -- Assignee: (was: Andrew Purtell) port HADOOP-4012 to branch-1 (splitting support for bzip2) -- Key: HADOOP-7823 URL: https://issues.apache.org/jira/browse/HADOOP-7823 Project: Hadoop Common Issue Type: New Feature Affects Versions: 0.20.205.0 Reporter: Tim Broberg Attachments: HADOOP-7823-branch-1-v2.patch, HADOOP-7823-branch-1-v3.patch, HADOOP-7823-branch-1-v3.patch, HADOOP-7823-branch-1-v4.patch, HADOOP-7823-branch-1.patch Please see HADOOP-4012 - Providing splitting support for bzip2 compressed files. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HADOOP-8552) Conflict: Same security.log.file for multiple users.
[ https://issues.apache.org/jira/browse/HADOOP-8552?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13415387#comment-13415387 ] Alejandro Abdelnur commented on HADOOP-8552: Devaraj, are you OK with this patch after Karthik's clarifications? Conflict: Same security.log.file for multiple users. - Key: HADOOP-8552 URL: https://issues.apache.org/jira/browse/HADOOP-8552 Project: Hadoop Common Issue Type: Bug Components: conf, security Affects Versions: 1.0.3, 2.0.0-alpha Reporter: Karthik Kambatla Assignee: Karthik Kambatla Attachments: HADOOP-8552_branch1.patch, HADOOP-8552_branch2.patch In log4j.properties, hadoop.security.log.file is set to SecurityAuth.audit. In the presence of multiple users, this can lead to a potential conflict. Adding username to the log file would avoid this scenario. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HADOOP-8552) Conflict: Same security.log.file for multiple users.
[ https://issues.apache.org/jira/browse/HADOOP-8552?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13415412#comment-13415412 ] Devaraj Das commented on HADOOP-8552: - Yes. Conflict: Same security.log.file for multiple users. - Key: HADOOP-8552 URL: https://issues.apache.org/jira/browse/HADOOP-8552 Project: Hadoop Common Issue Type: Bug Components: conf, security Affects Versions: 1.0.3, 2.0.0-alpha Reporter: Karthik Kambatla Assignee: Karthik Kambatla Attachments: HADOOP-8552_branch1.patch, HADOOP-8552_branch2.patch In log4j.properties, hadoop.security.log.file is set to SecurityAuth.audit. In the presence of multiple users, this can lead to a potential conflict. Adding username to the log file would avoid this scenario. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HADOOP-8552) Conflict: Same security.log.file for multiple users.
[ https://issues.apache.org/jira/browse/HADOOP-8552?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Alejandro Abdelnur updated HADOOP-8552: --- Resolution: Fixed Fix Version/s: 2.0.1-alpha 1.1.0 Hadoop Flags: Incompatible change,Reviewed Status: Resolved (was: Patch Available) Thanks Karthik. Committed to trunk, branch-1 and branch-2. Conflict: Same security.log.file for multiple users. - Key: HADOOP-8552 URL: https://issues.apache.org/jira/browse/HADOOP-8552 Project: Hadoop Common Issue Type: Bug Components: conf, security Affects Versions: 1.0.3, 2.0.0-alpha Reporter: Karthik Kambatla Assignee: Karthik Kambatla Fix For: 1.1.0, 2.0.1-alpha Attachments: HADOOP-8552_branch1.patch, HADOOP-8552_branch2.patch In log4j.properties, hadoop.security.log.file is set to SecurityAuth.audit. In the presence of multiple users, this can lead to a potential conflict. Adding username to the log file would avoid this scenario. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HADOOP-8552) Conflict: Same security.log.file for multiple users.
[ https://issues.apache.org/jira/browse/HADOOP-8552?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13415423#comment-13415423 ] Hudson commented on HADOOP-8552: Integrated in Hadoop-Hdfs-trunk-Commit #2545 (See [https://builds.apache.org/job/Hadoop-Hdfs-trunk-Commit/2545/]) HADOOP-8552. Conflict: Same security.log.file for multiple users. (kkambatl via tucu) (Revision 1362151) Result = SUCCESS tucu : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1362151 Files : * /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/conf/log4j.properties * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/packages/templates/conf/log4j.properties Conflict: Same security.log.file for multiple users. - Key: HADOOP-8552 URL: https://issues.apache.org/jira/browse/HADOOP-8552 Project: Hadoop Common Issue Type: Bug Components: conf, security Affects Versions: 1.0.3, 2.0.0-alpha Reporter: Karthik Kambatla Assignee: Karthik Kambatla Fix For: 1.1.0, 2.0.1-alpha Attachments: HADOOP-8552_branch1.patch, HADOOP-8552_branch2.patch In log4j.properties, hadoop.security.log.file is set to SecurityAuth.audit. In the presence of multiple users, this can lead to a potential conflict. Adding username to the log file would avoid this scenario. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HADOOP-8552) Conflict: Same security.log.file for multiple users.
[ https://issues.apache.org/jira/browse/HADOOP-8552?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13415426#comment-13415426 ] Hudson commented on HADOOP-8552: Integrated in Hadoop-Common-trunk-Commit #2480 (See [https://builds.apache.org/job/Hadoop-Common-trunk-Commit/2480/]) HADOOP-8552. Conflict: Same security.log.file for multiple users. (kkambatl via tucu) (Revision 1362151) Result = SUCCESS tucu : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1362151 Files : * /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/conf/log4j.properties * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/packages/templates/conf/log4j.properties Conflict: Same security.log.file for multiple users. - Key: HADOOP-8552 URL: https://issues.apache.org/jira/browse/HADOOP-8552 Project: Hadoop Common Issue Type: Bug Components: conf, security Affects Versions: 1.0.3, 2.0.0-alpha Reporter: Karthik Kambatla Assignee: Karthik Kambatla Fix For: 1.1.0, 2.0.1-alpha Attachments: HADOOP-8552_branch1.patch, HADOOP-8552_branch2.patch In log4j.properties, hadoop.security.log.file is set to SecurityAuth.audit. In the presence of multiple users, this can lead to a potential conflict. Adding username to the log file would avoid this scenario. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HADOOP-8552) Conflict: Same security.log.file for multiple users.
[ https://issues.apache.org/jira/browse/HADOOP-8552?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13415439#comment-13415439 ] Hudson commented on HADOOP-8552: Integrated in Hadoop-Mapreduce-trunk-Commit #2500 (See [https://builds.apache.org/job/Hadoop-Mapreduce-trunk-Commit/2500/]) HADOOP-8552. Conflict: Same security.log.file for multiple users. (kkambatl via tucu) (Revision 1362151) Result = FAILURE tucu : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1362151 Files : * /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/conf/log4j.properties * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/packages/templates/conf/log4j.properties Conflict: Same security.log.file for multiple users. - Key: HADOOP-8552 URL: https://issues.apache.org/jira/browse/HADOOP-8552 Project: Hadoop Common Issue Type: Bug Components: conf, security Affects Versions: 1.0.3, 2.0.0-alpha Reporter: Karthik Kambatla Assignee: Karthik Kambatla Fix For: 1.1.0, 2.0.1-alpha Attachments: HADOOP-8552_branch1.patch, HADOOP-8552_branch2.patch In log4j.properties, hadoop.security.log.file is set to SecurityAuth.audit. In the presence of multiple users, this can lead to a potential conflict. Adding username to the log file would avoid this scenario. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HADOOP-8589) ViewFs tests fail when tests dir is under Jenkins home dir
[ https://issues.apache.org/jira/browse/HADOOP-8589?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13415460#comment-13415460 ] Andrey Klochkov commented on HADOOP-8589: - The problem with the current code is that those 2 links can point to nested directories. In case of tests being run on a Jenkins master in it's default configuration one of the links the code will try to create will point to /var and another to /var/lib, and the creation of the second link will fail with Path /var already exists as dir; cannot create link. I'm proposing to use just the first component of the path for both links, i.e. cut off everything after /var in my example. It's already being done for the test link, and my patch is adding the same logic for the home link. ViewFs tests fail when tests dir is under Jenkins home dir -- Key: HADOOP-8589 URL: https://issues.apache.org/jira/browse/HADOOP-8589 Project: Hadoop Common Issue Type: Bug Components: fs, test Affects Versions: 0.23.1, 2.0.0-alpha Reporter: Andrey Klochkov Assignee: Andrey Klochkov Attachments: HADOOP-8859.patch TestFSMainOperationsLocalFileSystem fails in case when the test root directory is under the user's home directory, and the user's home dir is deeper than 2 levels from /. This happens with the default 1-node installation of Jenkins. This is the failure log: {code} org.apache.hadoop.fs.FileAlreadyExistsException: Path /var already exists as dir; cannot create link here at org.apache.hadoop.fs.viewfs.InodeTree.createLink(InodeTree.java:244) at org.apache.hadoop.fs.viewfs.InodeTree.init(InodeTree.java:334) at org.apache.hadoop.fs.viewfs.ViewFileSystem$1.init(ViewFileSystem.java:167) at org.apache.hadoop.fs.viewfs.ViewFileSystem.initialize(ViewFileSystem.java:167) at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2094) at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:79) at org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2128) at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2110) at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:290) at org.apache.hadoop.fs.viewfs.ViewFileSystemTestSetup.setupForViewFileSystem(ViewFileSystemTestSetup.java:76) at org.apache.hadoop.fs.viewfs.TestFSMainOperationsLocalFileSystem.setUp(TestFSMainOperationsLocalFileSystem.java:40) ... Standard Output 2012-07-11 22:07:20,239 INFO mortbay.log (Slf4jLog.java:info(67)) - Home dir base /var/lib {code} The reason for the failure is that the code tries to mount links for both /var and /var/lib, and it fails for the 2nd one as the /var is mounted already. The fix was provided in HADOOP-8036 but later it was reverted in HADOOP-8129. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HADOOP-8453) Add unit tests for winutils
[ https://issues.apache.org/jira/browse/HADOOP-8453?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13415492#comment-13415492 ] Bikas Saha commented on HADOOP-8453: Sure. Why dont you open a JIRA for that. We need to make sure these tests are executed on Windows. One option would be to check for that at startup of the test and return success for other OS's. Add unit tests for winutils --- Key: HADOOP-8453 URL: https://issues.apache.org/jira/browse/HADOOP-8453 Project: Hadoop Common Issue Type: Task Components: test Affects Versions: 1.1.0, 0.24.0 Reporter: Chuan Liu Assignee: Chuan Liu Priority: Minor Attachments: HADOOP-8453-branch-1-win-2.patch, HADOOP-8453-branch-1-win.patch In [Hadoop-8235|https://issues.apache.org/jira/browse/HADOOP-8235], we created a Windows console program, named ‘winutils’, to emulate some Linux command line utilities used by Hadoop. However no tests are provided in the original patch. As this code is quite complicated, and the complexity may even grow up in the future. We think unit tests are necessary to ensure code quality, as well as smooth future development. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HADOOP-8589) ViewFs tests fail when tests dir is under Jenkins home dir
[ https://issues.apache.org/jira/browse/HADOOP-8589?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13415525#comment-13415525 ] Ahmed Radwan commented on HADOOP-8589: -- bq. I'm proposing to use just the first component of the path for both links, i.e. cut off everything after /var in my example. It's already being done for the test link, and my patch is adding the same logic for the home link. * Can you clarify that in the ticket description since it is misleading? * Would it be better to check if (homeDirFirstComponent == testDirFirstComponent), and just add a single link, instead of adding the same link twice? ViewFs tests fail when tests dir is under Jenkins home dir -- Key: HADOOP-8589 URL: https://issues.apache.org/jira/browse/HADOOP-8589 Project: Hadoop Common Issue Type: Bug Components: fs, test Affects Versions: 0.23.1, 2.0.0-alpha Reporter: Andrey Klochkov Assignee: Andrey Klochkov Attachments: HADOOP-8859.patch TestFSMainOperationsLocalFileSystem fails in case when the test root directory is under the user's home directory, and the user's home dir is deeper than 2 levels from /. This happens with the default 1-node installation of Jenkins. This is the failure log: {code} org.apache.hadoop.fs.FileAlreadyExistsException: Path /var already exists as dir; cannot create link here at org.apache.hadoop.fs.viewfs.InodeTree.createLink(InodeTree.java:244) at org.apache.hadoop.fs.viewfs.InodeTree.init(InodeTree.java:334) at org.apache.hadoop.fs.viewfs.ViewFileSystem$1.init(ViewFileSystem.java:167) at org.apache.hadoop.fs.viewfs.ViewFileSystem.initialize(ViewFileSystem.java:167) at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2094) at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:79) at org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2128) at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2110) at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:290) at org.apache.hadoop.fs.viewfs.ViewFileSystemTestSetup.setupForViewFileSystem(ViewFileSystemTestSetup.java:76) at org.apache.hadoop.fs.viewfs.TestFSMainOperationsLocalFileSystem.setUp(TestFSMainOperationsLocalFileSystem.java:40) ... Standard Output 2012-07-11 22:07:20,239 INFO mortbay.log (Slf4jLog.java:info(67)) - Home dir base /var/lib {code} The reason for the failure is that the code tries to mount links for both /var and /var/lib, and it fails for the 2nd one as the /var is mounted already. The fix was provided in HADOOP-8036 but later it was reverted in HADOOP-8129. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HADOOP-8589) ViewFs tests fail when tests dir is under Jenkins home dir
[ https://issues.apache.org/jira/browse/HADOOP-8589?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13415545#comment-13415545 ] Daryn Sharp commented on HADOOP-8589: - I'd like to see viewfs tests only accessing directories mounted under the build.test.dir. There's been a number of recurring bugs centered around the mounting of the user's home dir. Are you aware the trash test tries to blow away your home directory to verify it won't delete an ancestor of the trash dir? Shining illustration of why everything should be under build.test.dir. ViewFs tests fail when tests dir is under Jenkins home dir -- Key: HADOOP-8589 URL: https://issues.apache.org/jira/browse/HADOOP-8589 Project: Hadoop Common Issue Type: Bug Components: fs, test Affects Versions: 0.23.1, 2.0.0-alpha Reporter: Andrey Klochkov Assignee: Andrey Klochkov Attachments: HADOOP-8859.patch TestFSMainOperationsLocalFileSystem fails in case when the test root directory is under the user's home directory, and the user's home dir is deeper than 2 levels from /. This happens with the default 1-node installation of Jenkins. This is the failure log: {code} org.apache.hadoop.fs.FileAlreadyExistsException: Path /var already exists as dir; cannot create link here at org.apache.hadoop.fs.viewfs.InodeTree.createLink(InodeTree.java:244) at org.apache.hadoop.fs.viewfs.InodeTree.init(InodeTree.java:334) at org.apache.hadoop.fs.viewfs.ViewFileSystem$1.init(ViewFileSystem.java:167) at org.apache.hadoop.fs.viewfs.ViewFileSystem.initialize(ViewFileSystem.java:167) at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2094) at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:79) at org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2128) at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2110) at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:290) at org.apache.hadoop.fs.viewfs.ViewFileSystemTestSetup.setupForViewFileSystem(ViewFileSystemTestSetup.java:76) at org.apache.hadoop.fs.viewfs.TestFSMainOperationsLocalFileSystem.setUp(TestFSMainOperationsLocalFileSystem.java:40) ... Standard Output 2012-07-11 22:07:20,239 INFO mortbay.log (Slf4jLog.java:info(67)) - Home dir base /var/lib {code} The reason for the failure is that the code tries to mount links for both /var and /var/lib, and it fails for the 2nd one as the /var is mounted already. The fix was provided in HADOOP-8036 but later it was reverted in HADOOP-8129. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HADOOP-8589) ViewFs tests fail when tests and home dirs are nested
[ https://issues.apache.org/jira/browse/HADOOP-8589?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Andrey Klochkov updated HADOOP-8589: Summary: ViewFs tests fail when tests and home dirs are nested (was: ViewFs tests fail when tests dir is under Jenkins home dir) ViewFs tests fail when tests and home dirs are nested - Key: HADOOP-8589 URL: https://issues.apache.org/jira/browse/HADOOP-8589 Project: Hadoop Common Issue Type: Bug Components: fs, test Affects Versions: 0.23.1, 2.0.0-alpha Reporter: Andrey Klochkov Assignee: Andrey Klochkov Attachments: HADOOP-8859.patch TestFSMainOperationsLocalFileSystem fails in case when the test root directory is under the user's home directory, and the user's home dir is deeper than 2 levels from /. This happens with the default 1-node installation of Jenkins. This is the failure log: {code} org.apache.hadoop.fs.FileAlreadyExistsException: Path /var already exists as dir; cannot create link here at org.apache.hadoop.fs.viewfs.InodeTree.createLink(InodeTree.java:244) at org.apache.hadoop.fs.viewfs.InodeTree.init(InodeTree.java:334) at org.apache.hadoop.fs.viewfs.ViewFileSystem$1.init(ViewFileSystem.java:167) at org.apache.hadoop.fs.viewfs.ViewFileSystem.initialize(ViewFileSystem.java:167) at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2094) at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:79) at org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2128) at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2110) at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:290) at org.apache.hadoop.fs.viewfs.ViewFileSystemTestSetup.setupForViewFileSystem(ViewFileSystemTestSetup.java:76) at org.apache.hadoop.fs.viewfs.TestFSMainOperationsLocalFileSystem.setUp(TestFSMainOperationsLocalFileSystem.java:40) ... Standard Output 2012-07-11 22:07:20,239 INFO mortbay.log (Slf4jLog.java:info(67)) - Home dir base /var/lib {code} The reason for the failure is that the code tries to mount links for both /var and /var/lib, and it fails for the 2nd one as the /var is mounted already. The fix was provided in HADOOP-8036 but later it was reverted in HADOOP-8129. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Created] (HADOOP-8599) Non empty response from FileSystem.getFileBlockLocations when asking for data beyond the end of file
Andrey Klochkov created HADOOP-8599: --- Summary: Non empty response from FileSystem.getFileBlockLocations when asking for data beyond the end of file Key: HADOOP-8599 URL: https://issues.apache.org/jira/browse/HADOOP-8599 Project: Hadoop Common Issue Type: Bug Components: fs Affects Versions: 2.0.0-alpha, 0.23.1, 1.0.3 Reporter: Andrey Klochkov When FileSystem.getFileBlockLocations(file,start,len) is called with start argument equal to the file size, the response is not empty. There is a test TestGetFileBlockLocations.testGetFileBlockLocations2 which uses randomly generated start and len arguments when calling FileSystem.getFileBlockLocations and the test fails randomly (when the generated start value equals to the file size). -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HADOOP-8599) Non empty response from FileSystem.getFileBlockLocations when asking for data beyond the end of file
[ https://issues.apache.org/jira/browse/HADOOP-8599?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13415576#comment-13415576 ] Andrey Klochkov commented on HADOOP-8599: - The build https://builds.apache.org/job/Hadoop-Common-0.23-Build/312/ failed due to this bug, this is the failure log: {code} org.apache.hadoop.fs.TestGetFileBlockLocations.testGetFileBlockLocations2 Failing for the past 1 build (Since #312 ) Took 0.29 sec. Stacktrace junit.framework.AssertionFailedError: null at junit.framework.Assert.fail(Assert.java:47) at junit.framework.Assert.assertTrue(Assert.java:20) at junit.framework.Assert.assertTrue(Assert.java:27) at org.apache.hadoop.fs.TestGetFileBlockLocations.oneTest(TestGetFileBlockLocations.java:90) at org.apache.hadoop.fs.TestGetFileBlockLocations.__CLR3_0_2h1r7rc12wg(TestGetFileBlockLocations.java:136) at org.apache.hadoop.fs.TestGetFileBlockLocations.testGetFileBlockLocations2(TestGetFileBlockLocations.java:131) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) at java.lang.reflect.Method.invoke(Method.java:597) at junit.framework.TestCase.runTest(TestCase.java:168) at junit.framework.TestCase.runBare(TestCase.java:134) at junit.framework.TestResult$1.protect(TestResult.java:110) at junit.framework.TestResult.runProtected(TestResult.java:128) at junit.framework.TestResult.run(TestResult.java:113) at junit.framework.TestCase.run(TestCase.java:124) at junit.framework.TestSuite.runTest(TestSuite.java:243) at junit.framework.TestSuite.run(TestSuite.java:238) at org.junit.internal.runners.JUnit38ClassRunner.run(JUnit38ClassRunner.java:83) at org.apache.maven.surefire.junit4.JUnit4TestSet.execute(JUnit4TestSet.java:53) at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:123) at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:104) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) at java.lang.reflect.Method.invoke(Method.java:597) at org.apache.maven.surefire.util.ReflectionUtils.invokeMethodWithArray(ReflectionUtils.java:164) at org.apache.maven.surefire.booter.ProviderFactory$ProviderProxy.invoke(ProviderFactory.java:110) at org.apache.maven.surefire.booter.SurefireStarter.invokeProvider(SurefireStarter.java:175) at org.apache.maven.surefire.booter.SurefireStarter.runSuitesInProcessWhenForked(SurefireStarter.java:81) at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:68) Standard Output 2012-07-15 09:13:27,734 INFO util.NativeCodeLoader (NativeCodeLoader.java:clinit(50)) - Loaded the native-hadoop library {code} Non empty response from FileSystem.getFileBlockLocations when asking for data beyond the end of file - Key: HADOOP-8599 URL: https://issues.apache.org/jira/browse/HADOOP-8599 Project: Hadoop Common Issue Type: Bug Components: fs Affects Versions: 1.0.3, 0.23.1, 2.0.0-alpha Reporter: Andrey Klochkov When FileSystem.getFileBlockLocations(file,start,len) is called with start argument equal to the file size, the response is not empty. There is a test TestGetFileBlockLocations.testGetFileBlockLocations2 which uses randomly generated start and len arguments when calling FileSystem.getFileBlockLocations and the test fails randomly (when the generated start value equals to the file size). -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HADOOP-8598) Server-side Trash
[ https://issues.apache.org/jira/browse/HADOOP-8598?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13415580#comment-13415580 ] Aaron T. Myers commented on HADOOP-8598: I think a quick and easy option for implementing server-side trash would be to just not require any configuration client-side. This could be done by adding an isTrashEnabled field to FsServerDefaults and changing the TrashPolicyDefault#isEnabled method to call Filesystem#getServerDefaults and return the value of isTrashEnabled. This won't result in a ton of extra RPCs to the NN since the HDFS implementation of getServerDefaults is cached in the DFSClient and rate-limited to one call per hour. Though it obviously wouldn't cover the issue of APIs besides the FsShell, I think this would cover a very large portion of the user problems we currently see with trash. Server-side Trash - Key: HADOOP-8598 URL: https://issues.apache.org/jira/browse/HADOOP-8598 Project: Hadoop Common Issue Type: New Feature Affects Versions: 2.0.0-alpha Reporter: Eli Collins Assignee: Eli Collins Priority: Critical There are a number of problems with Trash that continue to result in permanent data loss for users. The primary reasons trash is not used: - Trash is configured client-side and not enabled by default. - Trash is shell-only. FileSystem, WebHDFS, HttpFs, etc never use trash. - If trash fails, for example, because we can't create the trash directory or the move itself fails, trash is bypassed and the data is deleted. Trash was designed as a feature to help end users via the shell, however in my experience the primary use of trash is to help administrators implement data retention policies (this was also the motivation for HADOOP-7460). One could argue that (periodic read-only) snapshots are a better solution to this problem, however snapshots are not slated for Hadoop 2.x and trash is complimentary to snapshots (and backup) - eg you may create and delete data within your snapshot or backup window - so it makes sense to revisit trash's design. I think it's worth bringing trash's functionality in line with what users need. I propose we enable trash on a per-filesystem basis and implement it server-side. Ie trash becomes an HDFS feature enabled by administrators. Because the trash emptier lives in HDFS and users already have a per-filesystem trash directory we're mostly there already. The design preference from HADOOP-2514 was for trash to be implemented in user code however (a) in light of these problems, (b) we have a lot more user-facing APIs than the shell and (c) clients increasingly span file systems (via federation and symlinks) this design choice makes less sense. This is why we already use a per-filesystem trash/home directory instead of the user's client-configured one - otherwise trash would not work because renames can't span file systems. In short, HDFS trash would work similarly to how it does today, the difference is that client delete APIs would result in a rename into trash (ala TrashPolicyDefault#moveToTrash) if trash is enabled. Like today it would be renamed to the trash directory on the file system where the file being removed resides. The primary difference is that enablement and policy are configured server-side by adminstrators and is used regardless of the API used to access the filesytem. The one execption to this is that I think we should continue to support the explict skipTrash shell option. The rationale for skipTrash (HADOOP-6080) is that a move to trash may fail in cases where a rm may not, if a user has a home directory quota and does a rmr /tonsOfData, for example. Without a way to bypass this the user has no way (unless we revisit quotas, permissions or trash paths) to remove a directory they have permissions to remove without getting their quota adjusted by an admin. The skip trash API can be implemented by adding an explicit FileSystem API that bypasses trash and modifying the shell to use it when skipTrash is enabled. Given that users must explicitly specify skipTrash the API is less error prone. We could have the shell ask confirmation and annotate the API private to FsShell to discourage programatic use. This is not ideal but can be done compatibly (unlike redefining quotas, permissions or trash paths). In terms of compatibility, while this proposal is technically an incompatible change (client side configuration that disables trash and uses skipTrash with a previous FsShell release will now both be ignored if server-side trash is enabled, and non-HDFS file systems would need to make similar changes) I think it's worth targeting for Hadoop 2.x given that the new semantics preserve the current semantics.
[jira] [Commented] (HADOOP-8569) CMakeLists.txt: define _GNU_SOURCE and _LARGEFILE_SOURCE
[ https://issues.apache.org/jira/browse/HADOOP-8569?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13415584#comment-13415584 ] Andy Isaacson commented on HADOOP-8569: --- The downside to defining {{_GNU_SOURCE}} everywhere -- in both code that is known to need it, and in code that is intended to be portable -- is that it makes it very easy to accidentally break compilation on non-GNU platforms by unintentionally changing code from portable to gnu-specific. Suppose we have a project with {{foo-linux.c}} containing nonportable code, and {{generic.c}} containing POSIX portable code. If I define {{_GNU_SOURCE}} in {{CFLAGS}} then unintentionally adding a call to {{sync_file_range}} to {{generic.c}} will silently work on Linux, and won't break the build until you try building on Darwin or Solaris or whatever. If instead the project puts {{#define _GNU_SOURCE}} at the top of files intended to be platform-specific, then such portability breakage will be noticed immediately. The argument doesn't extend to LFS support -- it's entirely reasonable to use 64-bit-{{off_t}} everywhere including Linux-32. {code} +# note: can't enable -D_FILE_OFFSET_BITS=64: see MAPREDUCE-4258 +set(CMAKE_C_FLAGS ${CMAKE_C_FLAGS} -D_REENTRANT -D_GNU_SOURCE) +set(CMAKE_C_FLAGS ${CMAKE_C_FLAGS} -D_LARGEFILE_SOURCE) {code} If we don't have FILE_OFFSET_BITS=64, then shouldn't we also leave out LARGEFILE_SOURCE? (This is a serious question, I don't know how those two defines interact, I just vaguely remember that there's a complicated rule about how they should be used.) In summary -- I'd slightly prefer to limit the GNU_SOURCE to a {{#define}} in the files that we intend to be Linux-specific. The rest of this patch is good though, cleaning up the LFS defines and adding that comment about MAPREDUCE-4258. CMakeLists.txt: define _GNU_SOURCE and _LARGEFILE_SOURCE Key: HADOOP-8569 URL: https://issues.apache.org/jira/browse/HADOOP-8569 Project: Hadoop Common Issue Type: Bug Reporter: Colin Patrick McCabe Assignee: Colin Patrick McCabe Priority: Minor Attachments: HADOOP-8569.001.patch In the native code, we should define _GNU_SOURCE and _LARGEFILE_SOURCE so that all of the functions on Linux are available. _LARGEFILE enables fseeko and ftello; _GNU_SOURCE enables a variety of Linux-specific functions from glibc, including sync_file_range. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HADOOP-8589) ViewFs tests fail when tests and home dirs are nested
[ https://issues.apache.org/jira/browse/HADOOP-8589?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13415587#comment-13415587 ] Ahmed Radwan commented on HADOOP-8589: -- I agree Daryn. Yes, TestViewFsTrash tries to delete the user's home directory!! Which raise the question of what if this testcase fails and it successfully deletes the user's home directory. I think this specific issue need to be prioritized and addressed in a separate ticket. ViewFs tests fail when tests and home dirs are nested - Key: HADOOP-8589 URL: https://issues.apache.org/jira/browse/HADOOP-8589 Project: Hadoop Common Issue Type: Bug Components: fs, test Affects Versions: 0.23.1, 2.0.0-alpha Reporter: Andrey Klochkov Assignee: Andrey Klochkov Attachments: HADOOP-8859.patch TestFSMainOperationsLocalFileSystem fails in case when the test root directory is under the user's home directory, and the user's home dir is deeper than 2 levels from /. This happens with the default 1-node installation of Jenkins. This is the failure log: {code} org.apache.hadoop.fs.FileAlreadyExistsException: Path /var already exists as dir; cannot create link here at org.apache.hadoop.fs.viewfs.InodeTree.createLink(InodeTree.java:244) at org.apache.hadoop.fs.viewfs.InodeTree.init(InodeTree.java:334) at org.apache.hadoop.fs.viewfs.ViewFileSystem$1.init(ViewFileSystem.java:167) at org.apache.hadoop.fs.viewfs.ViewFileSystem.initialize(ViewFileSystem.java:167) at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2094) at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:79) at org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2128) at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2110) at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:290) at org.apache.hadoop.fs.viewfs.ViewFileSystemTestSetup.setupForViewFileSystem(ViewFileSystemTestSetup.java:76) at org.apache.hadoop.fs.viewfs.TestFSMainOperationsLocalFileSystem.setUp(TestFSMainOperationsLocalFileSystem.java:40) ... Standard Output 2012-07-11 22:07:20,239 INFO mortbay.log (Slf4jLog.java:info(67)) - Home dir base /var/lib {code} The reason for the failure is that the code tries to mount links for both /var and /var/lib, and it fails for the 2nd one as the /var is mounted already. The fix was provided in HADOOP-8036 but later it was reverted in HADOOP-8129. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HADOOP-8598) Server-side Trash
[ https://issues.apache.org/jira/browse/HADOOP-8598?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13415596#comment-13415596 ] Daryn Sharp commented on HADOOP-8598: - I haven't digested this jira, but be sure to keep viewfs in mind. Ie. {{FileSystem#getServerDefaults(Path)}} would be required to get the info for the right mount point. Server-side Trash - Key: HADOOP-8598 URL: https://issues.apache.org/jira/browse/HADOOP-8598 Project: Hadoop Common Issue Type: New Feature Affects Versions: 2.0.0-alpha Reporter: Eli Collins Assignee: Eli Collins Priority: Critical There are a number of problems with Trash that continue to result in permanent data loss for users. The primary reasons trash is not used: - Trash is configured client-side and not enabled by default. - Trash is shell-only. FileSystem, WebHDFS, HttpFs, etc never use trash. - If trash fails, for example, because we can't create the trash directory or the move itself fails, trash is bypassed and the data is deleted. Trash was designed as a feature to help end users via the shell, however in my experience the primary use of trash is to help administrators implement data retention policies (this was also the motivation for HADOOP-7460). One could argue that (periodic read-only) snapshots are a better solution to this problem, however snapshots are not slated for Hadoop 2.x and trash is complimentary to snapshots (and backup) - eg you may create and delete data within your snapshot or backup window - so it makes sense to revisit trash's design. I think it's worth bringing trash's functionality in line with what users need. I propose we enable trash on a per-filesystem basis and implement it server-side. Ie trash becomes an HDFS feature enabled by administrators. Because the trash emptier lives in HDFS and users already have a per-filesystem trash directory we're mostly there already. The design preference from HADOOP-2514 was for trash to be implemented in user code however (a) in light of these problems, (b) we have a lot more user-facing APIs than the shell and (c) clients increasingly span file systems (via federation and symlinks) this design choice makes less sense. This is why we already use a per-filesystem trash/home directory instead of the user's client-configured one - otherwise trash would not work because renames can't span file systems. In short, HDFS trash would work similarly to how it does today, the difference is that client delete APIs would result in a rename into trash (ala TrashPolicyDefault#moveToTrash) if trash is enabled. Like today it would be renamed to the trash directory on the file system where the file being removed resides. The primary difference is that enablement and policy are configured server-side by adminstrators and is used regardless of the API used to access the filesytem. The one execption to this is that I think we should continue to support the explict skipTrash shell option. The rationale for skipTrash (HADOOP-6080) is that a move to trash may fail in cases where a rm may not, if a user has a home directory quota and does a rmr /tonsOfData, for example. Without a way to bypass this the user has no way (unless we revisit quotas, permissions or trash paths) to remove a directory they have permissions to remove without getting their quota adjusted by an admin. The skip trash API can be implemented by adding an explicit FileSystem API that bypasses trash and modifying the shell to use it when skipTrash is enabled. Given that users must explicitly specify skipTrash the API is less error prone. We could have the shell ask confirmation and annotate the API private to FsShell to discourage programatic use. This is not ideal but can be done compatibly (unlike redefining quotas, permissions or trash paths). In terms of compatibility, while this proposal is technically an incompatible change (client side configuration that disables trash and uses skipTrash with a previous FsShell release will now both be ignored if server-side trash is enabled, and non-HDFS file systems would need to make similar changes) I think it's worth targeting for Hadoop 2.x given that the new semantics preserve the current semantics. In 2.x I think we should preserve FsShell based trash and support both it and server-side trash (defaults to disabled). For trunk/3.x I think we should remove the FsShell based trash entirely and enable server-side trash by default. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see:
[jira] [Commented] (HADOOP-8589) ViewFs tests fail when tests and home dirs are nested
[ https://issues.apache.org/jira/browse/HADOOP-8589?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13415601#comment-13415601 ] Daryn Sharp commented on HADOOP-8589: - What I'm trying to say is if the viewfs tests are confined to build.test.dir, then the mount path problems _and_ the trash problem are both solved. They are only orthogonal if you believe that viewfs should mount your home dir, and trust it to not munge the contents of you home directory, and/or your ~/.Trash if you are on OS X. This is at least the third jira to fix mounting home dirs. It always fixes for some breaks for others. Are there drawbacks to confining the tests to mounts within build.test.dir? ViewFs tests fail when tests and home dirs are nested - Key: HADOOP-8589 URL: https://issues.apache.org/jira/browse/HADOOP-8589 Project: Hadoop Common Issue Type: Bug Components: fs, test Affects Versions: 0.23.1, 2.0.0-alpha Reporter: Andrey Klochkov Assignee: Andrey Klochkov Attachments: HADOOP-8859.patch TestFSMainOperationsLocalFileSystem fails in case when the test root directory is under the user's home directory, and the user's home dir is deeper than 2 levels from /. This happens with the default 1-node installation of Jenkins. This is the failure log: {code} org.apache.hadoop.fs.FileAlreadyExistsException: Path /var already exists as dir; cannot create link here at org.apache.hadoop.fs.viewfs.InodeTree.createLink(InodeTree.java:244) at org.apache.hadoop.fs.viewfs.InodeTree.init(InodeTree.java:334) at org.apache.hadoop.fs.viewfs.ViewFileSystem$1.init(ViewFileSystem.java:167) at org.apache.hadoop.fs.viewfs.ViewFileSystem.initialize(ViewFileSystem.java:167) at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2094) at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:79) at org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2128) at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2110) at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:290) at org.apache.hadoop.fs.viewfs.ViewFileSystemTestSetup.setupForViewFileSystem(ViewFileSystemTestSetup.java:76) at org.apache.hadoop.fs.viewfs.TestFSMainOperationsLocalFileSystem.setUp(TestFSMainOperationsLocalFileSystem.java:40) ... Standard Output 2012-07-11 22:07:20,239 INFO mortbay.log (Slf4jLog.java:info(67)) - Home dir base /var/lib {code} The reason for the failure is that the code tries to mount links for both /var and /var/lib, and it fails for the 2nd one as the /var is mounted already. The fix was provided in HADOOP-8036 but later it was reverted in HADOOP-8129. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HADOOP-8569) CMakeLists.txt: define _GNU_SOURCE and _LARGEFILE_SOURCE
[ https://issues.apache.org/jira/browse/HADOOP-8569?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13415619#comment-13415619 ] Colin Patrick McCabe commented on HADOOP-8569: -- _GNU_SOURCE was defined previously in most (if not all) of our native projects. After I did the CMake conversion, the fact that it wasn't defined in the CMakeLists.txt was a bug, not a feature. That's what I'm trying to fix here. I realize that it's tempting to assume that code that you write without _GNU_SOURCE defined will automatically be portable. However, this is *NOT TRUE*. For example, even without _GNU_SOURCE defined, you still get the non-POSIX definition of strerror_r out of glibc. The only valid way to make sure your code is portable is to build and test it on multiple platforms. Any other strategy is just a waste of time. Defining GNU_SOURCE is similar to setting the correct DOCTYPE in your HTML file. It tells the browser (or compiler in this case) to turn off quirks mode and give you the real deal. I don't think the CheckFunctionExists stuff in the CMakeLists.txt will work consistently without _GNU_SOURCE defined. There are better ways to improve our portability. For example, we should probably have some OpenBSD jenkins build slaves. But let's not waste our time messing with macros. It really adds nothing but inconvenience. bq. If we don't have FILE_OFFSET_BITS=64, then shouldn't we also leave out LARGEFILE_SOURCE? _LARGEFILE_SOURCE exposes fseeko and ftello. _FILE_OFFSET_BITS changes the default off_t type to be 64 bits. Basically _LARGEFILE_SOURCE is something you need to define in addition to _FILE_OFFSET_BITS, but the two things do different things. CMakeLists.txt: define _GNU_SOURCE and _LARGEFILE_SOURCE Key: HADOOP-8569 URL: https://issues.apache.org/jira/browse/HADOOP-8569 Project: Hadoop Common Issue Type: Bug Reporter: Colin Patrick McCabe Assignee: Colin Patrick McCabe Priority: Minor Attachments: HADOOP-8569.001.patch In the native code, we should define _GNU_SOURCE and _LARGEFILE_SOURCE so that all of the functions on Linux are available. _LARGEFILE enables fseeko and ftello; _GNU_SOURCE enables a variety of Linux-specific functions from glibc, including sync_file_range. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HADOOP-8598) Server-side Trash
[ https://issues.apache.org/jira/browse/HADOOP-8598?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13415620#comment-13415620 ] Eli Collins commented on HADOOP-8598: - @ATM, like your idea, though it will only work with new clients (modified to check the server config) and FsShell it seems like a reasonable approach for branch-2. Server-side Trash - Key: HADOOP-8598 URL: https://issues.apache.org/jira/browse/HADOOP-8598 Project: Hadoop Common Issue Type: New Feature Affects Versions: 2.0.0-alpha Reporter: Eli Collins Assignee: Eli Collins Priority: Critical There are a number of problems with Trash that continue to result in permanent data loss for users. The primary reasons trash is not used: - Trash is configured client-side and not enabled by default. - Trash is shell-only. FileSystem, WebHDFS, HttpFs, etc never use trash. - If trash fails, for example, because we can't create the trash directory or the move itself fails, trash is bypassed and the data is deleted. Trash was designed as a feature to help end users via the shell, however in my experience the primary use of trash is to help administrators implement data retention policies (this was also the motivation for HADOOP-7460). One could argue that (periodic read-only) snapshots are a better solution to this problem, however snapshots are not slated for Hadoop 2.x and trash is complimentary to snapshots (and backup) - eg you may create and delete data within your snapshot or backup window - so it makes sense to revisit trash's design. I think it's worth bringing trash's functionality in line with what users need. I propose we enable trash on a per-filesystem basis and implement it server-side. Ie trash becomes an HDFS feature enabled by administrators. Because the trash emptier lives in HDFS and users already have a per-filesystem trash directory we're mostly there already. The design preference from HADOOP-2514 was for trash to be implemented in user code however (a) in light of these problems, (b) we have a lot more user-facing APIs than the shell and (c) clients increasingly span file systems (via federation and symlinks) this design choice makes less sense. This is why we already use a per-filesystem trash/home directory instead of the user's client-configured one - otherwise trash would not work because renames can't span file systems. In short, HDFS trash would work similarly to how it does today, the difference is that client delete APIs would result in a rename into trash (ala TrashPolicyDefault#moveToTrash) if trash is enabled. Like today it would be renamed to the trash directory on the file system where the file being removed resides. The primary difference is that enablement and policy are configured server-side by adminstrators and is used regardless of the API used to access the filesytem. The one execption to this is that I think we should continue to support the explict skipTrash shell option. The rationale for skipTrash (HADOOP-6080) is that a move to trash may fail in cases where a rm may not, if a user has a home directory quota and does a rmr /tonsOfData, for example. Without a way to bypass this the user has no way (unless we revisit quotas, permissions or trash paths) to remove a directory they have permissions to remove without getting their quota adjusted by an admin. The skip trash API can be implemented by adding an explicit FileSystem API that bypasses trash and modifying the shell to use it when skipTrash is enabled. Given that users must explicitly specify skipTrash the API is less error prone. We could have the shell ask confirmation and annotate the API private to FsShell to discourage programatic use. This is not ideal but can be done compatibly (unlike redefining quotas, permissions or trash paths). In terms of compatibility, while this proposal is technically an incompatible change (client side configuration that disables trash and uses skipTrash with a previous FsShell release will now both be ignored if server-side trash is enabled, and non-HDFS file systems would need to make similar changes) I think it's worth targeting for Hadoop 2.x given that the new semantics preserve the current semantics. In 2.x I think we should preserve FsShell based trash and support both it and server-side trash (defaults to disabled). For trunk/3.x I think we should remove the FsShell based trash entirely and enable server-side trash by default. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see:
[jira] [Created] (HADOOP-8600) libhdfs: fix some compiler warnings
Colin Patrick McCabe created HADOOP-8600: Summary: libhdfs: fix some compiler warnings Key: HADOOP-8600 URL: https://issues.apache.org/jira/browse/HADOOP-8600 Project: Hadoop Common Issue Type: Bug Affects Versions: 2.1.0-alpha Reporter: Colin Patrick McCabe Assignee: Colin Patrick McCabe Priority: Minor Fix some compiler warnings. Some of these are misuses of fprintf (forgetting the FILE* parameter), const mismatch warnings, format specifier warnings. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Work started] (HADOOP-8600) libhdfs: fix some compiler warnings
[ https://issues.apache.org/jira/browse/HADOOP-8600?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Work on HADOOP-8600 started by Colin Patrick McCabe. libhdfs: fix some compiler warnings --- Key: HADOOP-8600 URL: https://issues.apache.org/jira/browse/HADOOP-8600 Project: Hadoop Common Issue Type: Bug Affects Versions: 2.1.0-alpha Reporter: Colin Patrick McCabe Assignee: Colin Patrick McCabe Priority: Minor Fix some compiler warnings. Some of these are misuses of fprintf (forgetting the FILE* parameter), const mismatch warnings, format specifier warnings. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HADOOP-8599) Non empty response from FileSystem.getFileBlockLocations when asking for data beyond the end of file
[ https://issues.apache.org/jira/browse/HADOOP-8599?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Andrey Klochkov updated HADOOP-8599: Attachment: HADOOP-8859-branch-0.23.patch The patch fixes the condition of going beyond EOF in the FileSystem.getFileBlockLocations by changing {code} if (file.getLen() start) { return new BlockLocation[0]; {code} to {code} if (file.getLen() = start) { return new BlockLocation[0]; {code} An additional assert is added to the unit tests to verify the case. Non empty response from FileSystem.getFileBlockLocations when asking for data beyond the end of file - Key: HADOOP-8599 URL: https://issues.apache.org/jira/browse/HADOOP-8599 Project: Hadoop Common Issue Type: Bug Components: fs Affects Versions: 1.0.3, 0.23.1, 2.0.0-alpha Reporter: Andrey Klochkov Attachments: HADOOP-8859-branch-0.23.patch When FileSystem.getFileBlockLocations(file,start,len) is called with start argument equal to the file size, the response is not empty. There is a test TestGetFileBlockLocations.testGetFileBlockLocations2 which uses randomly generated start and len arguments when calling FileSystem.getFileBlockLocations and the test fails randomly (when the generated start value equals to the file size). -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HADOOP-8599) Non empty response from FileSystem.getFileBlockLocations when asking for data beyond the end of file
[ https://issues.apache.org/jira/browse/HADOOP-8599?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Andrey Klochkov updated HADOOP-8599: Assignee: Andrey Klochkov Status: Patch Available (was: Open) Non empty response from FileSystem.getFileBlockLocations when asking for data beyond the end of file - Key: HADOOP-8599 URL: https://issues.apache.org/jira/browse/HADOOP-8599 Project: Hadoop Common Issue Type: Bug Components: fs Affects Versions: 2.0.0-alpha, 0.23.1, 1.0.3 Reporter: Andrey Klochkov Assignee: Andrey Klochkov Attachments: HADOOP-8859-branch-0.23.patch When FileSystem.getFileBlockLocations(file,start,len) is called with start argument equal to the file size, the response is not empty. There is a test TestGetFileBlockLocations.testGetFileBlockLocations2 which uses randomly generated start and len arguments when calling FileSystem.getFileBlockLocations and the test fails randomly (when the generated start value equals to the file size). -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HADOOP-8599) Non empty response from FileSystem.getFileBlockLocations when asking for data beyond the end of file
[ https://issues.apache.org/jira/browse/HADOOP-8599?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13415699#comment-13415699 ] Hadoop QA commented on HADOOP-8599: --- +1 overall. Here are the results of testing the latest attachment http://issues.apache.org/jira/secure/attachment/12536733/HADOOP-8859-branch-0.23.patch against trunk revision . +1 @author. The patch does not contain any @author tags. +1 tests included. The patch appears to include 1 new or modified test files. +1 javac. The applied patch does not increase the total number of javac compiler warnings. +1 javadoc. The javadoc tool did not generate any warning messages. +1 eclipse:eclipse. The patch built with eclipse:eclipse. +1 findbugs. The patch does not introduce any new Findbugs (version 1.3.9) warnings. +1 release audit. The applied patch does not increase the total number of release audit warnings. +1 core tests. The patch passed unit tests in hadoop-common-project/hadoop-common. +1 contrib tests. The patch passed contrib unit tests. Test results: https://builds.apache.org/job/PreCommit-HADOOP-Build/1198//testReport/ Console output: https://builds.apache.org/job/PreCommit-HADOOP-Build/1198//console This message is automatically generated. Non empty response from FileSystem.getFileBlockLocations when asking for data beyond the end of file - Key: HADOOP-8599 URL: https://issues.apache.org/jira/browse/HADOOP-8599 Project: Hadoop Common Issue Type: Bug Components: fs Affects Versions: 1.0.3, 0.23.1, 2.0.0-alpha Reporter: Andrey Klochkov Assignee: Andrey Klochkov Attachments: HADOOP-8859-branch-0.23.patch When FileSystem.getFileBlockLocations(file,start,len) is called with start argument equal to the file size, the response is not empty. There is a test TestGetFileBlockLocations.testGetFileBlockLocations2 which uses randomly generated start and len arguments when calling FileSystem.getFileBlockLocations and the test fails randomly (when the generated start value equals to the file size). -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HADOOP-8522) ResetableGzipOutputStream creates invalid gzip files when finish() and resetState() are used
[ https://issues.apache.org/jira/browse/HADOOP-8522?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13415711#comment-13415711 ] Mike Percy commented on HADOOP-8522: Apologies for the delay in following up on this. Tom, do you expect GzipCodec to be able to read multi-member gzip files on JDK1.6 without native libs installed? ResetableGzipOutputStream creates invalid gzip files when finish() and resetState() are used Key: HADOOP-8522 URL: https://issues.apache.org/jira/browse/HADOOP-8522 Project: Hadoop Common Issue Type: Bug Components: io Affects Versions: 1.0.3, 2.0.0-alpha Reporter: Mike Percy Assignee: Mike Percy Attachments: HADOOP-8522-2a.patch ResetableGzipOutputStream creates invalid gzip files when finish() and resetState() are used. The issue is that finish() flushes the compressor buffer and writes the gzip CRC32 + data length trailer. After that, resetState() does not repeat the gzip header, but simply starts writing more deflate-compressed data. The resultant files are not readable by the Linux gunzip tool. ResetableGzipOutputStream should write valid multi-member gzip files. The gzip format is specified in [RFC 1952|https://tools.ietf.org/html/rfc1952]. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Resolved] (HADOOP-8423) MapFile.Reader.get() crashes jvm or throws EOFException on Snappy or LZO block-compressed data
[ https://issues.apache.org/jira/browse/HADOOP-8423?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Todd Lipcon resolved HADOOP-8423. - Resolution: Fixed Fix Version/s: 1.2.0 Committed to branch-1 for 1.2. Thanks for backporting, Harsh. MapFile.Reader.get() crashes jvm or throws EOFException on Snappy or LZO block-compressed data -- Key: HADOOP-8423 URL: https://issues.apache.org/jira/browse/HADOOP-8423 Project: Hadoop Common Issue Type: Bug Components: io Affects Versions: 0.20.2 Environment: Linux 2.6.32.23-0.3-default #1 SMP 2010-10-07 14:57:45 +0200 x86_64 x86_64 x86_64 GNU/Linux Reporter: Jason B Assignee: Todd Lipcon Fix For: 1.2.0, 2.1.0-alpha Attachments: HADOOP-8423-branch-1.patch, HADOOP-8423-branch-1.patch, MapFileCodecTest.java, hadoop-8423.txt I am using Cloudera distribution cdh3u1. When trying to check native codecs for better decompression performance such as Snappy or LZO, I ran into issues with random access using MapFile.Reader.get(key, value) method. First call of MapFile.Reader.get() works but a second call fails. Also I am getting different exceptions depending on number of entries in a map file. With LzoCodec and 10 record file, jvm gets aborted. At the same time the DefaultCodec works fine for all cases, as well as record compression for the native codecs. I created a simple test program (attached) that creates map files locally with sizes of 10 and 100 records for three codecs: Default, Snappy, and LZO. (The test requires corresponding native library available) The summary of problems are given below: Map Size: 100 Compression: RECORD == DefaultCodec: OK SnappyCodec: OK LzoCodec: OK Map Size: 10 Compression: RECORD == DefaultCodec: OK SnappyCodec: OK LzoCodec: OK Map Size: 100 Compression: BLOCK DefaultCodec: OK SnappyCodec: java.io.EOFException at org.apache.hadoop.io.compress.BlockDecompressorStream.getCompressedData(BlockDecompressorStream.java:114) LzoCodec: java.io.EOFException at org.apache.hadoop.io.compress.BlockDecompressorStream.getCompressedData(BlockDecompressorStream.java:114) Map Size: 10 Compression: BLOCK == DefaultCodec: OK SnappyCodec: java.lang.NoClassDefFoundError: Ljava/lang/InternalError at org.apache.hadoop.io.compress.snappy.SnappyDecompressor.decompressBytesDirect(Native Method) LzoCodec: # # A fatal error has been detected by the Java Runtime Environment: # # SIGSEGV (0xb) at pc=0x2b068ffcbc00, pid=6385, tid=47304763508496 # # JRE version: 6.0_21-b07 # Java VM: Java HotSpot(TM) 64-Bit Server VM (17.0-b17 mixed mode linux-amd64 ) # Problematic frame: # C [liblzo2.so.2+0x13c00] lzo1x_decompress+0x1a0 # -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HADOOP-8593) add the missed @Override to methods in Metric/Metric2 package
[ https://issues.apache.org/jira/browse/HADOOP-8593?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Suresh Srinivas updated HADOOP-8593: Affects Version/s: 3.0.0 1.0.0 add the missed @Override to methods in Metric/Metric2 package -- Key: HADOOP-8593 URL: https://issues.apache.org/jira/browse/HADOOP-8593 Project: Hadoop Common Issue Type: Improvement Components: metrics Affects Versions: 1.0.0, 3.0.0 Reporter: Brandon Li Assignee: Brandon Li Priority: Minor Fix For: 3.0.0 Attachments: HADOOP-8593.patch Adding @Override to the proper methods to take advantage of the compiler checking and make the code more readable. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HADOOP-8593) add the missed @Override to methods in Metric/Metric2 package
[ https://issues.apache.org/jira/browse/HADOOP-8593?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Suresh Srinivas updated HADOOP-8593: Resolution: Fixed Hadoop Flags: Reviewed Status: Resolved (was: Patch Available) I committed the patch. Thank you Brandon. add the missed @Override to methods in Metric/Metric2 package -- Key: HADOOP-8593 URL: https://issues.apache.org/jira/browse/HADOOP-8593 Project: Hadoop Common Issue Type: Improvement Components: metrics Affects Versions: 1.0.0, 3.0.0 Reporter: Brandon Li Assignee: Brandon Li Priority: Minor Fix For: 3.0.0 Attachments: HADOOP-8593.patch Adding @Override to the proper methods to take advantage of the compiler checking and make the code more readable. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HADOOP-8599) Non empty response from FileSystem.getFileBlockLocations when asking for data beyond the end of file
[ https://issues.apache.org/jira/browse/HADOOP-8599?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13415739#comment-13415739 ] Todd Lipcon commented on HADOOP-8599: - +1, looks good to me. Thanks Andrey. Non empty response from FileSystem.getFileBlockLocations when asking for data beyond the end of file - Key: HADOOP-8599 URL: https://issues.apache.org/jira/browse/HADOOP-8599 Project: Hadoop Common Issue Type: Bug Components: fs Affects Versions: 1.0.3, 0.23.1, 2.0.0-alpha Reporter: Andrey Klochkov Assignee: Andrey Klochkov Attachments: HADOOP-8859-branch-0.23.patch When FileSystem.getFileBlockLocations(file,start,len) is called with start argument equal to the file size, the response is not empty. There is a test TestGetFileBlockLocations.testGetFileBlockLocations2 which uses randomly generated start and len arguments when calling FileSystem.getFileBlockLocations and the test fails randomly (when the generated start value equals to the file size). -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HADOOP-8599) Non empty response from FileSystem.getFileBlockLocations when asking for data beyond the end of file
[ https://issues.apache.org/jira/browse/HADOOP-8599?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Todd Lipcon updated HADOOP-8599: Resolution: Fixed Fix Version/s: 2.2.0-alpha 3.0.0 0.23.3 Hadoop Flags: Reviewed Status: Resolved (was: Patch Available) Non empty response from FileSystem.getFileBlockLocations when asking for data beyond the end of file - Key: HADOOP-8599 URL: https://issues.apache.org/jira/browse/HADOOP-8599 Project: Hadoop Common Issue Type: Bug Components: fs Affects Versions: 1.0.3, 0.23.1, 2.0.0-alpha Reporter: Andrey Klochkov Assignee: Andrey Klochkov Fix For: 0.23.3, 3.0.0, 2.2.0-alpha Attachments: HADOOP-8859-branch-0.23.patch When FileSystem.getFileBlockLocations(file,start,len) is called with start argument equal to the file size, the response is not empty. There is a test TestGetFileBlockLocations.testGetFileBlockLocations2 which uses randomly generated start and len arguments when calling FileSystem.getFileBlockLocations and the test fails randomly (when the generated start value equals to the file size). -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HADOOP-8599) Non empty response from FileSystem.getFileBlockLocations when asking for data beyond the end of file
[ https://issues.apache.org/jira/browse/HADOOP-8599?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13415745#comment-13415745 ] Hudson commented on HADOOP-8599: Integrated in Hadoop-Hdfs-trunk-Commit #2551 (See [https://builds.apache.org/job/Hadoop-Hdfs-trunk-Commit/2551/]) HADOOP-8599. Non empty response from FileSystem.getFileBlockLocations when asking for data beyond the end of file. Contributed by Andrey Klochkov. (Revision 1362295) Result = SUCCESS todd : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1362295 Files : * /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FileSystem.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/TestGetFileBlockLocations.java Non empty response from FileSystem.getFileBlockLocations when asking for data beyond the end of file - Key: HADOOP-8599 URL: https://issues.apache.org/jira/browse/HADOOP-8599 Project: Hadoop Common Issue Type: Bug Components: fs Affects Versions: 1.0.3, 0.23.1, 2.0.0-alpha Reporter: Andrey Klochkov Assignee: Andrey Klochkov Fix For: 0.23.3, 3.0.0, 2.2.0-alpha Attachments: HADOOP-8859-branch-0.23.patch When FileSystem.getFileBlockLocations(file,start,len) is called with start argument equal to the file size, the response is not empty. There is a test TestGetFileBlockLocations.testGetFileBlockLocations2 which uses randomly generated start and len arguments when calling FileSystem.getFileBlockLocations and the test fails randomly (when the generated start value equals to the file size). -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HADOOP-8593) add the missed @Override to methods in Metric/Metric2 package
[ https://issues.apache.org/jira/browse/HADOOP-8593?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13415746#comment-13415746 ] Hudson commented on HADOOP-8593: Integrated in Hadoop-Hdfs-trunk-Commit #2551 (See [https://builds.apache.org/job/Hadoop-Hdfs-trunk-Commit/2551/]) HADOOP-8593. Add missed @Override annotations in Metric/Metrics2 package. Contributed by Brandon Li. (Revision 1362294) Result = SUCCESS suresh : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1362294 Files : * /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics/MetricsServlet.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics/MetricsUtil.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics/file/FileContext.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics/ganglia/GangliaContext.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics/ganglia/GangliaContext31.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics/jvm/JvmMetrics.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics/spi/AbstractMetricsContext.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics/spi/CompositeContext.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics/spi/MetricsRecordImpl.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics/spi/NoEmitMetricsContext.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics/spi/NullContext.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics/spi/NullContextWithUpdateThread.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics/util/MetricsIntValue.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics/util/MetricsLongValue.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics/util/MetricsTimeVaryingInt.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics/util/MetricsTimeVaryingLong.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics/util/MetricsTimeVaryingRate.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/impl/MetricsConfig.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/impl/MetricsSinkAdapter.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/impl/MetricsSourceAdapter.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/impl/MetricsSystemImpl.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/lib/MethodMetric.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/lib/MetricsSourceBuilder.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/lib/MutableGaugeLong.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/lib/MutableStat.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/sink/FileSink.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/sink/ganglia/AbstractGangliaSink.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/sink/ganglia/GangliaSink31.java add the missed @Override to methods in Metric/Metric2 package -- Key: HADOOP-8593 URL: https://issues.apache.org/jira/browse/HADOOP-8593 Project: Hadoop Common Issue Type: Improvement Components: metrics Affects Versions: 1.0.0, 3.0.0 Reporter: Brandon Li Assignee: Brandon Li Priority: Minor Fix For: 3.0.0 Attachments: HADOOP-8593.patch Adding @Override to the proper methods to take advantage of the compiler checking and make the code more readable. -- This message is automatically generated by JIRA. If you think it was sent incorrectly,
[jira] [Commented] (HADOOP-8599) Non empty response from FileSystem.getFileBlockLocations when asking for data beyond the end of file
[ https://issues.apache.org/jira/browse/HADOOP-8599?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13415750#comment-13415750 ] Hudson commented on HADOOP-8599: Integrated in Hadoop-Common-trunk-Commit #2486 (See [https://builds.apache.org/job/Hadoop-Common-trunk-Commit/2486/]) HADOOP-8599. Non empty response from FileSystem.getFileBlockLocations when asking for data beyond the end of file. Contributed by Andrey Klochkov. (Revision 1362295) Result = SUCCESS todd : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1362295 Files : * /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FileSystem.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/TestGetFileBlockLocations.java Non empty response from FileSystem.getFileBlockLocations when asking for data beyond the end of file - Key: HADOOP-8599 URL: https://issues.apache.org/jira/browse/HADOOP-8599 Project: Hadoop Common Issue Type: Bug Components: fs Affects Versions: 1.0.3, 0.23.1, 2.0.0-alpha Reporter: Andrey Klochkov Assignee: Andrey Klochkov Fix For: 0.23.3, 3.0.0, 2.2.0-alpha Attachments: HADOOP-8859-branch-0.23.patch When FileSystem.getFileBlockLocations(file,start,len) is called with start argument equal to the file size, the response is not empty. There is a test TestGetFileBlockLocations.testGetFileBlockLocations2 which uses randomly generated start and len arguments when calling FileSystem.getFileBlockLocations and the test fails randomly (when the generated start value equals to the file size). -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HADOOP-8593) add the missed @Override to methods in Metric/Metric2 package
[ https://issues.apache.org/jira/browse/HADOOP-8593?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13415751#comment-13415751 ] Hudson commented on HADOOP-8593: Integrated in Hadoop-Common-trunk-Commit #2486 (See [https://builds.apache.org/job/Hadoop-Common-trunk-Commit/2486/]) HADOOP-8593. Add missed @Override annotations in Metric/Metrics2 package. Contributed by Brandon Li. (Revision 1362294) Result = SUCCESS suresh : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1362294 Files : * /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics/MetricsServlet.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics/MetricsUtil.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics/file/FileContext.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics/ganglia/GangliaContext.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics/ganglia/GangliaContext31.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics/jvm/JvmMetrics.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics/spi/AbstractMetricsContext.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics/spi/CompositeContext.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics/spi/MetricsRecordImpl.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics/spi/NoEmitMetricsContext.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics/spi/NullContext.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics/spi/NullContextWithUpdateThread.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics/util/MetricsIntValue.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics/util/MetricsLongValue.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics/util/MetricsTimeVaryingInt.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics/util/MetricsTimeVaryingLong.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics/util/MetricsTimeVaryingRate.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/impl/MetricsConfig.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/impl/MetricsSinkAdapter.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/impl/MetricsSourceAdapter.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/impl/MetricsSystemImpl.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/lib/MethodMetric.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/lib/MetricsSourceBuilder.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/lib/MutableGaugeLong.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/lib/MutableStat.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/sink/FileSink.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/sink/ganglia/AbstractGangliaSink.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/sink/ganglia/GangliaSink31.java add the missed @Override to methods in Metric/Metric2 package -- Key: HADOOP-8593 URL: https://issues.apache.org/jira/browse/HADOOP-8593 Project: Hadoop Common Issue Type: Improvement Components: metrics Affects Versions: 1.0.0, 3.0.0 Reporter: Brandon Li Assignee: Brandon Li Priority: Minor Fix For: 3.0.0 Attachments: HADOOP-8593.patch Adding @Override to the proper methods to take advantage of the compiler checking and make the code more readable. -- This message is automatically generated by JIRA. If you think it was sent
[jira] [Commented] (HADOOP-8593) add the missed @Override to methods in Metric/Metric2 package
[ https://issues.apache.org/jira/browse/HADOOP-8593?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13415752#comment-13415752 ] Hudson commented on HADOOP-8593: Integrated in Hadoop-Mapreduce-trunk-Commit #2506 (See [https://builds.apache.org/job/Hadoop-Mapreduce-trunk-Commit/2506/]) HADOOP-8593. Add missed @Override annotations in Metric/Metrics2 package. Contributed by Brandon Li. (Revision 1362294) Result = FAILURE suresh : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1362294 Files : * /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics/MetricsServlet.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics/MetricsUtil.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics/file/FileContext.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics/ganglia/GangliaContext.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics/ganglia/GangliaContext31.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics/jvm/JvmMetrics.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics/spi/AbstractMetricsContext.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics/spi/CompositeContext.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics/spi/MetricsRecordImpl.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics/spi/NoEmitMetricsContext.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics/spi/NullContext.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics/spi/NullContextWithUpdateThread.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics/util/MetricsIntValue.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics/util/MetricsLongValue.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics/util/MetricsTimeVaryingInt.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics/util/MetricsTimeVaryingLong.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics/util/MetricsTimeVaryingRate.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/impl/MetricsConfig.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/impl/MetricsSinkAdapter.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/impl/MetricsSourceAdapter.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/impl/MetricsSystemImpl.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/lib/MethodMetric.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/lib/MetricsSourceBuilder.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/lib/MutableGaugeLong.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/lib/MutableStat.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/sink/FileSink.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/sink/ganglia/AbstractGangliaSink.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/sink/ganglia/GangliaSink31.java add the missed @Override to methods in Metric/Metric2 package -- Key: HADOOP-8593 URL: https://issues.apache.org/jira/browse/HADOOP-8593 Project: Hadoop Common Issue Type: Improvement Components: metrics Affects Versions: 1.0.0, 3.0.0 Reporter: Brandon Li Assignee: Brandon Li Priority: Minor Fix For: 3.0.0 Attachments: HADOOP-8593.patch Adding @Override to the proper methods to take advantage of the compiler checking and make the code more readable. -- This message is automatically generated by JIRA. If you think it was sent
[jira] [Commented] (HADOOP-8599) Non empty response from FileSystem.getFileBlockLocations when asking for data beyond the end of file
[ https://issues.apache.org/jira/browse/HADOOP-8599?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13415766#comment-13415766 ] Hudson commented on HADOOP-8599: Integrated in Hadoop-Mapreduce-trunk-Commit #2507 (See [https://builds.apache.org/job/Hadoop-Mapreduce-trunk-Commit/2507/]) HADOOP-8599. Non empty response from FileSystem.getFileBlockLocations when asking for data beyond the end of file. Contributed by Andrey Klochkov. (Revision 1362295) Result = FAILURE todd : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1362295 Files : * /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FileSystem.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/TestGetFileBlockLocations.java Non empty response from FileSystem.getFileBlockLocations when asking for data beyond the end of file - Key: HADOOP-8599 URL: https://issues.apache.org/jira/browse/HADOOP-8599 Project: Hadoop Common Issue Type: Bug Components: fs Affects Versions: 1.0.3, 0.23.1, 2.0.0-alpha Reporter: Andrey Klochkov Assignee: Andrey Klochkov Fix For: 0.23.3, 3.0.0, 2.2.0-alpha Attachments: HADOOP-8859-branch-0.23.patch When FileSystem.getFileBlockLocations(file,start,len) is called with start argument equal to the file size, the response is not empty. There is a test TestGetFileBlockLocations.testGetFileBlockLocations2 which uses randomly generated start and len arguments when calling FileSystem.getFileBlockLocations and the test fails randomly (when the generated start value equals to the file size). -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Created] (HADOOP-8601) Extend TestShell to cover Windows shell commands
Chuan Liu created HADOOP-8601: - Summary: Extend TestShell to cover Windows shell commands Key: HADOOP-8601 URL: https://issues.apache.org/jira/browse/HADOOP-8601 Project: Hadoop Common Issue Type: Test Affects Versions: 1-win Reporter: Chuan Liu Assignee: Chuan Liu The existing unit test only covers Linux shell commands. Since we begin to support Windows and use completely different commands on Windows, it make sense to extend TestShell to cover Windows use cases. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HADOOP-8453) Add unit tests for winutils
[ https://issues.apache.org/jira/browse/HADOOP-8453?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13415793#comment-13415793 ] Chuan Liu commented on HADOOP-8453: --- HADOOP-8601 is created to track JUnit for Windows shell commands. {quote} We need to make sure these tests are executed on Windows. One option would be to check for that at startup of the test and return success for other OS's. {quote} The Ant task uses 'if windows' to only execute the test if 'windows' is set to 'true'. This is on par with 'winutils' build task. Cmd batch scripts are not support on Linux. So I don't think we can check OS's inside the scripts. This is also true for other Linux only test scripts, e.g. test-libhdfs.sh. Add unit tests for winutils --- Key: HADOOP-8453 URL: https://issues.apache.org/jira/browse/HADOOP-8453 Project: Hadoop Common Issue Type: Task Components: test Affects Versions: 1.1.0, 0.24.0 Reporter: Chuan Liu Assignee: Chuan Liu Priority: Minor Attachments: HADOOP-8453-branch-1-win-2.patch, HADOOP-8453-branch-1-win.patch In [Hadoop-8235|https://issues.apache.org/jira/browse/HADOOP-8235], we created a Windows console program, named ‘winutils’, to emulate some Linux command line utilities used by Hadoop. However no tests are provided in the original patch. As this code is quite complicated, and the complexity may even grow up in the future. We think unit tests are necessary to ensure code quality, as well as smooth future development. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HADOOP-8453) Add unit tests for winutils
[ https://issues.apache.org/jira/browse/HADOOP-8453?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13415821#comment-13415821 ] Bikas Saha commented on HADOOP-8453: Sorry! I incorrectly remembered that this was a JUnit test. Having it as a junit test will let it work properly with the rest of test/reporting. Eg. tell you how many tests failed etc. I think we should exec the bat script inside a junit testcase when the OS is windows. We could do that as part of 8601 or even remove these bat scripts if those new testcases cover the ones in the bat script. +1. Add unit tests for winutils --- Key: HADOOP-8453 URL: https://issues.apache.org/jira/browse/HADOOP-8453 Project: Hadoop Common Issue Type: Task Components: test Affects Versions: 1.1.0, 0.24.0 Reporter: Chuan Liu Assignee: Chuan Liu Priority: Minor Attachments: HADOOP-8453-branch-1-win-2.patch, HADOOP-8453-branch-1-win.patch In [Hadoop-8235|https://issues.apache.org/jira/browse/HADOOP-8235], we created a Windows console program, named ‘winutils’, to emulate some Linux command line utilities used by Hadoop. However no tests are provided in the original patch. As this code is quite complicated, and the complexity may even grow up in the future. We think unit tests are necessary to ensure code quality, as well as smooth future development. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Created] (HADOOP-8602) Passive mode support for FTPFileSystem
nemon lou created HADOOP-8602: - Summary: Passive mode support for FTPFileSystem Key: HADOOP-8602 URL: https://issues.apache.org/jira/browse/HADOOP-8602 Project: Hadoop Common Issue Type: Improvement Components: fs Affects Versions: 2.0.0-alpha, 1.0.3 Reporter: nemon lou Priority: Minor FTPFileSystem uses active mode for default data connection mode.We shall be able to choose passive mode when active mode doesn't work (firewall for example). My thoughts is to add an option fs.ftp.data.connection.mode in core-site.xml.Since FTPClient(in org.apache.commons.net.ftp package) already supports passive mode, we just need to add a few code in FTPFileSystem .connect() method. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira