[jira] [Created] (HADOOP-17569) Building native code fails on Fedora 33
Kengo Seki created HADOOP-17569: --- Summary: Building native code fails on Fedora 33 Key: HADOOP-17569 URL: https://issues.apache.org/jira/browse/HADOOP-17569 Project: Hadoop Common Issue Type: Improvement Components: build, common Reporter: Kengo Seki I tried to build native code on Fedora 33, in which glibc 2.32 is installed by default, but it failed with the following error. {code:java} $ cat /etc/redhat-release Fedora release 33 (Thirty Three) $ sudo dnf info --installed glibc Installed Packages Name : glibc Version : 2.32 Release : 1.fc33 Architecture : x86_64 Size : 17 M Source : glibc-2.32-1.fc33.src.rpm Repository : @System >From repo: anaconda Summary : The GNU libc libraries URL : http://www.gnu.org/software/glibc/ License : LGPLv2+ and LGPLv2+ with exceptions and GPLv2+ and GPLv2+ with exceptions and BSD and Inner-Net and ISC and Public Domain and GFDL Description : The glibc package contains standard libraries which are used by : multiple programs on the system. In order to save disk space and : memory, as well as to make upgrading easier, common system code is : kept in one place and shared between programs. This particular package : contains the most important sets of shared libraries: the standard C : library and the standard math library. Without these two libraries, a : Linux system will not function. $ mvn clean compile -Pnative ... [INFO] Running make -j 1 VERBOSE=1 [WARNING] /usr/bin/cmake -S/home/vagrant/hadoop/hadoop-common-project/hadoop-common/src -B/home/vagrant/hadoop/hadoop-common-project/hadoop-common/target/native --check-build-system CMakeFiles/Makefile.cmake 0 [WARNING] /usr/bin/cmake -E cmake_progress_start /home/vagrant/hadoop/hadoop-common-project/hadoop-common/target/native/CMakeFiles /home/vagrant/hadoop/hadoop-common-project/hadoop-common/target/native//CMakeFiles/progress.marks [WARNING] make -f CMakeFiles/Makefile2 all [WARNING] make[1]: Entering directory '/home/vagrant/hadoop/hadoop-common-project/hadoop-common/target/native' [WARNING] make -f CMakeFiles/hadoop_static.dir/build.make CMakeFiles/hadoop_static.dir/depend [WARNING] make[2]: Entering directory '/home/vagrant/hadoop/hadoop-common-project/hadoop-common/target/native' [WARNING] cd /home/vagrant/hadoop/hadoop-common-project/hadoop-common/target/native && /usr/bin/cmake -E cmake_depends "Unix Makefiles" /home/vagrant/hadoop/hadoop-common-project/hadoop-common/src /home/vagrant/hadoop/hadoop-common-project/hadoop-common/src /home/vagrant/hadoop/hadoop-common-project/hadoop-common/target/native /home/vagrant/hadoop/hadoop-common-project/hadoop-common/target/native /home/vagrant/hadoop/hadoop-common-project/hadoop-common/target/native/CMakeFiles/hadoop_static.dir/DependInfo.cmake --color= [WARNING] Dependee "/home/vagrant/hadoop/hadoop-common-project/hadoop-common/target/native/CMakeFiles/hadoop_static.dir/DependInfo.cmake" is newer than depender "/home/vagrant/hadoop/hadoop-common-project/hadoop-common/target/native/CMakeFiles/hadoop_static.dir/depend.internal". [WARNING] Dependee "/home/vagrant/hadoop/hadoop-common-project/hadoop-common/target/native/CMakeFiles/CMakeDirectoryInformation.cmake" is newer than depender "/home/vagrant/hadoop/hadoop-common-project/hadoop-common/target/native/CMakeFiles/hadoop_static.dir/depend.internal". [WARNING] Scanning dependencies of target hadoop_static [WARNING] make[2]: Leaving directory '/home/vagrant/hadoop/hadoop-common-project/hadoop-common/target/native' [WARNING] make -f CMakeFiles/hadoop_static.dir/build.make CMakeFiles/hadoop_static.dir/build [WARNING] make[2]: Entering directory '/home/vagrant/hadoop/hadoop-common-project/hadoop-common/target/native' [WARNING] [ 2%] Building C object CMakeFiles/hadoop_static.dir/main/native/src/exception.c.o [WARNING] /usr/bin/cc -I/home/vagrant/hadoop/hadoop-common-project/hadoop-common/target/native/javah -I/home/vagrant/hadoop/hadoop-common-project/hadoop-common/src/main/native/src -I/home/vagrant/hadoop/hadoop-common-project/hadoop-common/src -I/home/vagrant/hadoop/hadoop-common-project/hadoop-common/src/src -I/home/vagrant/hadoop/hadoop-common-project/hadoop-common/target/native -I/usr/lib/jvm/java-1.8.0/include -I/usr/lib/jvm/java-1.8.0/include/linux -I/home/vagrant/hadoop/hadoop-common-project/hadoop-common/src/main/native/src/org/apache/hadoop/util -g -O2 -Wall -pthread -D_FILE_OFFSET_BITS=64 -D_GNU_SOURCE -std=gnu99 -o CMakeFiles/hadoop_static.dir/main/native/src/exception.c.o -c /home/vagrant/hadoop/hadoop-common-project/hadoop-common/src/main/native/src/exception.c [WARNING] make[2]: Leaving directory '/home/vagrant/hadoop/hadoop-common-project/hadoop-common/target/native' [WARNING] make[1]: Leaving directory
[jira] [Updated] (HADOOP-16764) Rewrite Python example codes using Python3
[ https://issues.apache.org/jira/browse/HADOOP-16764?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Kengo Seki updated HADOOP-16764: Status: Patch Available (was: Open) > Rewrite Python example codes using Python3 > -- > > Key: HADOOP-16764 > URL: https://issues.apache.org/jira/browse/HADOOP-16764 > Project: Hadoop Common > Issue Type: Sub-task > Components: documentation >Reporter: Kengo Seki >Assignee: Kengo Seki >Priority: Minor > -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Created] (HADOOP-16764) Rewrite Python example codes using Python3
Kengo Seki created HADOOP-16764: --- Summary: Rewrite Python example codes using Python3 Key: HADOOP-16764 URL: https://issues.apache.org/jira/browse/HADOOP-16764 Project: Hadoop Common Issue Type: Sub-task Components: documentation Reporter: Kengo Seki Assignee: Kengo Seki -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Commented] (HADOOP-16747) Support Python 3
[ https://issues.apache.org/jira/browse/HADOOP-16747?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16988548#comment-16988548 ] Kengo Seki commented on HADOOP-16747: - Example scripts in the document such as https://hadoop.apache.org/docs/current/hadoop-project-dist/hadoop-common/RackAwareness.html and https://hadoop.apache.org/docs/current/hadoop-streaming/HadoopStreaming.html#Hadoop_Aggregate_Package should be also updated ;) > Support Python 3 > > > Key: HADOOP-16747 > URL: https://issues.apache.org/jira/browse/HADOOP-16747 > Project: Hadoop Common > Issue Type: Improvement >Reporter: Akira Ajisaka >Priority: Major > > Python 2.x will become EoL at the end of 2019. > This is an umbrella JIRA to support Python 3 and drop Python 2.x. > Here is the python scripts that Hadoop has: > {noformat} > $ find . -name "*.py" | grep -v "hadoop-submarine" > ./hadoop-mapreduce-project/hadoop-mapreduce-examples/src/main/java/org/apache/hadoop/examples/terasort/job_history_summary.py > ./dev-support/determine-flaky-tests-hadoop.py > ./dev-support/bin/checkcompatibility.py > {noformat} -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Created] (HADOOP-14315) Python example in the rack awareness document doesn't work due to bad indentation
Kengo Seki created HADOOP-14315: --- Summary: Python example in the rack awareness document doesn't work due to bad indentation Key: HADOOP-14315 URL: https://issues.apache.org/jira/browse/HADOOP-14315 Project: Hadoop Common Issue Type: Bug Components: documentation Reporter: Kengo Seki Assignee: Kengo Seki Priority: Minor Running that example fails with: {code} File "example.py", line 28 address = '{0}/{1}'.format(ip, netmask) # format address string so it looks like 'ip/netmask' to make netaddr work ^ IndentationError: expected an indented block {code} -- This message was sent by Atlassian JIRA (v6.3.15#6346) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Commented] (HADOOP-13193) Upgrade to Apache Yetus 0.3.0
[ https://issues.apache.org/jira/browse/HADOOP-13193?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15305019#comment-15305019 ] Kengo Seki commented on HADOOP-13193: - Sorry for not explaining enough about qbt, and thanks for the additional explanation and committing Chris! > Upgrade to Apache Yetus 0.3.0 > - > > Key: HADOOP-13193 > URL: https://issues.apache.org/jira/browse/HADOOP-13193 > Project: Hadoop Common > Issue Type: Improvement > Components: documentation, test >Affects Versions: 3.0.0-alpha1 >Reporter: Allen Wittenauer >Assignee: Kengo Seki > Fix For: 2.8.0 > > Attachments: HADOOP-13193.1.patch > > > Upgrade yetus-wrapper to be 0.3.0 now that it has passed vote. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Updated] (HADOOP-13193) Upgrade to Apache Yetus 0.3.0
[ https://issues.apache.org/jira/browse/HADOOP-13193?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Kengo Seki updated HADOOP-13193: Attachment: HADOOP-13193.1.patch -01 * bump the version of Yetus to 0.3.0 * add qbt (quality build tool) command > Upgrade to Apache Yetus 0.3.0 > - > > Key: HADOOP-13193 > URL: https://issues.apache.org/jira/browse/HADOOP-13193 > Project: Hadoop Common > Issue Type: Improvement > Components: documentation, test >Affects Versions: 3.0.0-alpha1 >Reporter: Allen Wittenauer >Assignee: Kengo Seki > Attachments: HADOOP-13193.1.patch > > > Upgrade yetus-wrapper to be 0.3.0 now that it has passed vote. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Updated] (HADOOP-13193) Upgrade to Apache Yetus 0.3.0
[ https://issues.apache.org/jira/browse/HADOOP-13193?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Kengo Seki updated HADOOP-13193: Status: Patch Available (was: Open) > Upgrade to Apache Yetus 0.3.0 > - > > Key: HADOOP-13193 > URL: https://issues.apache.org/jira/browse/HADOOP-13193 > Project: Hadoop Common > Issue Type: Improvement > Components: documentation, test >Affects Versions: 3.0.0-alpha1 >Reporter: Allen Wittenauer >Assignee: Kengo Seki > Attachments: HADOOP-13193.1.patch > > > Upgrade yetus-wrapper to be 0.3.0 now that it has passed vote. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Assigned] (HADOOP-13193) Upgrade to Apache Yetus 0.3.0
[ https://issues.apache.org/jira/browse/HADOOP-13193?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Kengo Seki reassigned HADOOP-13193: --- Assignee: Kengo Seki (was: Allen Wittenauer) > Upgrade to Apache Yetus 0.3.0 > - > > Key: HADOOP-13193 > URL: https://issues.apache.org/jira/browse/HADOOP-13193 > Project: Hadoop Common > Issue Type: Improvement > Components: documentation, test >Affects Versions: 3.0.0-alpha1 >Reporter: Allen Wittenauer >Assignee: Kengo Seki > > Upgrade yetus-wrapper to be 0.3.0 now that it has passed vote. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Commented] (HADOOP-12691) Add CSRF Filter for REST APIs to Hadoop Common
[ https://issues.apache.org/jira/browse/HADOOP-12691?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15103219#comment-15103219 ] Kengo Seki commented on HADOOP-12691: - Directory hadoop-common at the top level seems to be newly created by mistake. https://github.com/apache/hadoop/tree/trunk/hadoop-common > Add CSRF Filter for REST APIs to Hadoop Common > -- > > Key: HADOOP-12691 > URL: https://issues.apache.org/jira/browse/HADOOP-12691 > Project: Hadoop Common > Issue Type: New Feature > Components: security >Reporter: Larry McCay >Assignee: Larry McCay > Fix For: 2.9.0 > > Attachments: CSRFProtectionforRESTAPIs.pdf, HADOOP-12691-001.patch, > HADOOP-12691-002.patch, HADOOP-12691-003.patch > > > CSRF prevention for REST APIs can be provided through a common servlet > filter. This filter would check for the existence of an expected > (configurable) HTTP header - such as X-XSRF-Header. > The fact that CSRF attacks are entirely browser based means that the above > approach can ensure that requests are coming from either: applications served > by the same origin as the REST API or that there is explicit policy > configuration that allows the setting of a header on XmlHttpRequest from > another origin. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HADOOP-12691) Add CSRF Filter for REST APIs to Hadoop Common
[ https://issues.apache.org/jira/browse/HADOOP-12691?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15103222#comment-15103222 ] Kengo Seki commented on HADOOP-12691: - No, your patch seems correct. I suspect it was applied with a wrong argument for "-p" option. > Add CSRF Filter for REST APIs to Hadoop Common > -- > > Key: HADOOP-12691 > URL: https://issues.apache.org/jira/browse/HADOOP-12691 > Project: Hadoop Common > Issue Type: New Feature > Components: security >Reporter: Larry McCay >Assignee: Larry McCay > Fix For: 2.9.0 > > Attachments: CSRFProtectionforRESTAPIs.pdf, HADOOP-12691-001.patch, > HADOOP-12691-002.patch, HADOOP-12691-003.patch > > > CSRF prevention for REST APIs can be provided through a common servlet > filter. This filter would check for the existence of an expected > (configurable) HTTP header - such as X-XSRF-Header. > The fact that CSRF attacks are entirely browser based means that the above > approach can ensure that requests are coming from either: applications served > by the same origin as the REST API or that there is explicit policy > configuration that allows the setting of a header on XmlHttpRequest from > another origin. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HADOOP-12691) Add CSRF Filter for REST APIs to Hadoop Common
[ https://issues.apache.org/jira/browse/HADOOP-12691?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15103210#comment-15103210 ] Kengo Seki commented on HADOOP-12691: - [~cnauroth], I'm afraid you seemed to put the new classes in the wrong position. > Add CSRF Filter for REST APIs to Hadoop Common > -- > > Key: HADOOP-12691 > URL: https://issues.apache.org/jira/browse/HADOOP-12691 > Project: Hadoop Common > Issue Type: New Feature > Components: security >Reporter: Larry McCay >Assignee: Larry McCay > Fix For: 2.9.0 > > Attachments: CSRFProtectionforRESTAPIs.pdf, HADOOP-12691-001.patch, > HADOOP-12691-002.patch, HADOOP-12691-003.patch > > > CSRF prevention for REST APIs can be provided through a common servlet > filter. This filter would check for the existence of an expected > (configurable) HTTP header - such as X-XSRF-Header. > The fact that CSRF attacks are entirely browser based means that the above > approach can ensure that requests are coming from either: applications served > by the same origin as the REST API or that there is explicit policy > configuration that allows the setting of a header on XmlHttpRequest from > another origin. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HADOOP-12681) start-build-env.sh fails in branch-2
[ https://issues.apache.org/jira/browse/HADOOP-12681?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Kengo Seki updated HADOOP-12681: Attachment: HADOOP-12681.branch-2.00.patch Attaching a patch. After applying this, test-patch works in docker mode on branch-2, as follows: {code} [sekikn@localhost hadoop]$ /home/sekikn/yetus/precommit/smart-apply-patch.sh --plugins=all ~/HADOOP-12681.branch-2.00.patch Processing: /home/sekikn/HADOOP-12681.branch-2.00.patch Patch file /home/sekikn/HADOOP-12681.branch-2.00.patch copied to /tmp/yetus-5197.32233 Applying the patch: Sun Dec 27 23:41:33 JST 2015 cd /home/sekikn/dev/hadoop git apply --binary -v --stat --apply -p0 /tmp/yetus-5197.32233/patch Applied patch dev-support/docker/Dockerfile cleanly. dev-support/docker/Dockerfile |4 1 file changed, 4 deletions(-) [sekikn@localhost hadoop]$ /home/sekikn/yetus/precommit/test-patch.sh --dirty-workspace --docker --project=hadoop MAPREDUCE-6584 /tmp/yetus-4717.16695 has been created Running in developer mode Processing: MAPREDUCE-6584 MAPREDUCE-6584 patch is being downloaded at Sun Dec 27 23:46:06 JST 2015 from https://issues.apache.org/jira/secure/attachment/12779516/MAPREDUCE-6584-branch-2.01.patch (snip) -1 overall _ _ __ | ___|_ _(_) |_ _ _ __ ___| | | |_ / _` | | | | | | '__/ _ \ | | _| (_| | | | |_| | | | __/_| |_| \__,_|_|_|\__,_|_| \___(_) | Vote | Subsystem | Runtime | Comment | 0 |reexec | 0m 0s | Docker mode activated. | +1 | @author | 0m 0s | The patch does not contain any @author | ||| tags. | -1 |test4tests | 0m 0s | The patch doesn't appear to include any | ||| new or modified tests. Please justify why | ||| no new tests are needed for this patch. | ||| Also please list what manual steps were | ||| performed to verify this patch. | +1 |mvninstall | 6m 10s| branch-2 passed | +1 | compile | 0m 21s| branch-2 passed | +1 | mvnsite | 0m 27s| branch-2 passed | +1 |mvneclipse | 0m 12s| branch-2 passed | +1 | javadoc | 0m 21s| branch-2 passed | +1 |mvninstall | 0m 24s| the patch passed | +1 | compile | 0m 21s| the patch passed | +1 | javac | 0m 21s| the patch passed | +1 | mvnsite | 0m 28s| the patch passed | +1 |mvneclipse | 0m 11s| the patch passed | +1 |whitespace | 0m 0s | Patch has no whitespace issues. | +1 | xml | 0m 0s | The patch has no ill-formed XML file. | +1 | javadoc | 0m 21s| the patch passed | +1 |asflicense | 0m 18s| Patch does not generate ASF License | ||| warnings. | || 10m 2s| || Subsystem || Report/Notes || | Docker | Client=1.7.1 Server=1.7.1 Image:yetus/hadoop:577e74f | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12779516/MAPREDUCE-6584-branch-2.01.patch | | JIRA Issue | MAPREDUCE-6584 | | Optional Tests | asflicense compile javac javadoc mvninstall mvnsite unit xml | | uname | Linux 7ba42874cad1 2.6.32-573.8.1.el6.x86_64 #1 SMP Tue Nov 10 18:01:38 UTC 2015 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality | /testptch/patchprocess/precommit/personality/provided.sh | | git revision | branch-2 / 42160d3 | | Default Java | 1.7.0_80 | | modules | C: hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core U: hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core | | Max memory used | 75MB | | Powered by | Apache Yetus 0.2.0-SNAPSHOT http://yetus.apache.org | {code} > start-build-env.sh fails in branch-2 > > > Key: HADOOP-12681 > URL: https://issues.apache.org/jira/browse/HADOOP-12681 > Project: Hadoop Common > Issue Type: Bug > Components: build >Reporter: Akira AJISAKA >Priority: Blocker > Labels: newbie > Attachments: HADOOP-12681.branch-2.00.patch > > > start-build-env.sh fails in branch-2. Found in MAPREDUCE-6584. > https://builds.apache.org/job/PreCommit-MAPREDUCE-Build/6236/console -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HADOOP-12681) start-build-env.sh fails in branch-2
[ https://issues.apache.org/jira/browse/HADOOP-12681?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Kengo Seki updated HADOOP-12681: Assignee: Kengo Seki Status: Patch Available (was: Open) > start-build-env.sh fails in branch-2 > > > Key: HADOOP-12681 > URL: https://issues.apache.org/jira/browse/HADOOP-12681 > Project: Hadoop Common > Issue Type: Bug > Components: build >Reporter: Akira AJISAKA >Assignee: Kengo Seki >Priority: Blocker > Labels: newbie > Attachments: HADOOP-12681.branch-2.00.patch > > > start-build-env.sh fails in branch-2. Found in MAPREDUCE-6584. > https://builds.apache.org/job/PreCommit-MAPREDUCE-Build/6236/console -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HADOOP-12273) releasedocmaker.py fails with stacktrace if --project option is not specified
[ https://issues.apache.org/jira/browse/HADOOP-12273?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15063263#comment-15063263 ] Kengo Seki commented on HADOOP-12273: - Hi [~lewuathe], the development of releasedocmaker has been stopped on Hadoop project and migrated to Yetus. HADOOP-12111 branch was already removed. You could backport this patch to trunk or branch-2, but I strongly recommend you to use the one included in [the latest Yetus release|http://yetus.apache.org/downloads/]. Many improvements and bug-fixes are applied. > releasedocmaker.py fails with stacktrace if --project option is not specified > - > > Key: HADOOP-12273 > URL: https://issues.apache.org/jira/browse/HADOOP-12273 > Project: Hadoop Common > Issue Type: Sub-task > Components: yetus >Reporter: Kengo Seki >Assignee: Kengo Seki >Priority: Trivial > Fix For: HADOOP-12111 > > Attachments: HADOOP-12273.HADOOP-12111.00.patch, > HADOOP-12273.HADOOP-12111.01.patch > > > It should show its usage instead. > {code} > [sekikn@mobile hadoop]$ dev-support/releasedocmaker.py --version 3.0.0 > Traceback (most recent call last): > File "dev-support/releasedocmaker.py", line 580, in > main() > File "dev-support/releasedocmaker.py", line 424, in main > title=projects[0] > TypeError: 'NoneType' object has no attribute '__getitem__' > {code} -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Assigned] (HADOOP-9637) Adding Native Fstat for Windows as needed by YARN
[ https://issues.apache.org/jira/browse/HADOOP-9637?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Kengo Seki reassigned HADOOP-9637: -- Assignee: Kengo Seki (was: Chuan Liu) > Adding Native Fstat for Windows as needed by YARN > - > > Key: HADOOP-9637 > URL: https://issues.apache.org/jira/browse/HADOOP-9637 > Project: Hadoop Common > Issue Type: Bug >Affects Versions: 2.1.0-beta, 3.0.0 >Reporter: Chuan Liu >Assignee: Kengo Seki > Fix For: 2.1.0-beta > > Attachments: HADOOP-9637-trunk.2.patch, HADOOP-9637-trunk.3.patch, > HADOOP-9637-trunk.patch > > > In the YARN, nodemanager need to enforce the log file can only be accessed by > the owner. At various places, {{SecureIOUtils.openForRead()}} was called to > enforce this check. We don't have {{NativeIO.Posix.getFstat()}} used by > {{SecureIOUtils.openForRead()}} on Windows, and this make the check fail on > Windows. The YARN unit tests > TestAggregatedLogFormat.testContainerLogsFileAccess and > TestContainerLogsPage.testContainerLogPageAccess fail on Windows because of > this. > The JIRA try to provide a Windows implementation of > {{NativeIO.Posix.getFstat()}}. > TestAggregatedLogFormat.testContainerLogsFileAccess test case fails on > Windows. The test case try to simulate a situation where first log file is > owned by different user (probably symlink) and second one by the user itself. > In this situation, the attempt to try to aggregate the logs should fail with > the error message "Owner ... for path ... did not match expected owner ...". > The check on file owner happens at {{AggregatedLogFormat.write()}} method. > The method calls {{SecureIOUtils.openForRead()}} to read the log files before > writing out to the OutputStream. > {{SecureIOUtils.openForRead()}} use {{NativeIO.Posix.getFstat()}} to get the > file owner and group. We don't have {{NativeIO.Posix.getFstat()}} > implementation on Windows; thus, the failure. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HADOOP-9637) Adding Native Fstat for Windows as needed by YARN
[ https://issues.apache.org/jira/browse/HADOOP-9637?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Kengo Seki updated HADOOP-9637: --- Assignee: Chuan Liu (was: Kengo Seki) > Adding Native Fstat for Windows as needed by YARN > - > > Key: HADOOP-9637 > URL: https://issues.apache.org/jira/browse/HADOOP-9637 > Project: Hadoop Common > Issue Type: Bug >Affects Versions: 2.1.0-beta, 3.0.0 >Reporter: Chuan Liu >Assignee: Chuan Liu > Fix For: 2.1.0-beta > > Attachments: HADOOP-9637-trunk.2.patch, HADOOP-9637-trunk.3.patch, > HADOOP-9637-trunk.patch > > > In the YARN, nodemanager need to enforce the log file can only be accessed by > the owner. At various places, {{SecureIOUtils.openForRead()}} was called to > enforce this check. We don't have {{NativeIO.Posix.getFstat()}} used by > {{SecureIOUtils.openForRead()}} on Windows, and this make the check fail on > Windows. The YARN unit tests > TestAggregatedLogFormat.testContainerLogsFileAccess and > TestContainerLogsPage.testContainerLogPageAccess fail on Windows because of > this. > The JIRA try to provide a Windows implementation of > {{NativeIO.Posix.getFstat()}}. > TestAggregatedLogFormat.testContainerLogsFileAccess test case fails on > Windows. The test case try to simulate a situation where first log file is > owned by different user (probably symlink) and second one by the user itself. > In this situation, the attempt to try to aggregate the logs should fail with > the error message "Owner ... for path ... did not match expected owner ...". > The check on file owner happens at {{AggregatedLogFormat.write()}} method. > The method calls {{SecureIOUtils.openForRead()}} to read the log files before > writing out to the OutputStream. > {{SecureIOUtils.openForRead()}} use {{NativeIO.Posix.getFstat()}} to get the > file owner and group. We don't have {{NativeIO.Posix.getFstat()}} > implementation on Windows; thus, the failure. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HADOOP-4258) the test patch script should check for filenames that differ only in case
[ https://issues.apache.org/jira/browse/HADOOP-4258?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14902886#comment-14902886 ] Kengo Seki commented on HADOOP-4258: Sorry [~jagadesh.kiran], I'm on vacation until Sep. 27th. I'll review your patch next Monday. > the test patch script should check for filenames that differ only in case > - > > Key: HADOOP-4258 > URL: https://issues.apache.org/jira/browse/HADOOP-4258 > Project: Hadoop Common > Issue Type: Sub-task > Components: build, test >Reporter: Owen O'Malley >Assignee: Jagadesh Kiran N > Labels: test-patch > Attachments: HADOOP-4258.001.patch, > HADOOP-4258.HADOOP-12111.00.patch, HADOOP-4258.HADOOP-12111.01.patch, > HADOOP-4258.HADOOP-12111.02.patch > > > It would be nice if the test patch script warned about filenames that differ > only in case. We recently had a patch committed that had a pair of colliding > filenames and subversion broke badly on my Mac. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HADOOP-4258) the test patch script should check for filenames that differ only in case
[ https://issues.apache.org/jira/browse/HADOOP-4258?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14746778#comment-14746778 ] Kengo Seki commented on HADOOP-4258: Thanks [~jagadesh.kiran], some comments: * Shortening only the filename is not sufficient. We must shorten "duplicatednames" in the file, because the arguement of add_test is used as the plugin name. See [Yetus precommit document about plugin name|https://github.com/apache/hadoop/blob/HADOOP-12111/dev-support/docs/precommit-advanced.md#common-plug-in-functions]. * The following is probably different from what Allen originally meant. This just sorts changed_files, while {{uniq -id}} outputs only duplicated lines ignoring case. {code} IFS=$'\n' changed_files=($(sort --ignore-case <<<"${CHANGED_FILES[*]}")) {code} * The following line will fail because the closing brace is missing. {code} for ((j=i+1; j < ${#changed_files[@]; j++)) { {code} * Address shellcheck and whitespace warnings. To make sure, the comments which are not addressed in this patch, except for mentioned above: bq. the problem that Owen O'Malley is actually talking about is a problem where a new file was added because it conflicted with a pre-existing file in the source tree. This patch would miss that situation. bq. $a and $b should be changed, like "$first_lower", "$second_lower". bq. if HADOOP-12257 gets committed, this should use _precheck instead of _postcheckout. > the test patch script should check for filenames that differ only in case > - > > Key: HADOOP-4258 > URL: https://issues.apache.org/jira/browse/HADOOP-4258 > Project: Hadoop Common > Issue Type: Sub-task > Components: build, test >Reporter: Owen O'Malley >Assignee: Jagadesh Kiran N > Labels: test-patch > Attachments: HADOOP-4258.001.patch, > HADOOP-4258.HADOOP-12111.00.patch, HADOOP-4258.HADOOP-12111.01.patch > > > It would be nice if the test patch script warned about filenames that differ > only in case. We recently had a patch committed that had a pair of colliding > filenames and subversion broke badly on my Mac. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HADOOP-12399) Wrong help messages in some test-patch plugins
[ https://issues.apache.org/jira/browse/HADOOP-12399?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14746465#comment-14746465 ] Kengo Seki commented on HADOOP-12399: - Thanks [~jagadesh.kiran], +1 non-binding. > Wrong help messages in some test-patch plugins > -- > > Key: HADOOP-12399 > URL: https://issues.apache.org/jira/browse/HADOOP-12399 > Project: Hadoop Common > Issue Type: Sub-task > Components: yetus >Affects Versions: HADOOP-12111 >Reporter: Kengo Seki >Assignee: Jagadesh Kiran N >Priority: Minor > Labels: newbie > Attachments: HADOOP-12399.HADOOP-12111.00.patch, > HADOOP-12399.HADOOP-12111.01.patch, HADOOP-12399.HADOOP-12111.02.patch > > > dev-support/personality/bigtop.sh: > {code} > 32 function bigtop_usage > 33 { > 34 echo "Bigtop specific:" > 35 echo "--bigtop-puppetsetup=[false|true] execute the bigtop dev setup > (needs sudo to root)" > 36 } > {code} > s/bigtop-puppetsetup/bigtop-puppet/. > dev-support/test-patch.d/gradle.sh: > {code} > 21 function gradle_usage > 22 { > 23 echo "gradle specific:" > 24 echo "--gradle-cmd=The 'gradle' command to use (default > 'gradle')" > 25 echo "--gradlew-cmd=The 'gradle' command to use (default > 'basedir/gradlew')" > 26 } > {code} > s/'gradle' command/'gradlew' command/ for the latter. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HADOOP-12393) asflicense is easily tricked
[ https://issues.apache.org/jira/browse/HADOOP-12393?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14744485#comment-14744485 ] Kengo Seki commented on HADOOP-12393: - Got it. Thanks! > asflicense is easily tricked > > > Key: HADOOP-12393 > URL: https://issues.apache.org/jira/browse/HADOOP-12393 > Project: Hadoop Common > Issue Type: Sub-task > Components: yetus >Affects Versions: HADOOP-12111 >Reporter: Allen Wittenauer > > asflicense needs to make sure that it gets at least one report file instead > of assuming nothing is wrong. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HADOOP-12393) asflicense is easily tricked
[ https://issues.apache.org/jira/browse/HADOOP-12393?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14744468#comment-14744468 ] Kengo Seki commented on HADOOP-12393: - Let me confirm, my understanding is: * asflicense should fail for Kafka, because its {{gradle rat}} generates only rat-report.html and rat-report.xml for now. * But because asflicense judges from only ant/mvn/gradle's status, it succeeds even though rat-report.txt does not exist. * So we should fix the plugin to check at least one of the assumed output exists. Is this what you meant? > asflicense is easily tricked > > > Key: HADOOP-12393 > URL: https://issues.apache.org/jira/browse/HADOOP-12393 > Project: Hadoop Common > Issue Type: Sub-task > Components: yetus >Affects Versions: HADOOP-12111 >Reporter: Allen Wittenauer > > asflicense needs to make sure that it gets at least one report file instead > of assuming nothing is wrong. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HADOOP-12401) Wrong function names in test-patch bigtop personality
[ https://issues.apache.org/jira/browse/HADOOP-12401?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14742592#comment-14742592 ] Kengo Seki commented on HADOOP-12401: - It's enough for {{gradlew toolchain}} to be called only once and it should be called before other gradle commands. So _precheck is suitable, I think. > Wrong function names in test-patch bigtop personality > - > > Key: HADOOP-12401 > URL: https://issues.apache.org/jira/browse/HADOOP-12401 > Project: Hadoop Common > Issue Type: Sub-task > Components: yetus >Affects Versions: HADOOP-12111 >Reporter: Kengo Seki > > In dev-support/personality/bigtop.sh: > {code} > 51 function bigtop_precheck_postinstall > 52 { > 53 if [[ ${BIGTOP_PUPPETSETUP} = "true" ]]; then > 54 pushd "${BASEDIR}" >/dev/null > 55 echo_and_redirect "${PATCH_DIR}/bigtop-branch-toolchain.txt" > "${GRADLEW}" toolchain > 56 popd >/dev/null > 57 fi > 58 } > 59 > 60 function bigtop_postapply_postinstall > 61 { > 62 if [[ ${BIGTOP_PUPPETSETUP} = "true" ]]; then > 63 pushd "${BASEDIR}" >/dev/null > 64 echo_and_redirect "${PATCH_DIR}/bigtop-patch-toolchain.txt" > "${GRADLEW}" toolchain > 65 popd >/dev/null > 66 fi > 67 } > {code} > Their names are not proper for test-patch plugin callback functions. Maybe it > should be like: > {code} > function bigtop_precompile > { > declare codebase=$1 > if [[ ${BIGTOP_PUPPETSETUP} = "true" && ( ${codebase} = "branch" || > ${codebase} = "patch" ) ]]; then > pushd "${BASEDIR}" >/dev/null > echo_and_redirect "${PATCH_DIR}/bigtop-${codebase}-toolchain.txt" > "${GRADLEW}" toolchain > popd >/dev/null > fi > } > {code} -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HADOOP-12399) Wrong help messages in some test-patch plugins
[ https://issues.apache.org/jira/browse/HADOOP-12399?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14742560#comment-14742560 ] Kengo Seki commented on HADOOP-12399: - Not BIGTOP_PUPPETSETUP, but bigtop-puppet. Just removing "setup" from the original code is enough. > Wrong help messages in some test-patch plugins > -- > > Key: HADOOP-12399 > URL: https://issues.apache.org/jira/browse/HADOOP-12399 > Project: Hadoop Common > Issue Type: Sub-task > Components: yetus >Affects Versions: HADOOP-12111 >Reporter: Kengo Seki >Assignee: Jagadesh Kiran N >Priority: Minor > Labels: newbie > Attachments: HADOOP-12399.HADOOP-12111.00.patch, > HADOOP-12399.HADOOP-12111.01.patch > > > dev-support/personality/bigtop.sh: > {code} > 32 function bigtop_usage > 33 { > 34 echo "Bigtop specific:" > 35 echo "--bigtop-puppetsetup=[false|true] execute the bigtop dev setup > (needs sudo to root)" > 36 } > {code} > s/bigtop-puppetsetup/bigtop-puppet/. > dev-support/test-patch.d/gradle.sh: > {code} > 21 function gradle_usage > 22 { > 23 echo "gradle specific:" > 24 echo "--gradle-cmd=The 'gradle' command to use (default > 'gradle')" > 25 echo "--gradlew-cmd=The 'gradle' command to use (default > 'basedir/gradlew')" > 26 } > {code} > s/'gradle' command/'gradlew' command/ for the latter. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Resolved] (HADOOP-12395) big top personality should skip asflicense
[ https://issues.apache.org/jira/browse/HADOOP-12395?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Kengo Seki resolved HADOOP-12395. - Resolution: Won't Fix Closing this because BIGTOP-2020 has been merged and {{gradlew rat}} works now. > big top personality should skip asflicense > -- > > Key: HADOOP-12395 > URL: https://issues.apache.org/jira/browse/HADOOP-12395 > Project: Hadoop Common > Issue Type: Sub-task > Components: yetus >Affects Versions: HADOOP-12111 >Reporter: Allen Wittenauer > > Bigtop doesn't support RAT checks as part of gradle, so it should probably > skip them for now. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Created] (HADOOP-12397) Incomplete comment for test-patch compile_cycle function
Kengo Seki created HADOOP-12397: --- Summary: Incomplete comment for test-patch compile_cycle function Key: HADOOP-12397 URL: https://issues.apache.org/jira/browse/HADOOP-12397 Project: Hadoop Common Issue Type: Sub-task Components: yetus Affects Versions: HADOOP-12111 Reporter: Kengo Seki Priority: Trivial Its comment says: {code} ## @description This will callout to _precompile, compile, and _postcompile {code} but it calls _rebuild also. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HADOOP-12398) filefilter function in test-patch flink personality is never called
[ https://issues.apache.org/jira/browse/HADOOP-12398?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Kengo Seki updated HADOOP-12398: Summary: filefilter function in test-patch flink personality is never called (was: filefilter function is never called in test-patch flink personality) > filefilter function in test-patch flink personality is never called > --- > > Key: HADOOP-12398 > URL: https://issues.apache.org/jira/browse/HADOOP-12398 > Project: Hadoop Common > Issue Type: Sub-task > Components: yetus >Affects Versions: HADOOP-12111 >Reporter: Kengo Seki > > Wrong function name. > {code} > 28 function fliblib_filefilter > 29 { > 30 local filename=$1 > 31 > 32 if [[ ${filename} =~ \.java$ > 33 || ${filename} =~ \.scala$ > 34 || ${filename} =~ pom.xml$ ]]; then > 35 add_test flinklib > 36 fi > 37 } > {code} -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Created] (HADOOP-12398) filefilter function is never called in test-patch flink personality
Kengo Seki created HADOOP-12398: --- Summary: filefilter function is never called in test-patch flink personality Key: HADOOP-12398 URL: https://issues.apache.org/jira/browse/HADOOP-12398 Project: Hadoop Common Issue Type: Sub-task Components: yetus Affects Versions: HADOOP-12111 Reporter: Kengo Seki Wrong function name. {code} 28 function fliblib_filefilter 29 { 30 local filename=$1 31 32 if [[ ${filename} =~ \.java$ 33 || ${filename} =~ \.scala$ 34 || ${filename} =~ pom.xml$ ]]; then 35 add_test flinklib 36 fi 37 } {code} -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Created] (HADOOP-12399) Wrong help messages in some test-patch plugins
Kengo Seki created HADOOP-12399: --- Summary: Wrong help messages in some test-patch plugins Key: HADOOP-12399 URL: https://issues.apache.org/jira/browse/HADOOP-12399 Project: Hadoop Common Issue Type: Sub-task Components: yetus Affects Versions: HADOOP-12111 Reporter: Kengo Seki Priority: Minor dev-support/personality/bigtop.sh: {code} 32 function bigtop_usage 33 { 34 echo "Bigtop specific:" 35 echo "--bigtop-puppetsetup=[false|true] execute the bigtop dev setup (needs sudo to root)" 36 } {code} s/bigtop-puppetsetup/bigtop-puppet/. dev-support/test-patch.d/gradle.sh: {code} 21 function gradle_usage 22 { 23 echo "gradle specific:" 24 echo "--gradle-cmd=The 'gradle' command to use (default 'gradle')" 25 echo "--gradlew-cmd=The 'gradle' command to use (default 'basedir/gradlew')" 26 } {code} s/'gradle' command/'gradlew' command/ for the latter. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Created] (HADOOP-12400) Wrong comment for scaladoc_rebuild function in test-patch scala plugin
Kengo Seki created HADOOP-12400: --- Summary: Wrong comment for scaladoc_rebuild function in test-patch scala plugin Key: HADOOP-12400 URL: https://issues.apache.org/jira/browse/HADOOP-12400 Project: Hadoop Common Issue Type: Sub-task Components: yetus Affects Versions: HADOOP-12111 Reporter: Kengo Seki Priority: Trivial {code} 62 ## @description Count and compare the number of JavaDoc warnings pre- and post- patch 63 ## @audience private 64 ## @stabilityevolving 65 ## @replaceable no 66 ## @return 0 on success 67 ## @return 1 on failure 68 function scaladoc_rebuild {code} s/JavaDoc/ScalaDoc/ -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HADOOP-12397) Incomplete comment for test-patch compile_cycle function
[ https://issues.apache.org/jira/browse/HADOOP-12397?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Kengo Seki updated HADOOP-12397: Labels: newbie (was: ) > Incomplete comment for test-patch compile_cycle function > > > Key: HADOOP-12397 > URL: https://issues.apache.org/jira/browse/HADOOP-12397 > Project: Hadoop Common > Issue Type: Sub-task > Components: yetus >Affects Versions: HADOOP-12111 >Reporter: Kengo Seki >Priority: Trivial > Labels: newbie > > Its comment says: > {code} > ## @description This will callout to _precompile, compile, and _postcompile > {code} > but it calls _rebuild also. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HADOOP-12398) filefilter function in test-patch flink personality is never called
[ https://issues.apache.org/jira/browse/HADOOP-12398?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Kengo Seki updated HADOOP-12398: Labels: newbie (was: ) > filefilter function in test-patch flink personality is never called > --- > > Key: HADOOP-12398 > URL: https://issues.apache.org/jira/browse/HADOOP-12398 > Project: Hadoop Common > Issue Type: Sub-task > Components: yetus >Affects Versions: HADOOP-12111 >Reporter: Kengo Seki > Labels: newbie > > Wrong function name. > {code} > 28 function fliblib_filefilter > 29 { > 30 local filename=$1 > 31 > 32 if [[ ${filename} =~ \.java$ > 33 || ${filename} =~ \.scala$ > 34 || ${filename} =~ pom.xml$ ]]; then > 35 add_test flinklib > 36 fi > 37 } > {code} -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HADOOP-12399) Wrong help messages in some test-patch plugins
[ https://issues.apache.org/jira/browse/HADOOP-12399?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Kengo Seki updated HADOOP-12399: Labels: newbie (was: ) > Wrong help messages in some test-patch plugins > -- > > Key: HADOOP-12399 > URL: https://issues.apache.org/jira/browse/HADOOP-12399 > Project: Hadoop Common > Issue Type: Sub-task > Components: yetus >Affects Versions: HADOOP-12111 >Reporter: Kengo Seki >Priority: Minor > Labels: newbie > > dev-support/personality/bigtop.sh: > {code} > 32 function bigtop_usage > 33 { > 34 echo "Bigtop specific:" > 35 echo "--bigtop-puppetsetup=[false|true] execute the bigtop dev setup > (needs sudo to root)" > 36 } > {code} > s/bigtop-puppetsetup/bigtop-puppet/. > dev-support/test-patch.d/gradle.sh: > {code} > 21 function gradle_usage > 22 { > 23 echo "gradle specific:" > 24 echo "--gradle-cmd=The 'gradle' command to use (default > 'gradle')" > 25 echo "--gradlew-cmd=The 'gradle' command to use (default > 'basedir/gradlew')" > 26 } > {code} > s/'gradle' command/'gradlew' command/ for the latter. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Created] (HADOOP-12401) Wrong function names in test-patch bigtop personality
Kengo Seki created HADOOP-12401: --- Summary: Wrong function names in test-patch bigtop personality Key: HADOOP-12401 URL: https://issues.apache.org/jira/browse/HADOOP-12401 Project: Hadoop Common Issue Type: Sub-task Components: yetus Affects Versions: HADOOP-12111 Reporter: Kengo Seki In dev-support/personality/bigtop.sh: {code} 51 function bigtop_precheck_postinstall 52 { 53 if [[ ${BIGTOP_PUPPETSETUP} = "true" ]]; then 54 pushd "${BASEDIR}" >/dev/null 55 echo_and_redirect "${PATCH_DIR}/bigtop-branch-toolchain.txt" "${GRADLEW}" toolchain 56 popd >/dev/null 57 fi 58 } 59 60 function bigtop_postapply_postinstall 61 { 62 if [[ ${BIGTOP_PUPPETSETUP} = "true" ]]; then 63 pushd "${BASEDIR}" >/dev/null 64 echo_and_redirect "${PATCH_DIR}/bigtop-patch-toolchain.txt" "${GRADLEW}" toolchain 65 popd >/dev/null 66 fi 67 } {code} Their names are not proper for test-patch plugin callback functions. Maybe it should be like: {code} function bigtop_precompile { declare codebase=$1 if [[ ${BIGTOP_PUPPETSETUP} = "true" && ( ${codebase} = "branch" || ${codebase} = "patch" ) ]]; then pushd "${BASEDIR}" >/dev/null echo_and_redirect "${PATCH_DIR}/bigtop-${codebase}-toolchain.txt" "${GRADLEW}" toolchain popd >/dev/null fi } {code} -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HADOOP-12397) Incomplete comment for test-patch compile_cycle function
[ https://issues.apache.org/jira/browse/HADOOP-12397?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14739941#comment-14739941 ] Kengo Seki commented on HADOOP-12397: - Thanks [~jagadesh.kiran], could you change the enumeration order in the comment to reflect the actual invocation order? > Incomplete comment for test-patch compile_cycle function > > > Key: HADOOP-12397 > URL: https://issues.apache.org/jira/browse/HADOOP-12397 > Project: Hadoop Common > Issue Type: Sub-task > Components: yetus >Affects Versions: HADOOP-12111 >Reporter: Kengo Seki >Assignee: Jagadesh Kiran N >Priority: Trivial > Labels: newbie > Attachments: HADOOP-12397.HADOOP-12111.00.patch > > > Its comment says: > {code} > ## @description This will callout to _precompile, compile, and _postcompile > {code} > but it calls _rebuild also. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HADOOP-12397) Incomplete comment for test-patch compile_cycle function
[ https://issues.apache.org/jira/browse/HADOOP-12397?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14739935#comment-14739935 ] Kengo Seki commented on HADOOP-12397: - Allen, the architecture and advanced document seem to mention the _rebuild phase already. Do you have something to worry about? > Incomplete comment for test-patch compile_cycle function > > > Key: HADOOP-12397 > URL: https://issues.apache.org/jira/browse/HADOOP-12397 > Project: Hadoop Common > Issue Type: Sub-task > Components: yetus >Affects Versions: HADOOP-12111 >Reporter: Kengo Seki >Assignee: Jagadesh Kiran N >Priority: Trivial > Labels: newbie > Attachments: HADOOP-12397.HADOOP-12111.00.patch > > > Its comment says: > {code} > ## @description This will callout to _precompile, compile, and _postcompile > {code} > but it calls _rebuild also. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HADOOP-12400) Wrong comment for scaladoc_rebuild function in test-patch scala plugin
[ https://issues.apache.org/jira/browse/HADOOP-12400?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14739966#comment-14739966 ] Kengo Seki commented on HADOOP-12400: - Thanks [~jagadesh.kiran], but I think it's enough to simply replace "JavaDoc" with "ScalaDoc" in the original comment. Because "_rebuild" is just a hook name and the new comment "Rebuild scala doc" is something inaccurate. This function works as the original comment says, via calling generic_pre_handler and generic_post_handler. > Wrong comment for scaladoc_rebuild function in test-patch scala plugin > -- > > Key: HADOOP-12400 > URL: https://issues.apache.org/jira/browse/HADOOP-12400 > Project: Hadoop Common > Issue Type: Sub-task > Components: yetus >Affects Versions: HADOOP-12111 >Reporter: Kengo Seki >Assignee: Jagadesh Kiran N >Priority: Trivial > Labels: newbie > Attachments: HADOOP-12400.HADOOP-12111.00.patch, > HADOOP-12400.HADOOP-12111.01.patch > > > {code} > 62 ## @description Count and compare the number of JavaDoc warnings pre- > and post- patch > 63 ## @audience private > 64 ## @stabilityevolving > 65 ## @replaceable no > 66 ## @return 0 on success > 67 ## @return 1 on failure > 68 function scaladoc_rebuild > {code} > s/JavaDoc/ScalaDoc/ -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HADOOP-12398) filefilter function in test-patch flink personality is never called
[ https://issues.apache.org/jira/browse/HADOOP-12398?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14739919#comment-14739919 ] Kengo Seki commented on HADOOP-12398: - [~jagadesh.kiran] Yes, that function is intended to be called from test-patch.sh. But because of its name, it won't be called. [As yetus precommit document explains|https://github.com/apache/hadoop/blob/HADOOP-12111/dev-support/docs/precommit-advanced.md#test-plug-ins], plugin name is passed to test-patch.sh via add_plugin function. In this case, the plugin name is "flinklib": {code} 26 add_plugin flinklib {code} so the function name should be "flinklib_filefilter". Could you rename the function rather than remove it? Thanks. > filefilter function in test-patch flink personality is never called > --- > > Key: HADOOP-12398 > URL: https://issues.apache.org/jira/browse/HADOOP-12398 > Project: Hadoop Common > Issue Type: Sub-task > Components: yetus >Affects Versions: HADOOP-12111 >Reporter: Kengo Seki >Assignee: Jagadesh Kiran N > Labels: newbie > Attachments: HADOOP-12398.HADOOP-12111.00.patch > > > Wrong function name. > {code} > 28 function fliblib_filefilter > 29 { > 30 local filename=$1 > 31 > 32 if [[ ${filename} =~ \.java$ > 33 || ${filename} =~ \.scala$ > 34 || ${filename} =~ pom.xml$ ]]; then > 35 add_test flinklib > 36 fi > 37 } > {code} -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Assigned] (HADOOP-12111) [Umbrella] Split test-patch off into its own TLP
[ https://issues.apache.org/jira/browse/HADOOP-12111?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Kengo Seki reassigned HADOOP-12111: --- Assignee: Kengo Seki > [Umbrella] Split test-patch off into its own TLP > > > Key: HADOOP-12111 > URL: https://issues.apache.org/jira/browse/HADOOP-12111 > Project: Hadoop Common > Issue Type: New Feature > Components: yetus >Affects Versions: HADOOP-12111 >Reporter: Allen Wittenauer >Assignee: Kengo Seki > > Given test-patch's tendency to get forked into a variety of different > projects, it makes a lot of sense to make an Apache TLP so that everyone can > benefit from a common code base. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HADOOP-12111) [Umbrella] Split test-patch off into its own TLP
[ https://issues.apache.org/jira/browse/HADOOP-12111?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Kengo Seki updated HADOOP-12111: Assignee: (was: Kengo Seki) > [Umbrella] Split test-patch off into its own TLP > > > Key: HADOOP-12111 > URL: https://issues.apache.org/jira/browse/HADOOP-12111 > Project: Hadoop Common > Issue Type: New Feature > Components: yetus >Affects Versions: HADOOP-12111 >Reporter: Allen Wittenauer > > Given test-patch's tendency to get forked into a variety of different > projects, it makes a lot of sense to make an Apache TLP so that everyone can > benefit from a common code base. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HADOOP-12257) rework build tool support; add gradle; add scala
[ https://issues.apache.org/jira/browse/HADOOP-12257?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14736852#comment-14736852 ] Kengo Seki commented on HADOOP-12257: - In addition to Sean's comments, I ran the updated test-patch for all 9 projects currently supported, and confirmed the results correctly reflect the problems which the patches have. Here's some examples. BigTop: {code} | Vote | Subsystem | Runtime | Comment | +1 | @author | 0m 00s| The patch does not contain any @author | ||| tags. | -1 |test4tests | 0m 00s| The patch doesn't appear to include any | ||| new or modified tests. Please justify why | ||| no new tests are needed for this patch. | ||| Also please list what manual steps were | ||| performed to verify this patch. | +1 |gradleboot | 0m 16s| master passed | +1 | compile | 0m 09s| master passed | -1 | scaladoc | 0m 09s| root in master failed. | +1 |gradleboot | 0m 14s| the patch passed | +1 | compile | 0m 10s| the patch passed | +1 |scalac | 0m 10s| the patch passed | +1 |whitespace | 0m 00s| Patch has no whitespace issues. | -1 | scaladoc | 0m 10s| root in the patch failed. | -1 |asflicense | 0m 10s| root in the patch failed. | || 1m 47s| || Subsystem || Report/Notes || | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12754699/BIGTOP-2019.1.patch | | JIRA Issue | BIGTOP-2019 | | git revision | master / 5efdbfb | | Optional Tests | asflicense scalac scaladoc unit compile | | uname | Darwin mobile.local 14.5.0 Darwin Kernel Version 14.5.0: Wed Jul 29 02:26:53 PDT 2015; root:xnu-2782.40.9~1/RELEASE_X86_64 x86_64 | | Build tool | gradle | | Personality | /Users/sekikn/hadoop/dev-support/personality/bigtop.sh | | Default Java | 1.7.0_80 | | scaladoc | /private/tmp/test-patch-bigtop/19287/branch-scaladoc-root.txt | | scaladoc | /private/tmp/test-patch-bigtop/19287/patch-scaladoc-root.txt | | asflicense | /private/tmp/test-patch-bigtop/19287/patch-asflicense-root.txt | {code} Kafka: failed to compile {code} | Vote | Subsystem | Runtime | Comment | +1 | @author | 0m 00s| The patch does not contain any @author | ||| tags. | +1 |test4tests | 0m 00s| The patch appears to include 11 new or | ||| modified test files. | +1 |gradleboot | 3m 47s| trunk passed | +1 | compile | 20m 12s | trunk passed | +1 | javadoc | 3m 37s| trunk passed | +1 | scaladoc | 10m 01s | trunk passed | +1 |gradleboot | 1m 55s| the patch passed | -1 | compile | 10m 53s | root in the patch failed. | -1 | javac | 10m 53s | root in the patch failed. | -1 |scalac | 10m 53s | root in the patch failed. | +1 |whitespace | 0m 01s| Patch has no whitespace issues. | -1 | javadoc | 1m 09s| root in the patch failed. | +1 | scaladoc | 2m 45s| the patch passed | +1 |asflicense | 0m 55s| Patch does not generate ASF License | ||| warnings. | || 62m 11s | || Subsystem || Report/Notes || | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12754305/KAFKA-2120_2015-09-04_17%3A49%3A01.patch | | JIRA Issue | KAFKA-2120 | | git revision | trunk / b8b1bca | | Optional Tests | asflicense javac javadoc unit findbugs compile scalac scaladoc | | uname | Darwin mobile.local 14.5.0 Darwin Kernel Version 14.5.0: Wed Jul 29 02:26:53 PDT 2015; root:xnu-2782.40.9~1/RELEASE_X86_64 x86_64 | | Build tool | gradle | | Personality | /Users/sekikn/hadoop/dev-support/personality/kafka.sh | | Default Java | 1.7.0_80 | | compile | /private/tmp/test-patch-kafka/38835/patch-compile-root.txt | | javac | /private/tmp/test-patch-kafka/38835/patch-compile-root.txt | | scalac | /private/tmp/test-patch-kafka/38835/patch-compile-root.txt | | findbugs | not supported by the gradle plugin | | javadoc | /private/tmp/test-patch-kafka/38835/patch-javadoc-root.txt | {code} Samza: succeeded {code} | Vote | Subsystem | Runtime | Comment | +1 | @author | 0m 00s| The patch does not contain
[jira] [Updated] (HADOOP-12233) if CHANGED_FILES is corrupt, find_changed_modules never returns
[ https://issues.apache.org/jira/browse/HADOOP-12233?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Kengo Seki updated HADOOP-12233: Attachment: HADOOP-12233.HADOOP-12111.02.patch We maybe overthought it. Is the 02.patch enough to solve this problem? > if CHANGED_FILES is corrupt, find_changed_modules never returns > --- > > Key: HADOOP-12233 > URL: https://issues.apache.org/jira/browse/HADOOP-12233 > Project: Hadoop Common > Issue Type: Sub-task > Components: yetus >Affects Versions: HADOOP-12111 >Reporter: Allen Wittenauer >Assignee: Kengo Seki > Fix For: HADOOP-12111 > > Attachments: HADOOP-12233.HADOOP-12111.00.patch, > HADOOP-12233.HADOOP-12111.01.patch, HADOOP-12233.HADOOP-12111.02.patch > > > In building some unit tests, did a negative tests and hit this condition. We > should put a limit on how many times we loop in the find_x_dirs code. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HADOOP-12233) if CHANGED_FILES is corrupt, find_changed_modules never returns
[ https://issues.apache.org/jira/browse/HADOOP-12233?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Kengo Seki updated HADOOP-12233: Status: Patch Available (was: Reopened) > if CHANGED_FILES is corrupt, find_changed_modules never returns > --- > > Key: HADOOP-12233 > URL: https://issues.apache.org/jira/browse/HADOOP-12233 > Project: Hadoop Common > Issue Type: Sub-task > Components: yetus >Affects Versions: HADOOP-12111 >Reporter: Allen Wittenauer >Assignee: Kengo Seki > Fix For: HADOOP-12111 > > Attachments: HADOOP-12233.HADOOP-12111.00.patch, > HADOOP-12233.HADOOP-12111.01.patch, HADOOP-12233.HADOOP-12111.02.patch > > > In building some unit tests, did a negative tests and hit this condition. We > should put a limit on how many times we loop in the find_x_dirs code. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Created] (HADOOP-12383) Document python version releasedocmaker supports
Kengo Seki created HADOOP-12383: --- Summary: Document python version releasedocmaker supports Key: HADOOP-12383 URL: https://issues.apache.org/jira/browse/HADOOP-12383 Project: Hadoop Common Issue Type: Sub-task Components: yetus Affects Versions: HADOOP-12111 Reporter: Kengo Seki Priority: Minor Maybe 2.5+? -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HADOOP-12380) Wrong grep command invocation in github_find_jira_title
[ https://issues.apache.org/jira/browse/HADOOP-12380?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14733769#comment-14733769 ] Kengo Seki commented on HADOOP-12380: - Thanks [~brahmareddy], +1 non-binding. {code} [sekikn@mobile hadoop]$ bash -x dev-support/test-patch.sh --basedir=../pig --project=pig --grep-cmd=/usr/bin/grep 18 (snip) + github_find_jira_title + declare title + declare maybe + declare retval + [[ ! -f /private/tmp/test-patch-pig/2968/github-pull.json ]] ++ /usr/bin/grep title /private/tmp/test-patch-pig/2968/github-pull.json ++ cut -f4 '-d"' + title=PIG-4373 (snip) {code} > Wrong grep command invocation in github_find_jira_title > --- > > Key: HADOOP-12380 > URL: https://issues.apache.org/jira/browse/HADOOP-12380 > Project: Hadoop Common > Issue Type: Sub-task > Components: yetus >Affects Versions: HADOOP-12111 >Reporter: Kengo Seki >Assignee: Brahma Reddy Battula > Labels: newbie > Attachments: HADOOP-12380.HADOOP-12111.01.patch > > > {code} > 139 title=$(GREP title "${PATCH_DIR}/github-pull.json" \ > 140 | cut -f4 -d\") > {code} > should be {{title=$($\{GREP} ...)}} -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HADOOP-12277) releasedocmaker index mode should create a readme.md in addition to a index.md
[ https://issues.apache.org/jira/browse/HADOOP-12277?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14733816#comment-14733816 ] Kengo Seki commented on HADOOP-12277: - Small nits: * 02.patch introduces 4 new pylint warnings. The first two should be fixed. I think the last two are ignorable because the source looks simple enough. {code} [sekikn@mobile hadoop]$ cat /private/tmp/test-patch-yetus/10595/diff-patch-pylint.txt dev-support/releasedocmaker.py:106: [C0326(bad-whitespace), ] Exactly one space required after comma versions.sort(key=LooseVersion,reverse=True) ^ dev-support/releasedocmaker.py:119: [C0326(bad-whitespace), ] Exactly one space required after comma versions.sort(key=LooseVersion,reverse=True) ^ dev-support/releasedocmaker.py:117: [C0111(missing-docstring), buildreadme] Missing function docstring dev-support/releasedocmaker.py:395: [R0915(too-many-statements), main] Too many statements (146/50) {code} * {{indexfile.close()}} after with statement is probably omittable. {code} 104 def buildindex(title, asf_license): 105 versions = glob("[0-9]*.[0-9]*.[0-9]*") 106 versions.sort(key=LooseVersion,reverse=True) 107 with open("index.md", "w") as indexfile: 108 if asf_license is True: 109 indexfile.write(ASF_LICENSE) 110 for version in versions: 111 indexfile.write("* %s v%s\n" % (title, version)) 112 for k in ("Changes", "Release Notes"): 113 indexfile.write("* [%s](%s/%s.%s.html)\n" \ 114 % (k, version, k.upper().replace(" ", ""), version)) 115 indexfile.close() 116 117 def buildreadme(title, asf_license): 118 versions = glob("[0-9]*.[0-9]*.[0-9]*") 119 versions.sort(key=LooseVersion,reverse=True) 120 with open("README.md", "w") as indexfile: 121 if asf_license is True: 122 indexfile.write(ASF_LICENSE) 123 for version in versions: 124 indexfile.write("* %s v%s\n" % (title, version)) 125 for k in ("Changes", "Release Notes"): 126 indexfile.write("* [%s](%s/%s.%s.md)\n" \ 127 % (k, version, k.upper().replace(" ", ""), version)) 128 indexfile.close() {code} Otherwise, +1 non-binding. > releasedocmaker index mode should create a readme.md in addition to a index.md > -- > > Key: HADOOP-12277 > URL: https://issues.apache.org/jira/browse/HADOOP-12277 > Project: Hadoop Common > Issue Type: Sub-task > Components: yetus >Affects Versions: HADOOP-12111 >Reporter: Allen Wittenauer >Assignee: Allen Wittenauer >Priority: Minor > Attachments: HADOOP-12277.HADOOP-12111.00.patch, > HADOOP-12277.HADOOP-12111.01.patch, HADOOP-12277.HADOOP-12111.02.patch > > > The content should be the same, however, rather than a mvn site-compatible > links, it should be github-compatible markdown links. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HADOOP-12341) patch file confuses test-patch (date format problems)
[ https://issues.apache.org/jira/browse/HADOOP-12341?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14734248#comment-14734248 ] Kengo Seki commented on HADOOP-12341: - bq. Thus that code was removed. Make sense. bq. Let's effectivelymerge smart-apply-patch and test-patch. Sounds good! By doing so, we can commonize some functions handling a patch and remove the dependency on smart-apply-patch from test-patch. But in my understanding, this change itself doesn't solve the original problem directly, is this right? > patch file confuses test-patch (date format problems) > - > > Key: HADOOP-12341 > URL: https://issues.apache.org/jira/browse/HADOOP-12341 > Project: Hadoop Common > Issue Type: Sub-task > Components: yetus >Affects Versions: HADOOP-12111 >Reporter: Allen Wittenauer > Attachments: HADOOP-12326.002.patch, > HADOOP-12375.HADOOP-12111.02.patch > > > This was attached to HADOOP-12326 . -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HADOOP-12341) patch file confuses test-patch (date format problems)
[ https://issues.apache.org/jira/browse/HADOOP-12341?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14733238#comment-14733238 ] Kengo Seki commented on HADOOP-12341: - To resolve this issue, I have some questions: * test-patch had skipped file type guessing if the file has .patch suffix, but this behaviour is changed on HADOOP-12129. Is there any concerns if we return this behaviour back? * Or, is there any concerns if we match the following second and third patterns against whole lines in the file? {code} 1662 fileOutput=$(head -n 1 "${patch}" | "${GREP}" -E "^(From [a-z0-9]* Mon Sep 17 00:00:00 2001)|(diff .*)|(Index: .*)$") {code} * Sean said: bq. we detect git patches made with format-patch, but we only ever use git apply, rather than the git am needed in that case. but I tried some patches made with format-patch, and {{git apply}} worked on all of them. If we can't find any example immediately, can we postpone the fix for this case? > patch file confuses test-patch (date format problems) > - > > Key: HADOOP-12341 > URL: https://issues.apache.org/jira/browse/HADOOP-12341 > Project: Hadoop Common > Issue Type: Sub-task > Components: yetus >Affects Versions: HADOOP-12111 >Reporter: Allen Wittenauer > Attachments: HADOOP-12326.002.patch, > HADOOP-12375.HADOOP-12111.02.patch > > > This was attached to HADOOP-12326 . -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HADOOP-4258) the test patch script should check for filenames that differ only in case
[ https://issues.apache.org/jira/browse/HADOOP-4258?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14732415#comment-14732415 ] Kengo Seki commented on HADOOP-4258: Thanks [~jagadesh.kiran], some comments: * K and L are not so readable too. How about "$first", "$second" or something? Also, $a and $b should be changed, like "$first_lower", "$second_lower". * Address other Allen's comments. About test-patch plugin hooks such as postcheckout or preapply, see [yetus precommit document|https://github.com/apache/hadoop/blob/HADOOP-12111/dev-support/docs/precommit-advanced.md#test-plug-ins]. * This code {code} 34 declare -a changed_files 35 i=0 36 # shellcheck disable=SC2153 37 while read line; do 38 changed_files[${i}]="${line}" 39 ((i=i+1)) 40 done < <(echo "${CHANGED_FILES}") {code} is intended to store newline-separated strings in an array, but very verbose (sorry, my fault). Instead, {code} declare -a changed_files changed_files=(${CHANGED_FILES}) {code} should work. * And, please don't forget to shorten the plugin name. ;) > the test patch script should check for filenames that differ only in case > - > > Key: HADOOP-4258 > URL: https://issues.apache.org/jira/browse/HADOOP-4258 > Project: Hadoop Common > Issue Type: Sub-task > Components: build, test >Reporter: Owen O'Malley >Assignee: Jagadesh Kiran N > Labels: test-patch > Attachments: HADOOP-4258.001.patch, HADOOP-4258.HADOOP-12111.00.patch > > > It would be nice if the test patch script warned about filenames that differ > only in case. We recently had a patch committed that had a pair of colliding > filenames and subversion broke badly on my Mac. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Created] (HADOOP-12380) Wrong grep command invocation in github_find_jira_title
Kengo Seki created HADOOP-12380: --- Summary: Wrong grep command invocation in github_find_jira_title Key: HADOOP-12380 URL: https://issues.apache.org/jira/browse/HADOOP-12380 Project: Hadoop Common Issue Type: Sub-task Components: yetus Affects Versions: HADOOP-12111 Reporter: Kengo Seki {code} 139 title=$(GREP title "${PATCH_DIR}/github-pull.json" \ 140 | cut -f4 -d\") {code} should be {{title=$($\{GREP} ...)}} -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HADOOP-12355) test-patch TAP plugin should use ${SED} instead of sed
[ https://issues.apache.org/jira/browse/HADOOP-12355?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14731060#comment-14731060 ] Kengo Seki commented on HADOOP-12355: - LGTM, +1 non-binding. > test-patch TAP plugin should use ${SED} instead of sed > -- > > Key: HADOOP-12355 > URL: https://issues.apache.org/jira/browse/HADOOP-12355 > Project: Hadoop Common > Issue Type: Sub-task > Components: yetus >Affects Versions: HADOOP-12111 >Reporter: Kengo Seki >Assignee: Jagadesh Kiran N >Priority: Trivial > Labels: newbie > Attachments: HADOOP-12355.HADOOP-12111.00.patch > > > for consistency and platform compatibility. > {code} > 54 if [[ -n "${filenames}" ]]; then > 55 module_failed_tests=$(echo "${filenames}" \ > 56 | sed -e "s,${TAP_LOG_DIR},,g" -e s,^/,,g ) > {code} -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HADOOP-12375) Incomplete checking for findbugs executable
[ https://issues.apache.org/jira/browse/HADOOP-12375?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14731040#comment-14731040 ] Kengo Seki commented on HADOOP-12375: - LGTM, +1 non-binding. BTW, my last comment bq. your patch seems to have an unnecessary header. Would you remove the first 3 lines? was a mistake. 00 and 01 are not applicable because a target file is wrong as Allen said, not because of the header. Sorry. > Incomplete checking for findbugs executable > --- > > Key: HADOOP-12375 > URL: https://issues.apache.org/jira/browse/HADOOP-12375 > Project: Hadoop Common > Issue Type: Sub-task > Components: yetus >Affects Versions: HADOOP-12111 >Reporter: Kengo Seki >Assignee: Jagadesh Kiran N > Labels: newbie > Attachments: HADOOP-12375.HADOOP-12111.00.patch, > HADOOP-12375.HADOOP-12111.01.patch, HADOOP-12375.HADOOP-12111.02.patch > > > In test-patch.d/findbugs.sh: > {code} > 65 function findbugs_is_installed > 66 { > 67 if [[ ! -e "${FINDBUGS_HOME}/bin/findbugs" ]]; then > 68 printf "\n\n%s is not executable.\n\n" "${FINDBUGS_HOME}/bin/findbugs" > 69 add_vote_table -1 findbugs "Findbugs is not installed." > 70 return 1 > 71 fi > 72 return 0 > 73 } > {code} > should be -x. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Assigned] (HADOOP-4258) the test patch script should check for filenames that differ only in case
[ https://issues.apache.org/jira/browse/HADOOP-4258?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Kengo Seki reassigned HADOOP-4258: -- Assignee: Jagadesh Kiran N > the test patch script should check for filenames that differ only in case > - > > Key: HADOOP-4258 > URL: https://issues.apache.org/jira/browse/HADOOP-4258 > Project: Hadoop Common > Issue Type: Sub-task > Components: build, test >Reporter: Owen O'Malley >Assignee: Jagadesh Kiran N > Labels: newbie, test-patch > Attachments: HADOOP-4258.001.patch > > > It would be nice if the test patch script warned about filenames that differ > only in case. We recently had a patch committed that had a pair of colliding > filenames and subversion broke badly on my Mac. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HADOOP-12341) patch file confuses test-patch (date format problems)
[ https://issues.apache.org/jira/browse/HADOOP-12341?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Kengo Seki updated HADOOP-12341: Attachment: HADOOP-12375.HADOOP-12111.02.patch Attaching another case. This was originally attached to HADOOP-12375. {{git apply}} works on it, but test-patch fails as follows: {code} /tmp/test-patch-hadoop/2043 has been created Running in developer mode Patch file /Users/sekikn/Desktop/HADOOP-12375.HADOOP-12111.02.patch copied to /private/tmp/test-patch-hadoop/2043 ERROR: Unsure how to process /Users/sekikn/Desktop/HADOOP-12375.HADOOP-12111.02.patch. {code} with bash -x: {code} Patch file /Users/sekikn/Desktop/HADOOP-12375.HADOOP-12111.02.patch copied to /private/tmp/test-patch-hadoop/2071 + guess_patch_file /private/tmp/test-patch-hadoop/2071/patch + local patch=/private/tmp/test-patch-hadoop/2071/patch + local fileOutput + [[ ! -f /private/tmp/test-patch-hadoop/2071/patch ]] + yetus_debug 'Trying to guess is /private/tmp/test-patch-hadoop/2071/patch is a patch file.' + [[ -n '' ]] ++ file /private/tmp/test-patch-hadoop/2071/patch + fileOutput='/private/tmp/test-patch-hadoop/2071/patch: ASCII text' + [[ /private/tmp/test-patch-hadoop/2071/patch: ASCII text =~ diff ]] ++ head -n 1 /private/tmp/test-patch-hadoop/2071/patch ++ grep -E '^(From [a-z0-9]* Mon Sep 17 00:00:00 2001)|(diff .*)|(Index: .*)$' + fileOutput= + [[ 1 == 0 ]] + return 1 + [[ 1 != 0 ]] + yetus_error 'ERROR: Unsure how to process /Users/sekikn/Desktop/HADOOP-12375.HADOOP-12111.02.patch.' + echo 'ERROR: Unsure how to process /Users/sekikn/Desktop/HADOOP-12375.HADOOP-12111.02.patch.' ERROR: Unsure how to process /Users/sekikn/Desktop/HADOOP-12375.HADOOP-12111.02.patch. {code} > patch file confuses test-patch (date format problems) > - > > Key: HADOOP-12341 > URL: https://issues.apache.org/jira/browse/HADOOP-12341 > Project: Hadoop Common > Issue Type: Sub-task > Components: yetus >Affects Versions: HADOOP-12111 >Reporter: Allen Wittenauer > Attachments: HADOOP-12326.002.patch, > HADOOP-12375.HADOOP-12111.02.patch > > > This was attached to HADOOP-12326 . -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HADOOP-12349) Misleading debug message in generic_locate_patch
[ https://issues.apache.org/jira/browse/HADOOP-12349?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14731089#comment-14731089 ] Kengo Seki commented on HADOOP-12349: - Thanks! But I think it's more user-friendly to leave a debug message that indicates the user is in generic_locate_patch and the download is failed, rather than simply removing it. Could you change the message to "generic_locate_patch: failed to download a patch" or something? > Misleading debug message in generic_locate_patch > > > Key: HADOOP-12349 > URL: https://issues.apache.org/jira/browse/HADOOP-12349 > Project: Hadoop Common > Issue Type: Sub-task > Components: yetus >Affects Versions: HADOOP-12111 >Reporter: Kengo Seki >Assignee: Jagadesh Kiran N >Priority: Minor > Labels: newbie > Attachments: HADOOP-12349.HADOOP-12111.00.patch > > > The following message in builtin-bugsystem.sh is duplicated with > jira_locate_patch and misleading. > {code} > yetus_debug "jira_locate_patch: not a JIRA." > {code} -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HADOOP-11897) test-patch.sh plugins should abbreviate the path
[ https://issues.apache.org/jira/browse/HADOOP-11897?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Kengo Seki updated HADOOP-11897: Component/s: yetus > test-patch.sh plugins should abbreviate the path > > > Key: HADOOP-11897 > URL: https://issues.apache.org/jira/browse/HADOOP-11897 > Project: Hadoop Common > Issue Type: Sub-task > Components: test, yetus >Affects Versions: HADOOP-12111 >Reporter: Allen Wittenauer >Priority: Trivial > Labels: newbie > > The current output of checkstyle, shellcheck, and whitespace results in very > long file name paths. It might be useful to abbreviate them in some way, > maybe removing the entire path and leaving just the filename or last > directory+filename. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HADOOP-4258) the test patch script should check for filenames that differ only in case
[ https://issues.apache.org/jira/browse/HADOOP-4258?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14731124#comment-14731124 ] Kengo Seki commented on HADOOP-4258: Thanks [~jagadesh.kiran], that's very helpful! When you update the patch, could you change its plugin name to something like "dupname"? The current name "duplicatednames" is too long as a plugin name. > the test patch script should check for filenames that differ only in case > - > > Key: HADOOP-4258 > URL: https://issues.apache.org/jira/browse/HADOOP-4258 > Project: Hadoop Common > Issue Type: Sub-task > Components: build, test >Reporter: Owen O'Malley > Labels: newbie, test-patch > Attachments: HADOOP-4258.001.patch > > > It would be nice if the test patch script warned about filenames that differ > only in case. We recently had a patch committed that had a pair of colliding > filenames and subversion broke badly on my Mac. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HADOOP-12349) Misleading debug message in generic_locate_patch
[ https://issues.apache.org/jira/browse/HADOOP-12349?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14731296#comment-14731296 ] Kengo Seki commented on HADOOP-12349: - Thanks, +1 non-binding. > Misleading debug message in generic_locate_patch > > > Key: HADOOP-12349 > URL: https://issues.apache.org/jira/browse/HADOOP-12349 > Project: Hadoop Common > Issue Type: Sub-task > Components: yetus >Affects Versions: HADOOP-12111 >Reporter: Kengo Seki >Assignee: Jagadesh Kiran N >Priority: Minor > Labels: newbie > Attachments: HADOOP-12349.HADOOP-12111.00.patch, > HADOOP-12349.HADOOP-12111.01.patch > > > The following message in builtin-bugsystem.sh is duplicated with > jira_locate_patch and misleading. > {code} > yetus_debug "jira_locate_patch: not a JIRA." > {code} -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HADOOP-11897) test-patch.sh plugins should abbreviate the path
[ https://issues.apache.org/jira/browse/HADOOP-11897?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14731263#comment-14731263 ] Kengo Seki commented on HADOOP-11897: - This may be fixed in the latest yetus code, but I found a strange output. {code} $ git checkout -b HADOOP-12111 origin/HADOOP-12111 # or simply `git checkout HADOOP-12111` if you've already fetched the remote branch $ dev-support/test-patch.sh --basedir=/path/to/another/hadoop/repo --project=hadoop https://issues.apache.org/jira/secure/attachment/12731148/HADOOP-7266.001.patch {code} will produce the following result (I skipped all mvn running with Ctrl-C): {code} || Subsystem || Report/Notes || | git revision | trunk / d16c4ee | | Optional Tests | asflicense javac javadoc mvninstall unit findbugs checkstyle | | uname | Darwin mobile.local 14.5.0 Darwin Kernel Version 14.5.0: Wed Jul 29 02:26:53 PDT 2015; root:xnu-2782.40.9~1/RELEASE_X86_64 x86_64 | | Build tool | maven | | Personality | /Users/sekikn/work/hadoop/dev-support/personality/hadoop.sh | | Default Java | 1.7.0_80 | | mvninstall | /private/tmp/test-patch-hadoop/61296/branch-mvninstall-root.txt | | javac | /private/tmp/test-patch-hadoop/61296/branch-javac-root.txt | | javadoc | /private/tmp/test-patch-hadoop/61296/branch-javadoc-hadoop-common-project_hadoop-common.txt | | javadoc | /private/tmp/test-patch-hadoop/61296/branch-javadoc-hadoop-tools_hadoop-streaming.txt | | checkstyle | /private/tmp/test-patch-hadoop/61296//private/tmp/test-patch-hadoop/61296/maven-branch-checkstyle-root.txt | | javac | /private/tmp/test-patch-hadoop/61296/patch-javac-root.txt | | asflicense | /private/tmp/test-patch-hadoop/61296/patch-asflicense-root.txt | | checkstyle | /private/tmp/test-patch-hadoop/61296//private/tmp/test-patch-hadoop/61296/maven-patch-checkstyle-root.txt | | mvninstall | /private/tmp/test-patch-hadoop/61296/patch-mvninstall-hadoop-common-project_hadoop-common.txt | | mvninstall | /private/tmp/test-patch-hadoop/61296/patch-mvninstall-hadoop-tools_hadoop-streaming.txt | | javadoc | /private/tmp/test-patch-hadoop/61296/patch-javadoc-hadoop-common-project_hadoop-common.txt | | javadoc | /private/tmp/test-patch-hadoop/61296/patch-javadoc-hadoop-tools_hadoop-streaming.txt | | eclipse | /private/tmp/test-patch-hadoop/61296/patch-eclipse-hadoop-common-project_hadoop-common.txt | | eclipse | /private/tmp/test-patch-hadoop/61296/patch-eclipse-hadoop-tools_hadoop-streaming.txt | | findbugs | /private/tmp/test-patch-hadoop/61296/patch-findbugs-hadoop-common-project_hadoop-common.txt | | findbugs | /private/tmp/test-patch-hadoop/61296/patch-findbugs-hadoop-tools_hadoop-streaming.txt | {code} The checkstyle path is very long, because the directory path is duplicated. As far as i investigated, shellcheck and whitespace output the correct path. > test-patch.sh plugins should abbreviate the path > > > Key: HADOOP-11897 > URL: https://issues.apache.org/jira/browse/HADOOP-11897 > Project: Hadoop Common > Issue Type: Test > Components: test >Reporter: Allen Wittenauer >Priority: Trivial > Labels: newbie > > The current output of checkstyle, shellcheck, and whitespace results in very > long file name paths. It might be useful to abbreviate them in some way, > maybe removing the entire path and leaving just the filename or last > directory+filename. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HADOOP-12246) Apache Hadoop should be listed under "big-data" and "hadoop" categories
[ https://issues.apache.org/jira/browse/HADOOP-12246?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14731798#comment-14731798 ] Kengo Seki commented on HADOOP-12246: - Hadoop uses git for the main source and the precommit assumes submitted patches are meant for the git repository. I'm a relative newcomer to hadoop development and not familiar with the ASF infra or doap.rdf, but experts will advise you in time. Anyway you can ignore the precommit error message. > Apache Hadoop should be listed under "big-data" and "hadoop" categories > --- > > Key: HADOOP-12246 > URL: https://issues.apache.org/jira/browse/HADOOP-12246 > Project: Hadoop Common > Issue Type: Task > Components: site >Reporter: Ajoy Bhatia >Assignee: Ajoy Bhatia >Priority: Trivial > Labels: categories > Fix For: site > > Attachments: site.patch > > Original Estimate: 2h > Remaining Estimate: 2h > > In the "Projects by category" list on > https://projects.apache.org/projects.html?category , Apache Hadoop is listed > in the database category only. > Apache Hadoop project should also be listed under big-data and hadoop > categories. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HADOOP-11897) test-patch.sh plugins should abbreviate the path
[ https://issues.apache.org/jira/browse/HADOOP-11897?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Kengo Seki updated HADOOP-11897: Issue Type: Sub-task (was: Test) Parent: HADOOP-12111 > test-patch.sh plugins should abbreviate the path > > > Key: HADOOP-11897 > URL: https://issues.apache.org/jira/browse/HADOOP-11897 > Project: Hadoop Common > Issue Type: Sub-task > Components: test >Reporter: Allen Wittenauer >Priority: Trivial > Labels: newbie > > The current output of checkstyle, shellcheck, and whitespace results in very > long file name paths. It might be useful to abbreviate them in some way, > maybe removing the entire path and leaving just the filename or last > directory+filename. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HADOOP-12312) Findbugs HTML report link shows 0 warnings despite errors
[ https://issues.apache.org/jira/browse/HADOOP-12312?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Kengo Seki updated HADOOP-12312: Attachment: Screen Shot 2015-09-04 at 00.25.44.png HADOOP-12312.HADOOP-12111.00.patch Attaching a patch. Abbreviated console output is as follows: {code} [sekikn@mobile hadoop]$ dev-support/test-patch.sh --basedir=../dev/hadoop --project=hadoop https://issues.apache.org/jira/secure/attachment/12751537/hadoop-reproduciable.git2c4208f.patch (snip) | Vote | Subsystem | Runtime | Comment | -1 | findbugs | 2m 53s| hadoop-common-project/hadoop-common | ||| introduced 2 new FindBugs issues. Reason | Tests FindBugs | module:hadoop-common-project/hadoop-common | Found reliance on default encoding in org.apache.hadoop.metrics2.sink.KafkaSink.putMetrics(MetricsRecord):in org.apache.hadoop.metrics2.sink.KafkaSink.putMetrics(MetricsRecord): String.getBytes() At KafkaSink.java:[line 135] | org.apache.hadoop.metrics2.sink.KafkaSink.putMetrics(MetricsRecord) invokes inefficient new String(String) constructor At KafkaSink.java:constructor At KafkaSink.java:[line 96] || Subsystem || Report/Notes || | findbugs | v3.0.0 | | findbugs | /private/tmp/test-patch-hadoop/58204/new-findbugs-hadoop-common-project_hadoop-common.html | {code} The screenshot of new-findbugs-hadoop-common-project_hadoop-common.html is also attached. It displays the correct information. > Findbugs HTML report link shows 0 warnings despite errors > - > > Key: HADOOP-12312 > URL: https://issues.apache.org/jira/browse/HADOOP-12312 > Project: Hadoop Common > Issue Type: Sub-task > Components: yetus >Affects Versions: HADOOP-12111 >Reporter: Varun Saxena >Assignee: Kengo Seki > Attachments: HADOOP-12312.HADOOP-12111.00.patch, Screen Shot > 2015-09-04 at 00.25.44.png, hadoop-reproduciable.git2c4208f.patch > > > Refer to Hadoop QA report below : > https://issues.apache.org/jira/browse/YARN-3232?focusedCommentId=14679146=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-14679146 > The report shows -1 for findbugs because there have been 7 findbugs warnings > introduced. But the HTML report in link is showing 0 findbugs warnings. > I verified locally and the warnings did indeed exist. > So there must be some problem in findbugs HTML report generation in > test-patch.sh > This inconsistency between -1 for findbugs and HTML report lead to these > findbugs warnings being leaked to trunk. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HADOOP-12375) Incomplete checking for findbugs executable
[ https://issues.apache.org/jira/browse/HADOOP-12375?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14729290#comment-14729290 ] Kengo Seki commented on HADOOP-12375: - Thanks [~jagadesh.kiran], but your patch seems to have an unnecessary header. Would you remove the first 3 lines? > Incomplete checking for findbugs executable > --- > > Key: HADOOP-12375 > URL: https://issues.apache.org/jira/browse/HADOOP-12375 > Project: Hadoop Common > Issue Type: Sub-task > Components: yetus >Affects Versions: HADOOP-12111 >Reporter: Kengo Seki >Assignee: Jagadesh Kiran N > Labels: newbie > Attachments: HADOOP-12375.HADOOP-12111.00.patch > > > In test-patch.d/findbugs.sh: > {code} > 65 function findbugs_is_installed > 66 { > 67 if [[ ! -e "${FINDBUGS_HOME}/bin/findbugs" ]]; then > 68 printf "\n\n%s is not executable.\n\n" "${FINDBUGS_HOME}/bin/findbugs" > 69 add_vote_table -1 findbugs "Findbugs is not installed." > 70 return 1 > 71 fi > 72 return 0 > 73 } > {code} > should be -x. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HADOOP-12312) Findbugs HTML report link shows 0 warnings despite errors
[ https://issues.apache.org/jira/browse/HADOOP-12312?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Kengo Seki updated HADOOP-12312: Status: Patch Available (was: Reopened) > Findbugs HTML report link shows 0 warnings despite errors > - > > Key: HADOOP-12312 > URL: https://issues.apache.org/jira/browse/HADOOP-12312 > Project: Hadoop Common > Issue Type: Sub-task > Components: yetus >Affects Versions: HADOOP-12111 >Reporter: Varun Saxena >Assignee: Kengo Seki > Attachments: HADOOP-12312.HADOOP-12111.00.patch, Screen Shot > 2015-09-04 at 00.25.44.png, hadoop-reproduciable.git2c4208f.patch > > > Refer to Hadoop QA report below : > https://issues.apache.org/jira/browse/YARN-3232?focusedCommentId=14679146=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-14679146 > The report shows -1 for findbugs because there have been 7 findbugs warnings > introduced. But the HTML report in link is showing 0 findbugs warnings. > I verified locally and the warnings did indeed exist. > So there must be some problem in findbugs HTML report generation in > test-patch.sh > This inconsistency between -1 for findbugs and HTML report lead to these > findbugs warnings being leaked to trunk. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HADOOP-4258) the test patch script should check for filenames that differ only in case
[ https://issues.apache.org/jira/browse/HADOOP-4258?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14729310#comment-14729310 ] Kengo Seki commented on HADOOP-4258: This feature is not yet incorporated in Yetus, so I'll make it a subtask of HADOOP-12111 for now. Thanks [~ozawa] for your reminder! > the test patch script should check for filenames that differ only in case > - > > Key: HADOOP-4258 > URL: https://issues.apache.org/jira/browse/HADOOP-4258 > Project: Hadoop Common > Issue Type: Test > Components: build, test >Reporter: Owen O'Malley >Assignee: Kengo Seki > Labels: newbie, test-patch > Attachments: HADOOP-4258.001.patch > > > It would be nice if the test patch script warned about filenames that differ > only in case. We recently had a patch committed that had a pair of colliding > filenames and subversion broke badly on my Mac. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HADOOP-4258) the test patch script should check for filenames that differ only in case
[ https://issues.apache.org/jira/browse/HADOOP-4258?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14729317#comment-14729317 ] Kengo Seki commented on HADOOP-4258: Cancelled patch because the current one is not for the HADOOP-12111 branch. > the test patch script should check for filenames that differ only in case > - > > Key: HADOOP-4258 > URL: https://issues.apache.org/jira/browse/HADOOP-4258 > Project: Hadoop Common > Issue Type: Sub-task > Components: build, test >Reporter: Owen O'Malley > Labels: newbie, test-patch > Attachments: HADOOP-4258.001.patch > > > It would be nice if the test patch script warned about filenames that differ > only in case. We recently had a patch committed that had a pair of colliding > filenames and subversion broke badly on my Mac. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HADOOP-4258) the test patch script should check for filenames that differ only in case
[ https://issues.apache.org/jira/browse/HADOOP-4258?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Kengo Seki updated HADOOP-4258: --- Assignee: (was: Kengo Seki) Status: Open (was: Patch Available) > the test patch script should check for filenames that differ only in case > - > > Key: HADOOP-4258 > URL: https://issues.apache.org/jira/browse/HADOOP-4258 > Project: Hadoop Common > Issue Type: Sub-task > Components: build, test >Reporter: Owen O'Malley > Labels: newbie, test-patch > Attachments: HADOOP-4258.001.patch > > > It would be nice if the test patch script warned about filenames that differ > only in case. We recently had a patch committed that had a pair of colliding > filenames and subversion broke badly on my Mac. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HADOOP-4258) the test patch script should check for filenames that differ only in case
[ https://issues.apache.org/jira/browse/HADOOP-4258?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Kengo Seki updated HADOOP-4258: --- Issue Type: Sub-task (was: Test) Parent: HADOOP-12111 > the test patch script should check for filenames that differ only in case > - > > Key: HADOOP-4258 > URL: https://issues.apache.org/jira/browse/HADOOP-4258 > Project: Hadoop Common > Issue Type: Sub-task > Components: build, test >Reporter: Owen O'Malley >Assignee: Kengo Seki > Labels: newbie, test-patch > Attachments: HADOOP-4258.001.patch > > > It would be nice if the test patch script warned about filenames that differ > only in case. We recently had a patch committed that had a pair of colliding > filenames and subversion broke badly on my Mac. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Created] (HADOOP-12375) Incomplete checking for findbugs executable
Kengo Seki created HADOOP-12375: --- Summary: Incomplete checking for findbugs executable Key: HADOOP-12375 URL: https://issues.apache.org/jira/browse/HADOOP-12375 Project: Hadoop Common Issue Type: Sub-task Components: yetus Affects Versions: HADOOP-12111 Reporter: Kengo Seki In test-patch.d/findbugs.sh: {code} 65 function findbugs_is_installed 66 { 67 if [[ ! -e "${FINDBUGS_HOME}/bin/findbugs" ]]; then 68 printf "\n\n%s is not executable.\n\n" "${FINDBUGS_HOME}/bin/findbugs" 69 add_vote_table -1 findbugs "Findbugs is not installed." 70 return 1 71 fi 72 return 0 73 } {code} should be -x. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HADOOP-12312) Findbugs HTML report link shows 0 warnings despite errors
[ https://issues.apache.org/jira/browse/HADOOP-12312?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14728490#comment-14728490 ] Kengo Seki commented on HADOOP-12312: - This code: {code} 315 #shellcheck disable=SC2016 316 new_findbugs_warnings=$("${FINDBUGS_HOME}/bin/filterBugs" -first patch \ 317 "${combined_xml}" "${newbugsbase}.xml" | ${AWK} '{print $1}') 318 if [[ $? != 0 ]]; then 319 popd >/dev/null 320 module_status ${i} -1 "" "${module} cannot run filterBugs (#1) from findbugs" 321 ((result=result+1)) 322 savestop=$(stop_clock) 323 MODULE_STATUS_TIMER[${i}]=${savestop} 324 ((i=i+1)) 325 continue 326 fi 327 328 #shellcheck disable=SC2016 329 new_findbugs_fixed_warnings=$("${FINDBUGS_HOME}/bin/filterBugs" -fixed patch \ 330 "${combined_xml}" "${newbugsbase}.xml" | ${AWK} '{print $1}') 331 if [[ $? != 0 ]]; then 332 popd >/dev/null 333 module_status ${i} -1 "" "${module} cannot run filterBugs (#2) from findbugs" 334 ((result=result+1)) 335 savestop=$(stop_clock) 336 MODULE_STATUS_TIMER[${i}]=${savestop} 337 ((i=i+1)) 338 continue 339 fi {code} writes to $\{newbugsbase}.xml twice, so new warnings will be removed. > Findbugs HTML report link shows 0 warnings despite errors > - > > Key: HADOOP-12312 > URL: https://issues.apache.org/jira/browse/HADOOP-12312 > Project: Hadoop Common > Issue Type: Sub-task > Components: yetus >Affects Versions: HADOOP-12111 >Reporter: Varun Saxena > Attachments: hadoop-reproduciable.git2c4208f.patch > > > Refer to Hadoop QA report below : > https://issues.apache.org/jira/browse/YARN-3232?focusedCommentId=14679146=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-14679146 > The report shows -1 for findbugs because there have been 7 findbugs warnings > introduced. But the HTML report in link is showing 0 findbugs warnings. > I verified locally and the warnings did indeed exist. > So there must be some problem in findbugs HTML report generation in > test-patch.sh > This inconsistency between -1 for findbugs and HTML report lead to these > findbugs warnings being leaked to trunk. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Assigned] (HADOOP-12312) Findbugs HTML report link shows 0 warnings despite errors
[ https://issues.apache.org/jira/browse/HADOOP-12312?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Kengo Seki reassigned HADOOP-12312: --- Assignee: Kengo Seki > Findbugs HTML report link shows 0 warnings despite errors > - > > Key: HADOOP-12312 > URL: https://issues.apache.org/jira/browse/HADOOP-12312 > Project: Hadoop Common > Issue Type: Sub-task > Components: yetus >Affects Versions: HADOOP-12111 >Reporter: Varun Saxena >Assignee: Kengo Seki > Attachments: hadoop-reproduciable.git2c4208f.patch > > > Refer to Hadoop QA report below : > https://issues.apache.org/jira/browse/YARN-3232?focusedCommentId=14679146=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-14679146 > The report shows -1 for findbugs because there have been 7 findbugs warnings > introduced. But the HTML report in link is showing 0 findbugs warnings. > I verified locally and the warnings did indeed exist. > So there must be some problem in findbugs HTML report generation in > test-patch.sh > This inconsistency between -1 for findbugs and HTML report lead to these > findbugs warnings being leaked to trunk. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HADOOP-12118) Validate xml configuration files with XML Schema
[ https://issues.apache.org/jira/browse/HADOOP-12118?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14728218#comment-14728218 ] Kengo Seki commented on HADOOP-12118: - If we proceed as my last comment, I'd like to wait the backport to branch-2 until the refactoring on trunk is finished, because it will change the output message format. > Validate xml configuration files with XML Schema > > > Key: HADOOP-12118 > URL: https://issues.apache.org/jira/browse/HADOOP-12118 > Project: Hadoop Common > Issue Type: Improvement >Reporter: Christopher Tubbs > Attachments: HADOOP-7947.branch-2.1.patch, hadoop-configuration.xsd > > > I spent an embarrassingly long time today trying to figure out why the > following wouldn't work. > {code} > > fs.defaultFS > hdfs://localhost:9000 > > {code} > I just kept getting an error about no authority for {{fs.defaultFS}}, with a > value of {{file:///}}, which made no sense... because I knew it was there. > The problem was that the {{core-site.xml}} was parsed entirely without any > validation. This seems incorrect. The very least that could be done is a > simple XML Schema validation against an XSD, before parsing. That way, users > will get immediate failures on common typos and other problems in the xml > configuration files. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HADOOP-12336) github_jira_bridge doesn't work
[ https://issues.apache.org/jira/browse/HADOOP-12336?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Kengo Seki updated HADOOP-12336: Attachment: HADOOP-12336.HADOOP-12111.01.patch Thanks! Attaching a revised patch. I confirmed it works using HADOOP-11820: {code} [sekikn@mobile hadoop]$ dev-support/test-patch.sh --basedir=../dev/hadoop --project=hadoop HADOOP-11820 (snip) HADOOP-11820 appears to be a Github PR. Switching Modes. GITHUB PR #2 is being downloaded at Tue Sep 1 01:00:41 JST 2015 from https://github.com/aw-altiscale/hadoop/pull/2 Patch from GITHUB PR #2 is being downloaded at Tue Sep 1 01:00:42 JST 2015 from https://github.com/aw-altiscale/hadoop/pull/2.patch (snip) -1 overall _ _ __ | ___|_ _(_) |_ _ _ __ ___| | | |_ / _` | | | | | | '__/ _ \ | | _| (_| | | | |_| | | | __/_| |_| \__,_|_|_|\__,_|_| \___(_) | Vote | Subsystem | Runtime | Comment | +1 | site | 0m 00s| HADOOP-12111 passed | 0 | @author | 0m 00s| Skipping @author checks as test-patch.sh | ||| has been patched. | -1 |test4tests | 0m 00s| The patch doesn't appear to include any | ||| new or modified tests. Please justify why | ||| no new tests are needed for this patch. | ||| Also please list what manual steps were | ||| performed to verify this patch. | +1 | site | 0m 00s| the patch passed | +1 |asflicense | 0m 26s| Patch does not generate ASF License | ||| warnings. | -1 |shellcheck | 0m 12s| The applied patch generated 3 new | ||| shellcheck issues (total was 34, now 37). | +1 |whitespace | 0m 00s| Patch has no whitespace issues. | || 1m 04s| || Subsystem || Report/Notes || | JIRA Issue | HADOOP-11820 | | GITHUB PR | https://github.com/aw-altiscale/hadoop/pull/2 | | git revision | HADOOP-12111 / b006c9a | | Optional Tests | asflicense site unit shellcheck | | uname | Darwin mobile.local 14.5.0 Darwin Kernel Version 14.5.0: Wed Jul 29 02:26:53 PDT 2015; root:xnu-2782.40.9~1/RELEASE_X86_64 x86_64 | | Build tool | maven | | Personality | /Users/sekikn/hadoop/dev-support/personality/hadoop.sh | | Default Java | 1.7.0_80 | | shellcheck | v0.3.6 | | shellcheck | /private/tmp/test-patch-hadoop/29857/diff-patch-shellcheck.txt | | Max memory used | 46MB | {code} > github_jira_bridge doesn't work > --- > > Key: HADOOP-12336 > URL: https://issues.apache.org/jira/browse/HADOOP-12336 > Project: Hadoop Common > Issue Type: Sub-task > Components: yetus >Affects Versions: HADOOP-12111 >Reporter: Allen Wittenauer >Assignee: Allen Wittenauer >Priority: Blocker > Attachments: HADOOP-12336.HADOOP-12111.00.patch, > HADOOP-12336.HADOOP-12111.01.patch > > > The github_jira_bridge (which allows the JIRA bugsystem to switch to github > mode) is failing. See comments. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HADOOP-12118) Validate xml configuration files with XML Schema
[ https://issues.apache.org/jira/browse/HADOOP-12118?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14724763#comment-14724763 ] Kengo Seki commented on HADOOP-12118: - Sorry [~gliptak] for the late response. I agree with the backport because it must be useful for the 2.x users. But I couldn't decide whether we should adopt xsd for validation or not. One possibility is, using xsd for the basic structure validation and xpath (for example) for the advanced validations I mentioned above to avoid direct xml walking, like: {code} XPath xpath = XPathFactory.newInstance().newXPath(); NodeList nodes = (NodeList) xpath.evaluate("/configuration/property/name/text()", new InputSource("core-site.xml"), XPathConstants.NODESET); Set s = new HashSet(); for (int i=0; iValidate xml configuration files with XML Schema > > > Key: HADOOP-12118 > URL: https://issues.apache.org/jira/browse/HADOOP-12118 > Project: Hadoop Common > Issue Type: Improvement >Reporter: Christopher Tubbs > Attachments: HADOOP-7947.branch-2.1.patch, hadoop-configuration.xsd > > > I spent an embarrassingly long time today trying to figure out why the > following wouldn't work. > {code} > > fs.defaultFS > hdfs://localhost:9000 > > {code} > I just kept getting an error about no authority for {{fs.defaultFS}}, with a > value of {{file:///}}, which made no sense... because I knew it was there. > The problem was that the {{core-site.xml}} was parsed entirely without any > validation. This seems incorrect. The very least that could be done is a > simple XML Schema validation against an XSD, before parsing. That way, users > will get immediate failures on common typos and other problems in the xml > configuration files. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HADOOP-12336) github_jira_bridge doesn't work
[ https://issues.apache.org/jira/browse/HADOOP-12336?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14721534#comment-14721534 ] Kengo Seki commented on HADOOP-12336: - I tried 00.patch on HADOOP-11820 and confirmed the PR is successfully applied. But some nits: {code} urlfromjira=$(${AWK} 'match($0,https://github.com/.*patch;){print substr($0,RSTART,RLENGTH)}' ${PATCH_DIR}/jira | tail -1) {code} * Using $\{GITHUB_BASE_URL} instead of a hard-coded URL seems better. Is there any reason? * https://github.com/[^ ]*patch seems safer, as well as jira_locate_patch does. github_jira_bridge doesn't work --- Key: HADOOP-12336 URL: https://issues.apache.org/jira/browse/HADOOP-12336 Project: Hadoop Common Issue Type: Sub-task Components: yetus Affects Versions: HADOOP-12111 Reporter: Allen Wittenauer Assignee: Allen Wittenauer Priority: Blocker Attachments: HADOOP-12336.HADOOP-12111.00.patch The github_jira_bridge (which allows the JIRA bugsystem to switch to github mode) is failing. See comments. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HADOOP-12340) test-patch docker mode fails in downloading findbugs with curl
[ https://issues.apache.org/jira/browse/HADOOP-12340?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14711564#comment-14711564 ] Kengo Seki commented on HADOOP-12340: - Oh, sorry. If it needs to be applied, please ignore. test-patch docker mode fails in downloading findbugs with curl -- Key: HADOOP-12340 URL: https://issues.apache.org/jira/browse/HADOOP-12340 Project: Hadoop Common Issue Type: Sub-task Components: yetus Affects Versions: HADOOP-12111 Reporter: Kengo Seki Assignee: Kengo Seki Fix For: HADOOP-12111 Attachments: HADOOP-12340.HADOOP-12111.00.patch, HADOOP-12340.HADOOP-12111.01.patch HADOOP-12129 replaced wget commands in test-patch with curl. But curl doesn't follow URL redirection by default, so docker mode fails in downloading findbugs for now: {code} $ dev-support/test-patch.sh --docker --project=hadoop /tmp/test.patch (snip) Step 11 : RUN mkdir -p /opt/findbugs curl https://sourceforge.net/projects/findbugs/files/findbugs/3.0.1/findbugs-noUpdateChecks-3.0.1.tar.gz/download -o /opt/findbugs.tar.gz tar xzf /opt/findbugs.tar.gz --strip-components 1 -C /opt/findbugs --- Running in 31c71013bb59 % Total% Received % Xferd Average Speed TimeTime Time Current Dload Upload Total SpentLeft Speed 100 397 100 3970 0168 0 0:00:02 0:00:02 --:--:-- 168 gzip: stdin: not in gzip format tar: Child returned status 1 tar: Error is not recoverable: exiting now The command '/bin/sh -c mkdir -p /opt/findbugs curl https://sourceforge.net/projects/findbugs/files/findbugs/3.0.1/findbugs-noUpdateChecks-3.0.1.tar.gz/download -o /opt/findbugs.tar.gz tar xzf /opt/findbugs.tar.gz --strip-components 1 -C /opt/findbugs' returned a non-zero code: 2 {code} Adding -L option will resolve this. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Created] (HADOOP-12355) test-patch TAP plugin should use $\{SED} instead of sed
Kengo Seki created HADOOP-12355: --- Summary: test-patch TAP plugin should use $\{SED} instead of sed Key: HADOOP-12355 URL: https://issues.apache.org/jira/browse/HADOOP-12355 Project: Hadoop Common Issue Type: Sub-task Components: yetus Affects Versions: HADOOP-12111 Reporter: Kengo Seki Priority: Trivial for consistency and platform compatibility. {code} 54 if [[ -n ${filenames} ]]; then 55 module_failed_tests=$(echo ${filenames} \ 56 | sed -e s,${TAP_LOG_DIR},,g -e s,^/,,g ) {code} -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HADOOP-12355) test-patch TAP plugin should use ${SED} instead of sed
[ https://issues.apache.org/jira/browse/HADOOP-12355?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Kengo Seki updated HADOOP-12355: Summary: test-patch TAP plugin should use ${SED} instead of sed (was: test-patch TAP plugin should use $\{SED} instead of sed) test-patch TAP plugin should use ${SED} instead of sed -- Key: HADOOP-12355 URL: https://issues.apache.org/jira/browse/HADOOP-12355 Project: Hadoop Common Issue Type: Sub-task Components: yetus Affects Versions: HADOOP-12111 Reporter: Kengo Seki Priority: Trivial Labels: newbie for consistency and platform compatibility. {code} 54 if [[ -n ${filenames} ]]; then 55 module_failed_tests=$(echo ${filenames} \ 56 | sed -e s,${TAP_LOG_DIR},,g -e s,^/,,g ) {code} -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HADOOP-12340) test-patch docker mode fails in downloading findbugs with curl
[ https://issues.apache.org/jira/browse/HADOOP-12340?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Kengo Seki updated HADOOP-12340: Attachment: HADOOP-12340.HADOOP-12111.01.patch -01: * replace -L with --location for consistency with other curl commands test-patch docker mode fails in downloading findbugs with curl -- Key: HADOOP-12340 URL: https://issues.apache.org/jira/browse/HADOOP-12340 Project: Hadoop Common Issue Type: Sub-task Components: yetus Affects Versions: HADOOP-12111 Reporter: Kengo Seki Assignee: Kengo Seki Fix For: HADOOP-12111 Attachments: HADOOP-12340.HADOOP-12111.00.patch, HADOOP-12340.HADOOP-12111.01.patch HADOOP-12129 replaced wget commands in test-patch with curl. But curl doesn't follow URL redirection by default, so docker mode fails in downloading findbugs for now: {code} $ dev-support/test-patch.sh --docker --project=hadoop /tmp/test.patch (snip) Step 11 : RUN mkdir -p /opt/findbugs curl https://sourceforge.net/projects/findbugs/files/findbugs/3.0.1/findbugs-noUpdateChecks-3.0.1.tar.gz/download -o /opt/findbugs.tar.gz tar xzf /opt/findbugs.tar.gz --strip-components 1 -C /opt/findbugs --- Running in 31c71013bb59 % Total% Received % Xferd Average Speed TimeTime Time Current Dload Upload Total SpentLeft Speed 100 397 100 3970 0168 0 0:00:02 0:00:02 --:--:-- 168 gzip: stdin: not in gzip format tar: Child returned status 1 tar: Error is not recoverable: exiting now The command '/bin/sh -c mkdir -p /opt/findbugs curl https://sourceforge.net/projects/findbugs/files/findbugs/3.0.1/findbugs-noUpdateChecks-3.0.1.tar.gz/download -o /opt/findbugs.tar.gz tar xzf /opt/findbugs.tar.gz --strip-components 1 -C /opt/findbugs' returned a non-zero code: 2 {code} Adding -L option will resolve this. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Created] (HADOOP-12354) github_find_jira_title in test-patch github plugin returns 0 even if jira_determine_issue failed
Kengo Seki created HADOOP-12354: --- Summary: github_find_jira_title in test-patch github plugin returns 0 even if jira_determine_issue failed Key: HADOOP-12354 URL: https://issues.apache.org/jira/browse/HADOOP-12354 Project: Hadoop Common Issue Type: Sub-task Components: yetus Affects Versions: HADOOP-12111 Reporter: Kengo Seki So the following check for $\{GITHUB_ISSUE} seems to be skipped and github_determine_issue seems to succeed almost always. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HADOOP-12354) github_find_jira_title in test-patch github plugin returns 0 even if jira_determine_issue failed
[ https://issues.apache.org/jira/browse/HADOOP-12354?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14711589#comment-14711589 ] Kengo Seki commented on HADOOP-12354: - I think it isn't. After applying HADOOP-12336.HADOOP-12111.00.patch, github_find_jira_title is still as follows: {code} function github_find_jira_title { declare title declare maybe declare retval if [[ ! -f ${PATCH_DIR}/github-pull.json ]]; then return 1 fi title=$(GREP title ${PATCH_DIR}/github-pull.json \ | cut -f4 -d\) # people typically do two types: JIRA-ISSUE: and [JIRA-ISSUE] # JIRA_ISSUE_RE is pretty strict so we need to chop that stuff # out first maybe=$(echo ${title} | cut -f2 -d\[ | cut -f1 -d\]) jira_determine_issue ${maybe} retval=$? if [[ ${retval} == 0 ]]; then return 0 fi maybe=$(echo ${title} | cut -f1 -d:) jira_determine_issue ${maybe} retval=$? if [[ ${retval} == 0 ]]; then return 0 fi } {code} Should it return other than 0 if both invocations of jira_determine_issue fail? github_find_jira_title in test-patch github plugin returns 0 even if jira_determine_issue failed Key: HADOOP-12354 URL: https://issues.apache.org/jira/browse/HADOOP-12354 Project: Hadoop Common Issue Type: Sub-task Components: yetus Affects Versions: HADOOP-12111 Reporter: Kengo Seki So the following check for $\{GITHUB_ISSUE} seems to be skipped and github_determine_issue seems to succeed almost always. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Created] (HADOOP-12349) Misleading debug message in generic_locate_patch
Kengo Seki created HADOOP-12349: --- Summary: Misleading debug message in generic_locate_patch Key: HADOOP-12349 URL: https://issues.apache.org/jira/browse/HADOOP-12349 Project: Hadoop Common Issue Type: Sub-task Components: yetus Affects Versions: HADOOP-12111 Reporter: Kengo Seki Priority: Minor The following message in builtin-bugsystem.sh is duplicated with jira_locate_patch and misleading. {code} yetus_debug jira_locate_patch: not a JIRA. {code} -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HADOOP-12340) test-patch docker mode fails in downloading findbugs with curl
[ https://issues.apache.org/jira/browse/HADOOP-12340?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Kengo Seki updated HADOOP-12340: Attachment: HADOOP-12340.HADOOP-12111.00.patch Attaching a patch. With this patch, docker mode succeeds as follows: {code} [sekikn@mobile hadoop]$ dev-support/test-patch.sh --basedir=../dev/hadoop --docker --patch-dir=foo --project=hadoop /tmp/test.patch (snip) Step 11 : RUN mkdir -p /opt/findbugs curl -L https://sourceforge.net/projects/findbugs/files/findbugs/3.0.1/findbugs-noUpdateChecks-3.0.1.tar.gz/download -o /opt/findbugs.tar.gz tar xzf /opt/findbugs.tar.gz --strip-components 1 -C /opt/findbugs --- Using cache --- 3b1253eaf7b3 (snip) Successfully built 7417b6ced5f2 (snip) | Vote | Subsystem | Runtime | Comment | 0 |reexec | 0m 6s | docker + precommit patch detected. | +1 | @author | 0m 0s | The patch does not contain any @author | ||| tags. | +1 |asflicense | 7m 8s | Patch does not generate ASF License | ||| warnings. | +1 |whitespace | 0m 0s | Patch has no whitespace issues. | || 8m 47s| || Subsystem || Report/Notes || | Docker | Client=1.7.1 Server=1.7.1 Image:test-patch-base-hadoop-86a1025 | | git revision | HADOOP-12111 / 2c4208f | | Optional Tests | asflicense | | uname | Linux 9721e2853d64 4.0.7-boot2docker #1 SMP Wed Jul 15 00:01:41 UTC 2015 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality | /testptch/patchprocess/precommit/personality/hadoop.sh | | Default Java | 1.7.0_79 | | Max memory used | 30MB | {code} test-patch docker mode fails in downloading findbugs with curl -- Key: HADOOP-12340 URL: https://issues.apache.org/jira/browse/HADOOP-12340 Project: Hadoop Common Issue Type: Sub-task Components: yetus Affects Versions: HADOOP-12111 Reporter: Kengo Seki Attachments: HADOOP-12340.HADOOP-12111.00.patch HADOOP-12129 replaced wget commands in test-patch with curl. But curl doesn't follow URL redirection by default, so docker mode fails in downloading findbugs for now: {code} $ dev-support/test-patch.sh --docker --project=hadoop /tmp/test.patch (snip) Step 11 : RUN mkdir -p /opt/findbugs curl https://sourceforge.net/projects/findbugs/files/findbugs/3.0.1/findbugs-noUpdateChecks-3.0.1.tar.gz/download -o /opt/findbugs.tar.gz tar xzf /opt/findbugs.tar.gz --strip-components 1 -C /opt/findbugs --- Running in 31c71013bb59 % Total% Received % Xferd Average Speed TimeTime Time Current Dload Upload Total SpentLeft Speed 100 397 100 3970 0168 0 0:00:02 0:00:02 --:--:-- 168 gzip: stdin: not in gzip format tar: Child returned status 1 tar: Error is not recoverable: exiting now The command '/bin/sh -c mkdir -p /opt/findbugs curl https://sourceforge.net/projects/findbugs/files/findbugs/3.0.1/findbugs-noUpdateChecks-3.0.1.tar.gz/download -o /opt/findbugs.tar.gz tar xzf /opt/findbugs.tar.gz --strip-components 1 -C /opt/findbugs' returned a non-zero code: 2 {code} Adding -L option will resolve this. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HADOOP-12340) test-patch docker mode fails in downloading findbugs with curl
[ https://issues.apache.org/jira/browse/HADOOP-12340?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Kengo Seki updated HADOOP-12340: Assignee: Kengo Seki Status: Patch Available (was: Open) test-patch docker mode fails in downloading findbugs with curl -- Key: HADOOP-12340 URL: https://issues.apache.org/jira/browse/HADOOP-12340 Project: Hadoop Common Issue Type: Sub-task Components: yetus Affects Versions: HADOOP-12111 Reporter: Kengo Seki Assignee: Kengo Seki Attachments: HADOOP-12340.HADOOP-12111.00.patch HADOOP-12129 replaced wget commands in test-patch with curl. But curl doesn't follow URL redirection by default, so docker mode fails in downloading findbugs for now: {code} $ dev-support/test-patch.sh --docker --project=hadoop /tmp/test.patch (snip) Step 11 : RUN mkdir -p /opt/findbugs curl https://sourceforge.net/projects/findbugs/files/findbugs/3.0.1/findbugs-noUpdateChecks-3.0.1.tar.gz/download -o /opt/findbugs.tar.gz tar xzf /opt/findbugs.tar.gz --strip-components 1 -C /opt/findbugs --- Running in 31c71013bb59 % Total% Received % Xferd Average Speed TimeTime Time Current Dload Upload Total SpentLeft Speed 100 397 100 3970 0168 0 0:00:02 0:00:02 --:--:-- 168 gzip: stdin: not in gzip format tar: Child returned status 1 tar: Error is not recoverable: exiting now The command '/bin/sh -c mkdir -p /opt/findbugs curl https://sourceforge.net/projects/findbugs/files/findbugs/3.0.1/findbugs-noUpdateChecks-3.0.1.tar.gz/download -o /opt/findbugs.tar.gz tar xzf /opt/findbugs.tar.gz --strip-components 1 -C /opt/findbugs' returned a non-zero code: 2 {code} Adding -L option will resolve this. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HADOOP-12301) Fix some test-patch plugins to count the diff lines correctly
[ https://issues.apache.org/jira/browse/HADOOP-12301?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Kengo Seki updated HADOOP-12301: Attachment: HADOOP-12301.HADOOP-12111.00.patch -00: * Change rubocop formatter from clang to emacs. The latter outputs only one line for one issue, so we can simply count lines as the number of issues. To be safe, I left the check for the number of fields, in case that the message part in a line contains a newline. * Change pylint.sh to count only lines matched with {code}/^.*:.*: \[.*\] /{code} (e.g., a.py:1: \[E0001(syntax-error), ] invalid syntax) to skip unnecessary lines. This pattern is compatible with a very early (2006) version of the --parseable option. https://bitbucket.org/logilab/pylint/commits/79b31c260ad15738a52b504b0ae29b52abf6367b?at=before-astroid#chg-reporters/text.py I confirmed they report correct numbers of issues using the above example patches. Rubocop: {code} | Vote | Subsystem | Runtime | Comment | -1 |rubocop | 0m 02s| The applied patch generated 2 new | | || rubocop issues (total was 77, now 79). || Subsystem || Report/Notes || | rubocop | v0.32.1 | | rubocop | /private/tmp/test-patch-hbase/83220/diff-patch-rubocop.txt | [sekikn@mobile hadoop]$ cat /private/tmp/test-patch-hbase/83220/diff-patch-rubocop.txt /Users/sekikn/hbase/bin/draining_servers.rb:165:4: C: Do not use `::` for method calls. /Users/sekikn/hbase/bin/draining_servers.rb:165:9: C: Do not use `::` for method calls. [sekikn@mobile hadoop]$ {code} Pylint: {code} | Vote | Subsystem | Runtime | Comment | -1 |pylint | 0m 05s| The applied patch generated 6 new pylint | ||| issues (total was 437, now 443). || Subsystem || Report/Notes || | pylint | v1.4.4 | | pylint | /private/tmp/test-patch-hadoop/61453/diff-patch-pylint.txt | [sekikn@mobile hadoop]$ cat /private/tmp/test-patch-hadoop/61453/diff-patch-pylint.txt dev-support/releasedocmaker.py:584: [W0311(bad-indentation), ] Bad indentation. Found 4 spaces, expected 2 dev-support/releasedocmaker.py:584: [C0326(bad-whitespace), ] No space allowed after bracket ( ( ) ) ^ dev-support/releasedocmaker.py:584: [C0326(bad-whitespace), ] No space allowed after bracket ( ( ) ) ^ dev-support/releasedocmaker.py:584: [C0326(bad-whitespace), ] No space allowed before bracket ( ( ) ) ^ dev-support/releasedocmaker.py:584: [C0326(bad-whitespace), ] No space allowed before bracket ( ( ) ) ^ dev-support/releasedocmaker.py:584: [W0104(pointless-statement), ] Statement seems to have no effect [sekikn@mobile hadoop]$ {code} Fix some test-patch plugins to count the diff lines correctly - Key: HADOOP-12301 URL: https://issues.apache.org/jira/browse/HADOOP-12301 Project: Hadoop Common Issue Type: Sub-task Components: yetus Affects Versions: HADOOP-12111 Reporter: Kengo Seki Attachments: HADOOP-12301.HADOOP-12111.00.patch 1. rubocop.sh counts only lines which have at least five fields separated by a colon: {code} calcdiffs ${PATCH_DIR}/branch-rubocop-result.txt ${PATCH_DIR}/patch-rubocop-result.txt ${PATCH_DIR}/diff-patch-rubocop.txt diffPostpatch=$(${AWK} -F: 'BEGIN {sum=0} 4NF {sum+=1} END {print sum}' ${PATCH_DIR}/diff-patch-rubocop.txt) if [[ ${diffPostpatch} -gt 0 ]] ; then # shellcheck disable=SC2016 numPrepatch=$(${AWK} -F: 'BEGIN {sum=0} 4NF {sum+=1} END {print sum}' ${PATCH_DIR}/branch-rubocop-result.txt) # shellcheck disable=SC2016 numPostpatch=$(${AWK} -F: 'BEGIN {sum=0} 4NF {sum+=1} END {print sum}' ${PATCH_DIR}/patch-rubocop-result.txt) {code} This is because the diff result can contain multiple lines for one issue. For example: {code} [sekikn@mobile hadoop]$ cat /private/tmp/test-patch-hbase/25821/diff-patch-rubocop.txt bin/draining_servers.rb:165:1: C: Do not introduce global variables. $foo {code} Checking the number of fields is intended to skip the second and third lines in the above diff file. But four or more colons can appear in the source code itself, for example: {code} | Vote | Subsystem | Runtime | Comment | -1 |rubocop | 0m 02s| The applied patch generated 4 new | | || rubocop issues (total was 77, now 81). || Subsystem ||
[jira] [Updated] (HADOOP-12301) Fix some test-patch plugins to count the diff lines correctly
[ https://issues.apache.org/jira/browse/HADOOP-12301?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Kengo Seki updated HADOOP-12301: Assignee: Kengo Seki Status: Patch Available (was: Open) Fix some test-patch plugins to count the diff lines correctly - Key: HADOOP-12301 URL: https://issues.apache.org/jira/browse/HADOOP-12301 Project: Hadoop Common Issue Type: Sub-task Components: yetus Affects Versions: HADOOP-12111 Reporter: Kengo Seki Assignee: Kengo Seki Attachments: HADOOP-12301.HADOOP-12111.00.patch 1. rubocop.sh counts only lines which have at least five fields separated by a colon: {code} calcdiffs ${PATCH_DIR}/branch-rubocop-result.txt ${PATCH_DIR}/patch-rubocop-result.txt ${PATCH_DIR}/diff-patch-rubocop.txt diffPostpatch=$(${AWK} -F: 'BEGIN {sum=0} 4NF {sum+=1} END {print sum}' ${PATCH_DIR}/diff-patch-rubocop.txt) if [[ ${diffPostpatch} -gt 0 ]] ; then # shellcheck disable=SC2016 numPrepatch=$(${AWK} -F: 'BEGIN {sum=0} 4NF {sum+=1} END {print sum}' ${PATCH_DIR}/branch-rubocop-result.txt) # shellcheck disable=SC2016 numPostpatch=$(${AWK} -F: 'BEGIN {sum=0} 4NF {sum+=1} END {print sum}' ${PATCH_DIR}/patch-rubocop-result.txt) {code} This is because the diff result can contain multiple lines for one issue. For example: {code} [sekikn@mobile hadoop]$ cat /private/tmp/test-patch-hbase/25821/diff-patch-rubocop.txt bin/draining_servers.rb:165:1: C: Do not introduce global variables. $foo {code} Checking the number of fields is intended to skip the second and third lines in the above diff file. But four or more colons can appear in the source code itself, for example: {code} | Vote | Subsystem | Runtime | Comment | -1 |rubocop | 0m 02s| The applied patch generated 4 new | | || rubocop issues (total was 77, now 81). || Subsystem || Report/Notes || | rubocop | v0.32.1 | | rubocop | /private/tmp/test-patch-hbase/5632/diff-patch-rubocop.txt | [sekikn@mobile hadoop]$ cat /private/tmp/test-patch-hbase/5632/diff-patch-rubocop.txt bin/draining_servers.rb:165:4: C: Do not use :: for method calls. foo::bar::baz ^^ bin/draining_servers.rb:165:9: C: Do not use :: for method calls. foo::bar::baz ^^ [sekikn@mobile hadoop]$ {code} In this case, new rubocop issues should be 2, but counted as 4 incorrectly. More reliable way to count is needed. 2. pylint.sh has the same problem. In addition, I removed awk's '-F:' option by mistake on HADOOP-12286. It can report a wrong number for now. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HADOOP-12301) Fix some test-patch plugins to count the diff lines correctly
[ https://issues.apache.org/jira/browse/HADOOP-12301?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Kengo Seki updated HADOOP-12301: Description: 1. rubocop.sh counts only lines which have at least five fields separated by a colon: {code} calcdiffs ${PATCH_DIR}/branch-rubocop-result.txt ${PATCH_DIR}/patch-rubocop-result.txt ${PATCH_DIR}/diff-patch-rubocop.txt diffPostpatch=$(${AWK} -F: 'BEGIN {sum=0} 4NF {sum+=1} END {print sum}' ${PATCH_DIR}/diff-patch-rubocop.txt) if [[ ${diffPostpatch} -gt 0 ]] ; then # shellcheck disable=SC2016 numPrepatch=$(${AWK} -F: 'BEGIN {sum=0} 4NF {sum+=1} END {print sum}' ${PATCH_DIR}/branch-rubocop-result.txt) # shellcheck disable=SC2016 numPostpatch=$(${AWK} -F: 'BEGIN {sum=0} 4NF {sum+=1} END {print sum}' ${PATCH_DIR}/patch-rubocop-result.txt) {code} This is because the diff result can contain multiple lines for one issue. For example: {code} [sekikn@mobile hadoop]$ cat /private/tmp/test-patch-hbase/25821/diff-patch-rubocop.txt bin/draining_servers.rb:165:1: C: Do not introduce global variables. $foo {code} Checking the number of fields is intended to skip the second and third lines in the above diff file. But four or more colons can appear in the source code itself, for example: {code} | Vote | Subsystem | Runtime | Comment | -1 |rubocop | 0m 02s| The applied patch generated 4 new | | || rubocop issues (total was 77, now 81). || Subsystem || Report/Notes || | rubocop | v0.32.1 | | rubocop | /private/tmp/test-patch-hbase/5632/diff-patch-rubocop.txt | [sekikn@mobile hadoop]$ cat /private/tmp/test-patch-hbase/5632/diff-patch-rubocop.txt bin/draining_servers.rb:165:4: C: Do not use :: for method calls. foo::bar::baz ^^ bin/draining_servers.rb:165:9: C: Do not use :: for method calls. foo::bar::baz ^^ [sekikn@mobile hadoop]$ {code} In this case, new rubocop issues should be 2, but counted as 4 incorrectly. More reliable way to count is needed. 2. pylint.sh has the same problem. In addition, I removed awk's '-F:' option by mistake on HADOOP-12286. It can report a wrong number for now. was: Some test-patch plugins have problems about counting the diff lines. 1. ruby-lint.sh counts only lines which have five or more fields: {code} diffPostpatch=$(${AWK} -F: 'BEGIN {sum=0} 4NF {sum+=1} END {print sum}' ${PATCH_DIR}/diff-patch-ruby-lint.txt) if [[ ${diffPostpatch} -gt 0 ]] ; then # shellcheck disable=SC2016 numPrepatch=$(${AWK} -F: 'BEGIN {sum=0} 4NF {sum+=1} END {print sum}' ${PATCH_DIR}/branch-ruby-lint-result.txt) # shellcheck disable=SC2016 numPostpatch=$(${AWK} -F: 'BEGIN {sum=0} 4NF {sum+=1} END {print sum}' ${PATCH_DIR}/patch-ruby-lint-result.txt) {code} This is because the diff result can contain multiple lines for one issue. For example: {code} [sekikn@mobile hadoop]$ cat /tmp/test-patch-hbase/1149/diff-patch-rubocop.txt bin/draining_servers.rb:165:1: C: Do not introduce global variables. $FOO bin/draining_servers.rb:166:1: C: Do not introduce global variables. $BAR {code} Checking the number of fields is intended to skip the second, third, fifth and sixth lines above. But four or more colons can appear in the source code itself, so more reliable way is needed. 2. pylint.sh has the same problem. In addition, I removed awk's '-F:' option by mistake on HADOOP-12286. It must report a wrong number for now. 3. rubocop.sh should also follow ruby-lint.sh. Or, change its formatter to emacs or simple style (currently clang). They are parser-friendly because they outputs one line by one issue, so we can count issues by 'wc -l' like perlcritic.sh. https://github.com/bbatsov/rubocop#emacs-style-formatter Fix some test-patch plugins to count the diff lines correctly - Key: HADOOP-12301 URL: https://issues.apache.org/jira/browse/HADOOP-12301 Project: Hadoop Common Issue Type: Sub-task Components: yetus Affects Versions: HADOOP-12111 Reporter: Kengo Seki 1. rubocop.sh counts only lines which have at least five fields separated by a colon: {code} calcdiffs ${PATCH_DIR}/branch-rubocop-result.txt ${PATCH_DIR}/patch-rubocop-result.txt ${PATCH_DIR}/diff-patch-rubocop.txt diffPostpatch=$(${AWK} -F: 'BEGIN {sum=0} 4NF {sum+=1} END {print sum}' ${PATCH_DIR}/diff-patch-rubocop.txt) if [[ ${diffPostpatch} -gt 0 ]] ; then # shellcheck disable=SC2016 numPrepatch=$(${AWK} -F: 'BEGIN {sum=0} 4NF {sum+=1} END {print sum}' ${PATCH_DIR}/branch-rubocop-result.txt) # shellcheck disable=SC2016 numPostpatch=$(${AWK} -F: 'BEGIN
[jira] [Commented] (HADOOP-12301) Fix some test-patch plugins to count the diff lines correctly
[ https://issues.apache.org/jira/browse/HADOOP-12301?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14708070#comment-14708070 ] Kengo Seki commented on HADOOP-12301: - I misunderstood ruby-lint output. It outputs only one line for one issue (I ran ruby-lint against all of the .rb files under HBase and confirmed it) and the current implementation is reasonable. I edited JIRA description so as to reflect reality. Fix some test-patch plugins to count the diff lines correctly - Key: HADOOP-12301 URL: https://issues.apache.org/jira/browse/HADOOP-12301 Project: Hadoop Common Issue Type: Sub-task Components: yetus Affects Versions: HADOOP-12111 Reporter: Kengo Seki 1. rubocop.sh counts only lines which have at least five fields separated by a colon: {code} calcdiffs ${PATCH_DIR}/branch-rubocop-result.txt ${PATCH_DIR}/patch-rubocop-result.txt ${PATCH_DIR}/diff-patch-rubocop.txt diffPostpatch=$(${AWK} -F: 'BEGIN {sum=0} 4NF {sum+=1} END {print sum}' ${PATCH_DIR}/diff-patch-rubocop.txt) if [[ ${diffPostpatch} -gt 0 ]] ; then # shellcheck disable=SC2016 numPrepatch=$(${AWK} -F: 'BEGIN {sum=0} 4NF {sum+=1} END {print sum}' ${PATCH_DIR}/branch-rubocop-result.txt) # shellcheck disable=SC2016 numPostpatch=$(${AWK} -F: 'BEGIN {sum=0} 4NF {sum+=1} END {print sum}' ${PATCH_DIR}/patch-rubocop-result.txt) {code} This is because the diff result can contain multiple lines for one issue. For example: {code} [sekikn@mobile hadoop]$ cat /private/tmp/test-patch-hbase/25821/diff-patch-rubocop.txt bin/draining_servers.rb:165:1: C: Do not introduce global variables. $foo {code} Checking the number of fields is intended to skip the second and third lines in the above diff file. But four or more colons can appear in the source code itself, for example: {code} | Vote | Subsystem | Runtime | Comment | -1 |rubocop | 0m 02s| The applied patch generated 4 new | | || rubocop issues (total was 77, now 81). || Subsystem || Report/Notes || | rubocop | v0.32.1 | | rubocop | /private/tmp/test-patch-hbase/5632/diff-patch-rubocop.txt | [sekikn@mobile hadoop]$ cat /private/tmp/test-patch-hbase/5632/diff-patch-rubocop.txt bin/draining_servers.rb:165:4: C: Do not use :: for method calls. foo::bar::baz ^^ bin/draining_servers.rb:165:9: C: Do not use :: for method calls. foo::bar::baz ^^ [sekikn@mobile hadoop]$ {code} In this case, new rubocop issues should be 2, but counted as 4 incorrectly. More reliable way to count is needed. 2. pylint.sh has the same problem. In addition, I removed awk's '-F:' option by mistake on HADOOP-12286. It can report a wrong number for now. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HADOOP-12301) Fix some test-patch plugins to count the diff lines correctly
[ https://issues.apache.org/jira/browse/HADOOP-12301?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14708080#comment-14708080 ] Kengo Seki commented on HADOOP-12301: - A pylint example. In this case, new pylint issues should be 6, but counted as 10 incorrectly. {code} | Vote | Subsystem | Runtime | Comment | -1 |pylint | 0m 05s| The applied patch generated 10 new | ||| pylint issues (total was 437, now 447). || Subsystem || Report/Notes || | pylint | v1.4.4 | | pylint | /private/tmp/test-patch-hadoop/40919/diff-patch-pylint.txt | [sekikn@mobile hadoop]$ cat /private/tmp/test-patch-hadoop/40919/diff-patch-pylint.txt dev-support/releasedocmaker.py:584: [W0311(bad-indentation), ] Bad indentation. Found 4 spaces, expected 2 dev-support/releasedocmaker.py:584: [C0326(bad-whitespace), ] No space allowed after bracket ( ( ) ) ^ dev-support/releasedocmaker.py:584: [C0326(bad-whitespace), ] No space allowed after bracket ( ( ) ) ^ dev-support/releasedocmaker.py:584: [C0326(bad-whitespace), ] No space allowed before bracket ( ( ) ) ^ dev-support/releasedocmaker.py:584: [C0326(bad-whitespace), ] No space allowed before bracket ( ( ) ) ^ dev-support/releasedocmaker.py:584: [W0104(pointless-statement), ] Statement seems to have no effect [sekikn@mobile hadoop]$ {code} Fix some test-patch plugins to count the diff lines correctly - Key: HADOOP-12301 URL: https://issues.apache.org/jira/browse/HADOOP-12301 Project: Hadoop Common Issue Type: Sub-task Components: yetus Affects Versions: HADOOP-12111 Reporter: Kengo Seki 1. rubocop.sh counts only lines which have at least five fields separated by a colon: {code} calcdiffs ${PATCH_DIR}/branch-rubocop-result.txt ${PATCH_DIR}/patch-rubocop-result.txt ${PATCH_DIR}/diff-patch-rubocop.txt diffPostpatch=$(${AWK} -F: 'BEGIN {sum=0} 4NF {sum+=1} END {print sum}' ${PATCH_DIR}/diff-patch-rubocop.txt) if [[ ${diffPostpatch} -gt 0 ]] ; then # shellcheck disable=SC2016 numPrepatch=$(${AWK} -F: 'BEGIN {sum=0} 4NF {sum+=1} END {print sum}' ${PATCH_DIR}/branch-rubocop-result.txt) # shellcheck disable=SC2016 numPostpatch=$(${AWK} -F: 'BEGIN {sum=0} 4NF {sum+=1} END {print sum}' ${PATCH_DIR}/patch-rubocop-result.txt) {code} This is because the diff result can contain multiple lines for one issue. For example: {code} [sekikn@mobile hadoop]$ cat /private/tmp/test-patch-hbase/25821/diff-patch-rubocop.txt bin/draining_servers.rb:165:1: C: Do not introduce global variables. $foo {code} Checking the number of fields is intended to skip the second and third lines in the above diff file. But four or more colons can appear in the source code itself, for example: {code} | Vote | Subsystem | Runtime | Comment | -1 |rubocop | 0m 02s| The applied patch generated 4 new | | || rubocop issues (total was 77, now 81). || Subsystem || Report/Notes || | rubocop | v0.32.1 | | rubocop | /private/tmp/test-patch-hbase/5632/diff-patch-rubocop.txt | [sekikn@mobile hadoop]$ cat /private/tmp/test-patch-hbase/5632/diff-patch-rubocop.txt bin/draining_servers.rb:165:4: C: Do not use :: for method calls. foo::bar::baz ^^ bin/draining_servers.rb:165:9: C: Do not use :: for method calls. foo::bar::baz ^^ [sekikn@mobile hadoop]$ {code} In this case, new rubocop issues should be 2, but counted as 4 incorrectly. More reliable way to count is needed. 2. pylint.sh has the same problem. In addition, I removed awk's '-F:' option by mistake on HADOOP-12286. It can report a wrong number for now. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HADOOP-12233) if CHANGED_FILES is corrupt, find_changed_modules never returns
[ https://issues.apache.org/jira/browse/HADOOP-12233?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Kengo Seki updated HADOOP-12233: Attachment: HADOOP-12233.HADOOP-12111.01.patch bq. Would something simple like verifying that we're still inside $\{BASEDIR} be sufficient? It sounds better because it's not unnatural and we can remove magic numbers from the source. Attaching a revised patch. If a changed file is not in $\{BASEDIR}, test-patch will fail as follows: {code} [sekikn@mobile hadoop]$ cat /tmp/test.patch diff --git /a/dev-support/releasedocmaker.py b/dev-support/releasedocmaker.py index 3c398be..05aa347 100755 --- a/dev-support/releasedocmaker.py +++ b/dev-support/releasedocmaker.py @@ -581,3 +581,4 @@ def main(): if __name__ == __main__: main() + [sekikn@mobile hadoop]$ dev-support/test-patch.sh --basedir=../dev/hadoop --project=hadoop /tmp/test.patch --debug (snip) [Sat Aug 22 13:55:42 JST 2015 DEBUG]: Find pom.xml dir for: /a/dev-support dev-support/test-patch.sh: line 1123: cd: /a: No such file or directory [Sat Aug 22 13:55:42 JST 2015 DEBUG]: ERROR: /a/dev-support is not in /Users/sekikn/dev/hadoop. ERROR: pom.xml is not found. Make sure the target is a maven-based project. -1 overall {code} Valid patches still pass: {code} [sekikn@mobile hadoop]$ cat /tmp/test.patch diff --git a/dev-support/releasedocmaker.py b/dev-support/releasedocmaker.py index 3c398be..05aa347 100755 --- a/dev-support/releasedocmaker.py +++ b/dev-support/releasedocmaker.py @@ -581,3 +581,4 @@ def main(): if __name__ == __main__: main() + [sekikn@mobile hadoop]$ dev-support/test-patch.sh --basedir=../dev/hadoop --project=hadoop /tmp/test.patch (snip) +1 overall {code} if CHANGED_FILES is corrupt, find_changed_modules never returns --- Key: HADOOP-12233 URL: https://issues.apache.org/jira/browse/HADOOP-12233 Project: Hadoop Common Issue Type: Sub-task Components: yetus Affects Versions: HADOOP-12111 Reporter: Allen Wittenauer Assignee: Kengo Seki Attachments: HADOOP-12233.HADOOP-12111.00.patch, HADOOP-12233.HADOOP-12111.01.patch In building some unit tests, did a negative tests and hit this condition. We should put a limit on how many times we loop in the find_x_dirs code. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HADOOP-12061) Incorrect command in 2.7.0 single cluster setup document
[ https://issues.apache.org/jira/browse/HADOOP-12061?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Kengo Seki updated HADOOP-12061: Attachment: HADOOP-12061.branch-2.002.patch Sure. Would you try this one? Incorrect command in 2.7.0 single cluster setup document Key: HADOOP-12061 URL: https://issues.apache.org/jira/browse/HADOOP-12061 Project: Hadoop Common Issue Type: Bug Components: documentation Affects Versions: 2.7.0 Reporter: Kengo Seki Assignee: Kengo Seki Priority: Minor Labels: newbie Attachments: HADOOP-12061.001.patch, HADOOP-12061.002.patch, HADOOP-12061.branch-2.002.patch There seems to be a trivial fault in the single node cluster setup document on branch-2. Setup passphraseless ssh says as follows: {code} $ ssh-keygen -t dsa -P '' -f ~/.ssh/id_dsa $ cat ~/.ssh/id_dsa.pub ~/.ssh/authorized_keys $ export HADOOP\_PREFIX=/usr/local/hadoop {code} But it should be: {code} $ ssh-keygen -t dsa -P '' -f ~/.ssh/id_dsa $ cat ~/.ssh/id_dsa.pub ~/.ssh/authorized_keys $ chmod 0700 ~/.ssh/authorized_keys {code} -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Created] (HADOOP-12340) test-patch docker mode fails in downloading findbugs with curl
Kengo Seki created HADOOP-12340: --- Summary: test-patch docker mode fails in downloading findbugs with curl Key: HADOOP-12340 URL: https://issues.apache.org/jira/browse/HADOOP-12340 Project: Hadoop Common Issue Type: Sub-task Components: yetus Affects Versions: HADOOP-12111 Reporter: Kengo Seki HADOOP-12129 replaced wget commands in test-patch with curl. But curl doesn't follow URL redirection by default, so docker mode fails in downloading findbugs for now: {code} $ dev-support/test-patch.sh --docker --project=hadoop /tmp/test.patch (snip) Step 11 : RUN mkdir -p /opt/findbugs curl https://sourceforge.net/projects/findbugs/files/findbugs/3.0.1/findbugs-noUpdateChecks-3.0.1.tar.gz/download -o /opt/findbugs.tar.gz tar xzf /opt/findbugs.tar.gz --strip-components 1 -C /opt/findbugs --- Running in 31c71013bb59 % Total% Received % Xferd Average Speed TimeTime Time Current Dload Upload Total SpentLeft Speed 100 397 100 3970 0168 0 0:00:02 0:00:02 --:--:-- 168 gzip: stdin: not in gzip format tar: Child returned status 1 tar: Error is not recoverable: exiting now The command '/bin/sh -c mkdir -p /opt/findbugs curl https://sourceforge.net/projects/findbugs/files/findbugs/3.0.1/findbugs-noUpdateChecks-3.0.1.tar.gz/download -o /opt/findbugs.tar.gz tar xzf /opt/findbugs.tar.gz --strip-components 1 -C /opt/findbugs' returned a non-zero code: 2 {code} Adding -L option will resolve this. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Created] (HADOOP-12339) Wrong help message about --curl-cmd option
Kengo Seki created HADOOP-12339: --- Summary: Wrong help message about --curl-cmd option Key: HADOOP-12339 URL: https://issues.apache.org/jira/browse/HADOOP-12339 Project: Hadoop Common Issue Type: Sub-task Components: yetus Affects Versions: HADOOP-12111 Reporter: Kengo Seki Priority: Minor {code} --curl-cmd=cmd The 'wget' command to use (default 'curl') {code} :) -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HADOOP-12233) if CHANGED_FILES is corrupt, find_changed_modules never returns
[ https://issues.apache.org/jira/browse/HADOOP-12233?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Kengo Seki updated HADOOP-12233: Attachment: HADOOP-12233.HADOOP-12111.00.patch -00: * put a limit on the loops in find_buildfile_dir and module_skipdir in the same way as find_changed_modules I confirmed that both of them stop against some artificial patch (without going into detail so that a malicious patch starving jenkins is not submitted easily) with the following message: {code} ERROR: pom.xml is not found. Make sure the target is a maven-based project. {code} if CHANGED_FILES is corrupt, find_changed_modules never returns --- Key: HADOOP-12233 URL: https://issues.apache.org/jira/browse/HADOOP-12233 Project: Hadoop Common Issue Type: Sub-task Components: yetus Affects Versions: HADOOP-12111 Reporter: Allen Wittenauer Attachments: HADOOP-12233.HADOOP-12111.00.patch In building some unit tests, did a negative tests and hit this condition. We should put a limit on how many times we loop in the find_x_dirs code. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HADOOP-12303) test-patch pylint plugin fails silently and votes +1 incorrectly
[ https://issues.apache.org/jira/browse/HADOOP-12303?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Kengo Seki updated HADOOP-12303: Assignee: Kengo Seki Status: Patch Available (was: Open) test-patch pylint plugin fails silently and votes +1 incorrectly Key: HADOOP-12303 URL: https://issues.apache.org/jira/browse/HADOOP-12303 Project: Hadoop Common Issue Type: Sub-task Components: yetus Affects Versions: HADOOP-12111 Reporter: Kengo Seki Assignee: Kengo Seki Attachments: HADOOP-12303.HADOOP-12111.00.patch This patch {code} [sekikn@mobile hadoop]$ cat /tmp/test.patch diff --git a/dev-support/releasedocmaker.py b/dev-support/releasedocmaker.py index 37bd58a..7cd6dd3 100755 --- a/dev-support/releasedocmaker.py +++ b/dev-support/releasedocmaker.py @@ -580,4 +580,4 @@ def main(): sys.exit(1) if __name__ == __main__: -main() +main( ) {code} is supposed to cause the following pylint errors. {code} C:583, 0: No space allowed after bracket main( ) ^ (bad-whitespace) C:583, 0: No space allowed before bracket main( ) ^ (bad-whitespace) {code} But the system locale is set as follows, pylint check is passed, and there is no pylint output. {code} [sekikn@mobile hadoop]$ locale LANG= LC_COLLATE=C LC_CTYPE=UTF-8 LC_MESSAGES=C LC_MONETARY=C LC_NUMERIC=C LC_TIME=C LC_ALL= [sekikn@mobile hadoop]$ dev-support/test-patch.sh --basedir=/Users/sekikn/dev/hadoop --project=hadoop /tmp/test.patch (snip) | Vote | Subsystem | Runtime | Comment | +1 | @author | 0m 00s| The patch does not contain any @author | ||| tags. | +1 |asflicense | 0m 21s| Patch does not generate ASF License | ||| warnings. | +1 |pylint | 0m 01s| There were no new pylint issues. | +1 |whitespace | 0m 00s| Patch has no whitespace issues. | || 0m 24s| (snip) [sekikn@mobile hadoop]$ cat /private/tmp/test-patch-hadoop/8656/branch-pylint-result.txt [sekikn@mobile hadoop]$ cat /private/tmp/test-patch-hadoop/8656/patch-pylint-result.txt [sekikn@mobile hadoop]$ cat /private/tmp/test-patch-hadoop/8656/diff-patch-pylint.txt {code} Removing '2/dev/null' from pylint.sh reveals the root cause. Setting LC_ALL or LC_CTYPE such like 'en_US.UTF-8' solves this problem. {code} pylint plugin: prepatch Running pylint against modified python scripts. No config file found, using default configuration Traceback (most recent call last): File /usr/local/bin/pylint, line 11, in module sys.exit(run_pylint()) File /Library/Python/2.7/site-packages/pylint/__init__.py, line 23, in run_pylint Run(sys.argv[1:]) File /Library/Python/2.7/site-packages/pylint/lint.py, line 1332, in __init__ linter.check(args) File /Library/Python/2.7/site-packages/pylint/lint.py, line 747, in check self._do_check(files_or_modules) File /Library/Python/2.7/site-packages/pylint/lint.py, line 869, in _do_check self.check_astroid_module(ast_node, walker, rawcheckers, tokencheckers) File /Library/Python/2.7/site-packages/pylint/lint.py, line 944, in check_astroid_module checker.process_tokens(tokens) File /Library/Python/2.7/site-packages/pylint/checkers/format.py, line 743, in process_tokens self.check_indent_level(token, indents[-1]+1, line_num) File /Library/Python/2.7/site-packages/pylint/checkers/format.py, line 963, in check_indent_level expected * unit_size)) File /Library/Python/2.7/site-packages/pylint/checkers/__init__.py, line 101, in add_message self.linter.add_message(msg_id, line, node, args, confidence) File /Library/Python/2.7/site-packages/pylint/utils.py, line 410, in add_message (abspath, path, module, obj, line or 1, col_offset or 0), msg, confidence)) File /Library/Python/2.7/site-packages/pylint/reporters/text.py, line 61, in handle_message self.write_message(msg) File /Library/Python/2.7/site-packages/pylint/reporters/text.py, line 51, in write_message self.writeln(msg.format(self._template)) File /Library/Python/2.7/site-packages/pylint/reporters/__init__.py, line 94, in
[jira] [Updated] (HADOOP-12233) if CHANGED_FILES is corrupt, find_changed_modules never returns
[ https://issues.apache.org/jira/browse/HADOOP-12233?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Kengo Seki updated HADOOP-12233: Assignee: Kengo Seki Status: Patch Available (was: Open) if CHANGED_FILES is corrupt, find_changed_modules never returns --- Key: HADOOP-12233 URL: https://issues.apache.org/jira/browse/HADOOP-12233 Project: Hadoop Common Issue Type: Sub-task Components: yetus Affects Versions: HADOOP-12111 Reporter: Allen Wittenauer Assignee: Kengo Seki Attachments: HADOOP-12233.HADOOP-12111.00.patch In building some unit tests, did a negative tests and hit this condition. We should put a limit on how many times we loop in the find_x_dirs code. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HADOOP-12303) test-patch pylint plugin fails silently and votes +1 incorrectly
[ https://issues.apache.org/jira/browse/HADOOP-12303?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Kengo Seki updated HADOOP-12303: Attachment: HADOOP-12303.HADOOP-12111.00.patch Attaching a patch. I confirmed the pylint plugin detects the above case correctly and passes some simple regression tests. {code} [sekikn@mobile hadoop]$ cat /tmp/test.patch diff --git a/dev-support/releasedocmaker.py b/dev-support/releasedocmaker.py index 3c398be..6c862ee 100755 --- a/dev-support/releasedocmaker.py +++ b/dev-support/releasedocmaker.py @@ -580,4 +580,4 @@ def main(): sys.exit(1) if __name__ == __main__: -main() +main( ) [sekikn@mobile hadoop]$ locale LANG= LC_COLLATE=C LC_CTYPE=UTF-8 LC_MESSAGES=C LC_MONETARY=C LC_NUMERIC=C LC_TIME=C LC_ALL= [sekikn@mobile hadoop]$ dev-support/test-patch.sh --basedir=/Users/sekikn/dev/hadoop --project=hadoop /tmp/test.patch (snip) | Vote | Subsystem | Runtime | Comment | +1 | @author | 0m 00s| The patch does not contain any @author | ||| tags. | +1 |asflicense | 0m 31s| Patch does not generate ASF License | ||| warnings. | -1 |pylint | 0m 00s| Something bad seems to have happened in | ||| running pylint. Please check pylint | ||| stderr files. | +1 |whitespace | 0m 00s| Patch has no whitespace issues. | || 0m 32s| || Subsystem || Report/Notes || | git revision | HADOOP-12111 / 565d9bf | | Optional Tests | asflicense pylint | | uname | Darwin mobile.local 14.4.0 Darwin Kernel Version 14.4.0: Thu May 28 11:35:04 PDT 2015; root:xnu-2782.30.5~1/RELEASE_X86_64 x86_64 | | Build tool | maven | | Personality | /Users/sekikn/hadoop/dev-support/personality/hadoop.sh | | Default Java | 1.7.0_80 | | pylint | prepatch stderr: /private/tmp/test-patch-hadoop/48440/pylint.48440.19203 | | pylint | postpatch stderr: /private/tmp/test-patch-hadoop/48440/pylint.48440.3046 | | Max memory used | 46MB | Finished build. [sekikn@mobile hadoop]$ cat /private/tmp/test-patch-hadoop/48440/pylint.48440.3046 No config file found, using default configuration Traceback (most recent call last): File /usr/local/bin/pylint, line 11, in module sys.exit(run_pylint()) File /Library/Python/2.7/site-packages/pylint/__init__.py, line 23, in run_pylint Run(sys.argv[1:]) File /Library/Python/2.7/site-packages/pylint/lint.py, line 1332, in __init__ linter.check(args) File /Library/Python/2.7/site-packages/pylint/lint.py, line 747, in check self._do_check(files_or_modules) File /Library/Python/2.7/site-packages/pylint/lint.py, line 869, in _do_check self.check_astroid_module(ast_node, walker, rawcheckers, tokencheckers) File /Library/Python/2.7/site-packages/pylint/lint.py, line 944, in check_astroid_module checker.process_tokens(tokens) File /Library/Python/2.7/site-packages/pylint/checkers/format.py, line 743, in process_tokens self.check_indent_level(token, indents[-1]+1, line_num) File /Library/Python/2.7/site-packages/pylint/checkers/format.py, line 963, in check_indent_level expected * unit_size)) File /Library/Python/2.7/site-packages/pylint/checkers/__init__.py, line 101, in add_message self.linter.add_message(msg_id, line, node, args, confidence) File /Library/Python/2.7/site-packages/pylint/utils.py, line 410, in add_message (abspath, path, module, obj, line or 1, col_offset or 0), msg, confidence)) File /Library/Python/2.7/site-packages/pylint/reporters/text.py, line 61, in handle_message self.write_message(msg) File /Library/Python/2.7/site-packages/pylint/reporters/text.py, line 51, in write_message self.writeln(msg.format(self._template)) File /Library/Python/2.7/site-packages/pylint/reporters/__init__.py, line 94, in writeln print(self.encode(string), file=self.out) File /Library/Python/2.7/site-packages/pylint/reporters/__init__.py, line 84, in encode locale.getdefaultlocale()[1] or File /System/Library/Frameworks/Python.framework/Versions/2.7/lib/python2.7/locale.py, line 511, in getdefaultlocale return _parse_localename(localename) File /System/Library/Frameworks/Python.framework/Versions/2.7/lib/python2.7/locale.py, line 443, in _parse_localename raise ValueError, 'unknown locale: %s' %
[jira] [Updated] (HADOOP-12303) test-patch pylint plugin fails silently and votes +1 incorrectly
[ https://issues.apache.org/jira/browse/HADOOP-12303?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Kengo Seki updated HADOOP-12303: Attachment: HADOOP-12303.HADOOP-12111.01.patch -01: * address shellcheck issues test-patch pylint plugin fails silently and votes +1 incorrectly Key: HADOOP-12303 URL: https://issues.apache.org/jira/browse/HADOOP-12303 Project: Hadoop Common Issue Type: Sub-task Components: yetus Affects Versions: HADOOP-12111 Reporter: Kengo Seki Assignee: Kengo Seki Attachments: HADOOP-12303.HADOOP-12111.00.patch, HADOOP-12303.HADOOP-12111.01.patch This patch {code} [sekikn@mobile hadoop]$ cat /tmp/test.patch diff --git a/dev-support/releasedocmaker.py b/dev-support/releasedocmaker.py index 37bd58a..7cd6dd3 100755 --- a/dev-support/releasedocmaker.py +++ b/dev-support/releasedocmaker.py @@ -580,4 +580,4 @@ def main(): sys.exit(1) if __name__ == __main__: -main() +main( ) {code} is supposed to cause the following pylint errors. {code} C:583, 0: No space allowed after bracket main( ) ^ (bad-whitespace) C:583, 0: No space allowed before bracket main( ) ^ (bad-whitespace) {code} But the system locale is set as follows, pylint check is passed, and there is no pylint output. {code} [sekikn@mobile hadoop]$ locale LANG= LC_COLLATE=C LC_CTYPE=UTF-8 LC_MESSAGES=C LC_MONETARY=C LC_NUMERIC=C LC_TIME=C LC_ALL= [sekikn@mobile hadoop]$ dev-support/test-patch.sh --basedir=/Users/sekikn/dev/hadoop --project=hadoop /tmp/test.patch (snip) | Vote | Subsystem | Runtime | Comment | +1 | @author | 0m 00s| The patch does not contain any @author | ||| tags. | +1 |asflicense | 0m 21s| Patch does not generate ASF License | ||| warnings. | +1 |pylint | 0m 01s| There were no new pylint issues. | +1 |whitespace | 0m 00s| Patch has no whitespace issues. | || 0m 24s| (snip) [sekikn@mobile hadoop]$ cat /private/tmp/test-patch-hadoop/8656/branch-pylint-result.txt [sekikn@mobile hadoop]$ cat /private/tmp/test-patch-hadoop/8656/patch-pylint-result.txt [sekikn@mobile hadoop]$ cat /private/tmp/test-patch-hadoop/8656/diff-patch-pylint.txt {code} Removing '2/dev/null' from pylint.sh reveals the root cause. Setting LC_ALL or LC_CTYPE such like 'en_US.UTF-8' solves this problem. {code} pylint plugin: prepatch Running pylint against modified python scripts. No config file found, using default configuration Traceback (most recent call last): File /usr/local/bin/pylint, line 11, in module sys.exit(run_pylint()) File /Library/Python/2.7/site-packages/pylint/__init__.py, line 23, in run_pylint Run(sys.argv[1:]) File /Library/Python/2.7/site-packages/pylint/lint.py, line 1332, in __init__ linter.check(args) File /Library/Python/2.7/site-packages/pylint/lint.py, line 747, in check self._do_check(files_or_modules) File /Library/Python/2.7/site-packages/pylint/lint.py, line 869, in _do_check self.check_astroid_module(ast_node, walker, rawcheckers, tokencheckers) File /Library/Python/2.7/site-packages/pylint/lint.py, line 944, in check_astroid_module checker.process_tokens(tokens) File /Library/Python/2.7/site-packages/pylint/checkers/format.py, line 743, in process_tokens self.check_indent_level(token, indents[-1]+1, line_num) File /Library/Python/2.7/site-packages/pylint/checkers/format.py, line 963, in check_indent_level expected * unit_size)) File /Library/Python/2.7/site-packages/pylint/checkers/__init__.py, line 101, in add_message self.linter.add_message(msg_id, line, node, args, confidence) File /Library/Python/2.7/site-packages/pylint/utils.py, line 410, in add_message (abspath, path, module, obj, line or 1, col_offset or 0), msg, confidence)) File /Library/Python/2.7/site-packages/pylint/reporters/text.py, line 61, in handle_message self.write_message(msg) File /Library/Python/2.7/site-packages/pylint/reporters/text.py, line 51, in write_message self.writeln(msg.format(self._template)) File
[jira] [Commented] (HADOOP-12303) test-patch pylint plugin fails silently and votes +1 incorrectly
[ https://issues.apache.org/jira/browse/HADOOP-12303?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14695679#comment-14695679 ] Kengo Seki commented on HADOOP-12303: - In the above situation, pylint returns 1. {code} [sekikn@mobile hadoop]$ pylint dev-support/releasedocmaker.py /dev/null No config file found, using default configuration Traceback (most recent call last): File /usr/local/bin/pylint, line 11, in module sys.exit(run_pylint()) File /Library/Python/2.7/site-packages/pylint/__init__.py, line 23, in run_pylint Run(sys.argv[1:]) File /Library/Python/2.7/site-packages/pylint/lint.py, line 1332, in __init__ linter.check(args) File /Library/Python/2.7/site-packages/pylint/lint.py, line 747, in check self._do_check(files_or_modules) File /Library/Python/2.7/site-packages/pylint/lint.py, line 869, in _do_check self.check_astroid_module(ast_node, walker, rawcheckers, tokencheckers) File /Library/Python/2.7/site-packages/pylint/lint.py, line 944, in check_astroid_module checker.process_tokens(tokens) File /Library/Python/2.7/site-packages/pylint/checkers/format.py, line 727, in process_tokens self.new_line(TokenWrapper(tokens), idx-1, idx+1) File /Library/Python/2.7/site-packages/pylint/checkers/format.py, line 473, in new_line self.check_lines(line, line_num) File /Library/Python/2.7/site-packages/pylint/checkers/format.py, line 932, in check_lines self.add_message('line-too-long', line=i, args=(len(line), max_chars)) File /Library/Python/2.7/site-packages/pylint/checkers/__init__.py, line 101, in add_message self.linter.add_message(msg_id, line, node, args, confidence) File /Library/Python/2.7/site-packages/pylint/utils.py, line 410, in add_message (abspath, path, module, obj, line or 1, col_offset or 0), msg, confidence)) File /Library/Python/2.7/site-packages/pylint/reporters/text.py, line 61, in handle_message self.write_message(msg) File /Library/Python/2.7/site-packages/pylint/reporters/text.py, line 51, in write_message self.writeln(msg.format(self._template)) File /Library/Python/2.7/site-packages/pylint/reporters/__init__.py, line 94, in writeln print(self.encode(string), file=self.out) File /Library/Python/2.7/site-packages/pylint/reporters/__init__.py, line 84, in encode locale.getdefaultlocale()[1] or File /System/Library/Frameworks/Python.framework/Versions/2.7/lib/python2.7/locale.py, line 511, in getdefaultlocale return _parse_localename(localename) File /System/Library/Frameworks/Python.framework/Versions/2.7/lib/python2.7/locale.py, line 443, in _parse_localename raise ValueError, 'unknown locale: %s' % localename ValueError: unknown locale: UTF-8 [sekikn@mobile hadoop]$ echo $? 1 {code} But pylint can also return 1 even if linting succeeded. {code} [sekikn@mobile hadoop]$ cat a.py a.py import simplejson print dir(simplejson) [sekikn@mobile hadoop]$ pylint --reports=n a.py No config file found, using default configuration * Module a F: 2, 0: Unable to import 'simplejson' (import-error) [sekikn@mobile hadoop]$ echo $? 1 {code} So the status code doesn't help, though this example is somewhat artificial. Probably we need to distinguish them by the stderr contents. test-patch pylint plugin fails silently and votes +1 incorrectly Key: HADOOP-12303 URL: https://issues.apache.org/jira/browse/HADOOP-12303 Project: Hadoop Common Issue Type: Sub-task Components: yetus Affects Versions: HADOOP-12111 Reporter: Kengo Seki This patch {code} [sekikn@mobile hadoop]$ cat /tmp/test.patch diff --git a/dev-support/releasedocmaker.py b/dev-support/releasedocmaker.py index 37bd58a..7cd6dd3 100755 --- a/dev-support/releasedocmaker.py +++ b/dev-support/releasedocmaker.py @@ -580,4 +580,4 @@ def main(): sys.exit(1) if __name__ == __main__: -main() +main( ) {code} is supposed to cause the following pylint errors. {code} C:583, 0: No space allowed after bracket main( ) ^ (bad-whitespace) C:583, 0: No space allowed before bracket main( ) ^ (bad-whitespace) {code} But the system locale is set as follows, pylint check is passed, and there is no pylint output. {code} [sekikn@mobile hadoop]$ locale LANG= LC_COLLATE=C LC_CTYPE=UTF-8 LC_MESSAGES=C LC_MONETARY=C LC_NUMERIC=C LC_TIME=C LC_ALL= [sekikn@mobile hadoop]$ dev-support/test-patch.sh --basedir=/Users/sekikn/dev/hadoop --project=hadoop /tmp/test.patch (snip) | Vote | Subsystem | Runtime | Comment | +1 | @author | 0m 00s| The patch does not contain any @author | ||| tags.
[jira] [Updated] (HADOOP-12314) check_unittests in test-patch.sh can return a wrong status
[ https://issues.apache.org/jira/browse/HADOOP-12314?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Kengo Seki updated HADOOP-12314: Assignee: Kengo Seki Status: Patch Available (was: Open) check_unittests in test-patch.sh can return a wrong status -- Key: HADOOP-12314 URL: https://issues.apache.org/jira/browse/HADOOP-12314 Project: Hadoop Common Issue Type: Sub-task Components: yetus Affects Versions: HADOOP-12111 Reporter: Kengo Seki Assignee: Kengo Seki Attachments: HADOOP-12314.HADOOP-12111.00.patch Follow-up from HADOOP-12247. check_unittests returns the value of $\{result}, but the status of *_process_tests is added to $\{results}. -- This message was sent by Atlassian JIRA (v6.3.4#6332)