GitHub user zentol opened a pull request:

    https://github.com/apache/flink/pull/1703

    [FLINK-3491] Prevent failure of HDFSCopyUtilitiesTest on Windows

    This PR contains two commits to that prevent this test from failing on 
Windows.
    
    The first commit resolves the problem raised in the JIRA: it changes how 
the URI is generated, using new Path(file).toUri() instead of file.toUri(), 
since the latter fails for Windows paths.
    
    After this change, i got a new exception when running the test:
    ```
    testCopyFromLocal(org.apache.flink.streaming.util.HDFSCopyUtilitiesTest)  
Time elapsed: 1.892 sec  <<< ERROR!
    java.lang.NullPointerException: null
            at java.lang.ProcessBuilder.start(ProcessBuilder.java:1012)
            at org.apache.hadoop.util.Shell.runCommand(Shell.java:445)
            at org.apache.hadoop.util.Shell.run(Shell.java:418)
            at 
org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:650)
            at org.apache.hadoop.util.Shell.execCommand(Shell.java:739)
            at org.apache.hadoop.util.Shell.execCommand(Shell.java:722)
            at 
org.apache.hadoop.fs.RawLocalFileSystem.setPermission(RawLocalFileSystem.java:631)
            at 
org.apache.hadoop.fs.FilterFileSystem.setPermission(FilterFileSystem.java:468)
            at 
org.apache.hadoop.fs.ChecksumFileSystem.create(ChecksumFileSystem.java:456)
            at 
org.apache.hadoop.fs.ChecksumFileSystem.create(ChecksumFileSystem.java:424)
            at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:907)
            at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:888)
            at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:785)
            at org.apache.hadoop.fs.FileUtil.copy(FileUtil.java:365)
            at org.apache.hadoop.fs.FileUtil.copy(FileUtil.java:338)
            at org.apache.hadoop.fs.FileUtil.copy(FileUtil.java:289)
            at 
org.apache.hadoop.fs.LocalFileSystem.copyFromLocalFile(LocalFileSystem.java:82)
            at 
org.apache.hadoop.fs.FileSystem.copyFromLocalFile(FileSystem.java:1837)
            at 
org.apache.flink.streaming.util.HDFSCopyFromLocal$1.run(HDFSCopyFromLocal.java:49)
    
    ```
    
    After googling a bit it appears that several hadoop versions can't deal 
with windows paths unless it has access to fancy libraries/dll's. (see 
https://issues.apache.org/jira/browse/SPARK-6961) As such I added a  second 
commit that disables the test when run on windows. 

You can merge this pull request into a Git repository by running:

    $ git pull https://github.com/zentol/flink 3491_test_hdfscopy

Alternatively you can review and apply these changes as the patch at:

    https://github.com/apache/flink/pull/1703.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

    This closes #1703
    
----
commit 124da4220968ce331ebd9c45cdf35e4a4074848b
Author: zentol <s.mo...@web.de>
Date:   2016-02-24T10:14:23Z

    [FLINK-3491] Prevent URIException in HDFSCopyTest

commit c458f305ad23f9f075993133ffcf5ed5dd606eb4
Author: zentol <s.mo...@web.de>
Date:   2016-02-24T10:14:52Z

    Disable HDFSCopyUtilitiesTest on Windows

----


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---

Reply via email to