Hello Gary,

I wanted to reproduce your problem, but had problems with the line numbers in 
the stack trace. Can you check why you have different ones? When I check it on 
my system the line numbers match the 1.2.1 sourcde. And if I actually disable 
stack-trace-trimming(commited) in surefire, it actually prints a helpful error:

...
Caused by: java.lang.RuntimeException: Error while running command to get file 
permissions : java.io.IOException: Cannot run program "ls": CreateProcess 
error=2, Das System kann die angegebene Datei nicht finden
        at java.lang.ProcessBuilder.start(ProcessBuilder.java:1041)
        at org.apache.hadoop.util.Shell.runCommand(Shell.java:200)
        at org.apache.hadoop.util.Shell.run(Shell.java:182)
        at 
org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:375)
        at org.apache.hadoop.util.Shell.execCommand(Shell.java:461)
        at org.apache.hadoop.util.Shell.execCommand(Shell.java:444)
        at org.apache.hadoop.fs.FileUtil.execCommand(FileUtil.java:712)
        at 
org.apache.hadoop.fs.RawLocalFileSystem$RawLocalFileStatus.loadPermissionInfo(RawLocalFileSystem.java:448)
        at 
org.apache.hadoop.fs.RawLocalFileSystem$RawLocalFileStatus.getPermission(RawLocalFileSystem.java:423)
        at 
org.apache.hadoop.util.DiskChecker.mkdirsWithExistsAndPermissionCheck(DiskChecker.java:146)
        at org.apache.hadoop.util.DiskChecker.checkDir(DiskChecker.java:162)
        at 
org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:1704)
        at 
org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:1651)
        at 
org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:1626)
        at 
org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.java:421)
        at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:284)
        at 
org.apache.commons.vfs2.provider.hdfs.test.HdfsFileProviderTest.<clinit>(HdfsFileProviderTest.java:95)
...

And it actually means, that "ls.exe" is not in the PATH. So yes, the test does 
not work on all Windows systems, it requires at least a ls.exe. I will remove 
the automatic running of those tests on the Windows platform (again). (but with 
a better named profile).

As a quick fix it should be enough to add any ls.exe, in my case it was the 
portable git distribution (from github):

%LOCALAPPDATA%\GitHub\PortableGit_015aa71ef18c047ce8509ffb2f9e4bb0e3e73f13\bin\ls.exe

Gruss
Bernd










Am Tue, 10 Jun 2014 11:02:19 -0400 schrieb Gary Gregory
<garydgreg...@gmail.com>:

> woa... cygwin? I have that installed but it does not help.
> 
> How about this: can please you turn off HDFS testing on Windows like
> it was before.
> 
> I'll be happy to test patches for you on my set up, Windows 7
> Professional 64-bit Service Pack 1.
> 
> My understanding is that the HDFS jars we use do not run on Windows
> out of the box because they rely on calling OS commands that are *Nix
> specific.
> 
> Gary
> 
> 
> On Tue, Jun 10, 2014 at 9:45 AM, Bernd Eckenfels
> <e...@zusammenkunft.net> wrote:
> 
> > Hello,
> >
> > they do work for me, hm. Windows 7 x64 de. I will try some other
> > environments. Maybe it picks up some cygwin stuff or something on my
> > system?
> >
> > Gruss
> > Bernd
> >
> >
> > Am Mon, 9 Jun 2014 18:05:18 -0400
> > schrieb Gary Gregory <garydgreg...@gmail.com>:
> >
> > > Ecki enabled the HDFS tests on Windows but they sure fail for me,
> > > see below.
> > >
> > > Do they work for anyone else on Windows?
> > >
> > > Tests run: 2, Failures: 0, Errors: 2, Skipped: 0, Time elapsed:
> > > 3.681 sec <<< FAILURE! - in
> > > org.apache.commons.vfs2.provider.hdfs.test.HdfsFileProviderTest
> > > org.apache.commons.vfs2.provider.hdfs.test.HdfsFileProviderTest
> > > Time elapsed: 3.68 sec  <<< ERROR!
> > > java.lang.ExceptionInInitializerError: null
> > >         at
> > >
> > org.apache.hadoop.fs.RawLocalFileSystem$RawLocalFileStatus.loadPermissionInfo(RawLocalFileSystem.java:468)
> > >         at
> > >
> > org.apache.hadoop.fs.RawLocalFileSystem$RawLocalFileStatus.getPermission(RawLocalFileSystem.java:418)
> > >         at
> > >
> > org.apache.hadoop.util.DiskChecker.mkdirsWithExistsAndPermissionCheck(DiskChecker.java:146)
> > >         at
> > > org.apache.hadoop.util.DiskChecker.checkDir(DiskChecker.java:162)
> > > at
> > >
> > org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:1643)
> > >         at
> > >
> > org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:1590)
> > >         at
> > >
> > org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:1565)
> > >         at
> > >
> > org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.java:421)
> > >         at
> > > org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:284)
> > >         at
> > >
> > org.apache.commons.vfs2.provider.hdfs.test.HdfsFileProviderTest.<clinit>(HdfsFileProviderTest.java:95)
> > >
> > > org.apache.commons.vfs2.provider.hdfs.test.HdfsFileProviderTest
> > > Time elapsed: 3.681 sec  <<< ERROR!
> > > java.lang.NoClassDefFoundError: Could not initialize class
> > > org.apache.commons.vfs2.provider.hdfs.test.HdfsFileProviderTest
> > >         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native
> > > Method) at
> > >
> > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
> > >         at
> > >
> > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> > >         at java.lang.reflect.Method.invoke(Method.java:606)
> > >         at
> > >
> > org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47)
> > >         at
> > >
> > org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
> > >         at
> > >
> > org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44)
> > >         at
> > >
> > org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33)
> > >         at
> > > org.junit.runners.ParentRunner.run(ParentRunner.java:309) at
> > >
> > org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:264)
> > >         at
> > >
> > org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:153)
> > >         at
> > >
> > org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:124)
> > >         at
> > >
> > org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoader(ForkedBooter.java:200)
> > >         at
> > >
> > org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:153)
> > >         at
> > > org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:103)
> > >
> > > Gary
> > >
> > >
> >
> >
> > ---------------------------------------------------------------------
> > To unsubscribe, e-mail: dev-unsubscr...@commons.apache.org
> > For additional commands, e-mail: dev-h...@commons.apache.org
> >
> >
> 
> 


---------------------------------------------------------------------
To unsubscribe, e-mail: dev-unsubscr...@commons.apache.org
For additional commands, e-mail: dev-h...@commons.apache.org

Reply via email to