The build still breaks on Windows. Can you fix it please?

Gary


On Mon, Jun 16, 2014 at 8:23 PM, Bernd Eckenfels <e...@zusammenkunft.net>
wrote:

> Hello Gary,
>
> I wanted to reproduce your problem, but had problems with the line numbers
> in the stack trace. Can you check why you have different ones? When I check
> it on my system the line numbers match the 1.2.1 sourcde. And if I actually
> disable stack-trace-trimming(commited) in surefire, it actually prints a
> helpful error:
>
> ...
> Caused by: java.lang.RuntimeException: Error while running command to get
> file permissions : java.io.IOException: Cannot run program "ls":
> CreateProcess error=2, Das System kann die angegebene Datei nicht finden
>         at java.lang.ProcessBuilder.start(ProcessBuilder.java:1041)
>         at org.apache.hadoop.util.Shell.runCommand(Shell.java:200)
>         at org.apache.hadoop.util.Shell.run(Shell.java:182)
>         at
> org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:375)
>         at org.apache.hadoop.util.Shell.execCommand(Shell.java:461)
>         at org.apache.hadoop.util.Shell.execCommand(Shell.java:444)
>         at org.apache.hadoop.fs.FileUtil.execCommand(FileUtil.java:712)
>         at
> org.apache.hadoop.fs.RawLocalFileSystem$RawLocalFileStatus.loadPermissionInfo(RawLocalFileSystem.java:448)
>         at
> org.apache.hadoop.fs.RawLocalFileSystem$RawLocalFileStatus.getPermission(RawLocalFileSystem.java:423)
>         at
> org.apache.hadoop.util.DiskChecker.mkdirsWithExistsAndPermissionCheck(DiskChecker.java:146)
>         at
> org.apache.hadoop.util.DiskChecker.checkDir(DiskChecker.java:162)
>         at
> org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:1704)
>         at
> org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:1651)
>         at
> org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:1626)
>         at
> org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.java:421)
>         at
> org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:284)
>         at
> org.apache.commons.vfs2.provider.hdfs.test.HdfsFileProviderTest.<clinit>(HdfsFileProviderTest.java:95)
> ...
>
> And it actually means, that "ls.exe" is not in the PATH. So yes, the test
> does not work on all Windows systems, it requires at least a ls.exe. I will
> remove the automatic running of those tests on the Windows platform
> (again). (but with a better named profile).
>
> As a quick fix it should be enough to add any ls.exe, in my case it was
> the portable git distribution (from github):
>
>
> %LOCALAPPDATA%\GitHub\PortableGit_015aa71ef18c047ce8509ffb2f9e4bb0e3e73f13\bin\ls.exe
>
> Gruss
> Bernd
>
>
>
>
>
>
>
>
>
>
> Am Tue, 10 Jun 2014 11:02:19 -0400 schrieb Gary Gregory
> <garydgreg...@gmail.com>:
>
> > woa... cygwin? I have that installed but it does not help.
> >
> > How about this: can please you turn off HDFS testing on Windows like
> > it was before.
> >
> > I'll be happy to test patches for you on my set up, Windows 7
> > Professional 64-bit Service Pack 1.
> >
> > My understanding is that the HDFS jars we use do not run on Windows
> > out of the box because they rely on calling OS commands that are *Nix
> > specific.
> >
> > Gary
> >
> >
> > On Tue, Jun 10, 2014 at 9:45 AM, Bernd Eckenfels
> > <e...@zusammenkunft.net> wrote:
> >
> > > Hello,
> > >
> > > they do work for me, hm. Windows 7 x64 de. I will try some other
> > > environments. Maybe it picks up some cygwin stuff or something on my
> > > system?
> > >
> > > Gruss
> > > Bernd
> > >
> > >
> > > Am Mon, 9 Jun 2014 18:05:18 -0400
> > > schrieb Gary Gregory <garydgreg...@gmail.com>:
> > >
> > > > Ecki enabled the HDFS tests on Windows but they sure fail for me,
> > > > see below.
> > > >
> > > > Do they work for anyone else on Windows?
> > > >
> > > > Tests run: 2, Failures: 0, Errors: 2, Skipped: 0, Time elapsed:
> > > > 3.681 sec <<< FAILURE! - in
> > > > org.apache.commons.vfs2.provider.hdfs.test.HdfsFileProviderTest
> > > > org.apache.commons.vfs2.provider.hdfs.test.HdfsFileProviderTest
> > > > Time elapsed: 3.68 sec  <<< ERROR!
> > > > java.lang.ExceptionInInitializerError: null
> > > >         at
> > > >
> > >
> org.apache.hadoop.fs.RawLocalFileSystem$RawLocalFileStatus.loadPermissionInfo(RawLocalFileSystem.java:468)
> > > >         at
> > > >
> > >
> org.apache.hadoop.fs.RawLocalFileSystem$RawLocalFileStatus.getPermission(RawLocalFileSystem.java:418)
> > > >         at
> > > >
> > >
> org.apache.hadoop.util.DiskChecker.mkdirsWithExistsAndPermissionCheck(DiskChecker.java:146)
> > > >         at
> > > > org.apache.hadoop.util.DiskChecker.checkDir(DiskChecker.java:162)
> > > > at
> > > >
> > >
> org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:1643)
> > > >         at
> > > >
> > >
> org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:1590)
> > > >         at
> > > >
> > >
> org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:1565)
> > > >         at
> > > >
> > >
> org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.java:421)
> > > >         at
> > > > org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:284)
> > > >         at
> > > >
> > >
> org.apache.commons.vfs2.provider.hdfs.test.HdfsFileProviderTest.<clinit>(HdfsFileProviderTest.java:95)
> > > >
> > > > org.apache.commons.vfs2.provider.hdfs.test.HdfsFileProviderTest
> > > > Time elapsed: 3.681 sec  <<< ERROR!
> > > > java.lang.NoClassDefFoundError: Could not initialize class
> > > > org.apache.commons.vfs2.provider.hdfs.test.HdfsFileProviderTest
> > > >         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native
> > > > Method) at
> > > >
> > >
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
> > > >         at
> > > >
> > >
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> > > >         at java.lang.reflect.Method.invoke(Method.java:606)
> > > >         at
> > > >
> > >
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47)
> > > >         at
> > > >
> > >
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
> > > >         at
> > > >
> > >
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44)
> > > >         at
> > > >
> > >
> org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33)
> > > >         at
> > > > org.junit.runners.ParentRunner.run(ParentRunner.java:309) at
> > > >
> > >
> org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:264)
> > > >         at
> > > >
> > >
> org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:153)
> > > >         at
> > > >
> > >
> org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:124)
> > > >         at
> > > >
> > >
> org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoader(ForkedBooter.java:200)
> > > >         at
> > > >
> > >
> org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:153)
> > > >         at
> > > >
> org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:103)
> > > >
> > > > Gary
> > > >
> > > >
> > >
> > >
> > > ---------------------------------------------------------------------
> > > To unsubscribe, e-mail: dev-unsubscr...@commons.apache.org
> > > For additional commands, e-mail: dev-h...@commons.apache.org
> > >
> > >
> >
> >
>
>
> ---------------------------------------------------------------------
> To unsubscribe, e-mail: dev-unsubscr...@commons.apache.org
> For additional commands, e-mail: dev-h...@commons.apache.org
>
>


-- 
E-Mail: garydgreg...@gmail.com | ggreg...@apache.org
Java Persistence with Hibernate, Second Edition
<http://www.manning.com/bauer3/>
JUnit in Action, Second Edition <http://www.manning.com/tahchiev/>
Spring Batch in Action <http://www.manning.com/templier/>
Blog: http://garygregory.wordpress.com
Home: http://garygregory.com/
Tweet! http://twitter.com/GaryGregory

Reply via email to