Yes, been on the books for a while ... https://issues.apache.org/jira/browse/SPARK-2356 That one just may always be a known 'gotcha' in Windows; it's kind of a Hadoop gotcha. I don't know that Spark 100% works on Windows and it isn't tested on Windows.
On Wed, Feb 25, 2015 at 11:05 PM, boci <boci.b...@gmail.com> wrote: > Thanks your fast answer... > in windows it's not working, because hadoop (surprise suprise) need > winutils.exe. Without this it's not working, but if you not set the hadoop > directory You simply get > > 15/02/26 00:03:16 ERROR Shell: Failed to locate the winutils binary in the > hadoop binary path > java.io.IOException: Could not locate executable null\bin\winutils.exe in > the Hadoop binaries. > > b0c1 > > > ---------------------------------------------------------------------------------------------------------------------------------- > Skype: boci13, Hangout: boci.b...@gmail.com > > On Wed, Feb 25, 2015 at 11:50 PM, Sean Owen <so...@cloudera.com> wrote: >> >> Spark and Hadoop should be listed as 'provided' dependency in your >> Maven or SBT build. But that should make it available at compile time. >> >> On Wed, Feb 25, 2015 at 10:42 PM, boci <boci.b...@gmail.com> wrote: >> > Hi, >> > >> > I have a little question. I want to develop a spark based application, >> > but >> > spark depend to hadoop-client library. I think it's not necessary (spark >> > standalone) so I excluded from sbt file.. the result is interesting. My >> > trait where I create the spark context not compiled. >> > >> > The error: >> > ... >> > scala.reflect.internal.Types$TypeError: bad symbolic reference. A >> > signature >> > in SparkContext.class refers to term mapred >> > [error] in package org.apache.hadoop which is not available. >> > [error] It may be completely missing from the current classpath, or the >> > version on >> > [error] the classpath might be incompatible with the version used when >> > compiling SparkContext.class. >> > ... >> > >> > I used this class for integration test. I'm using windows and I don't >> > want >> > to using hadoop for integration test. How can I solve this? >> > >> > Thanks >> > Janos >> > > > --------------------------------------------------------------------- To unsubscribe, e-mail: user-unsubscr...@spark.apache.org For additional commands, e-mail: user-h...@spark.apache.org