Yes, been on the books for a while ...
https://issues.apache.org/jira/browse/SPARK-2356
That one just may always be a known 'gotcha' in Windows; it's kind of
a Hadoop gotcha. I don't know that Spark 100% works on Windows and it
isn't tested on Windows.

On Wed, Feb 25, 2015 at 11:05 PM, boci <boci.b...@gmail.com> wrote:
> Thanks your fast answer...
> in windows it's not working, because hadoop (surprise suprise) need
> winutils.exe. Without this it's not working, but if you not set the hadoop
> directory You simply get
>
> 15/02/26 00:03:16 ERROR Shell: Failed to locate the winutils binary in the
> hadoop binary path
> java.io.IOException: Could not locate executable null\bin\winutils.exe in
> the Hadoop binaries.
>
> b0c1
>
>
> ----------------------------------------------------------------------------------------------------------------------------------
> Skype: boci13, Hangout: boci.b...@gmail.com
>
> On Wed, Feb 25, 2015 at 11:50 PM, Sean Owen <so...@cloudera.com> wrote:
>>
>> Spark and Hadoop should be listed as 'provided' dependency in your
>> Maven or SBT build. But that should make it available at compile time.
>>
>> On Wed, Feb 25, 2015 at 10:42 PM, boci <boci.b...@gmail.com> wrote:
>> > Hi,
>> >
>> > I have a little question. I want to develop a spark based application,
>> > but
>> > spark depend to hadoop-client library. I think it's not necessary (spark
>> > standalone) so I excluded from sbt file.. the result is interesting. My
>> > trait where I create the spark context not compiled.
>> >
>> > The error:
>> > ...
>> >  scala.reflect.internal.Types$TypeError: bad symbolic reference. A
>> > signature
>> > in SparkContext.class refers to term mapred
>> > [error] in package org.apache.hadoop which is not available.
>> > [error] It may be completely missing from the current classpath, or the
>> > version on
>> > [error] the classpath might be incompatible with the version used when
>> > compiling SparkContext.class.
>> > ...
>> >
>> > I used this class for integration test. I'm using windows and I don't
>> > want
>> > to using hadoop for integration test. How can I solve this?
>> >
>> > Thanks
>> > Janos
>> >
>
>

---------------------------------------------------------------------
To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
For additional commands, e-mail: user-h...@spark.apache.org

Reply via email to