Thanks your fast answer... in windows it's not working, because hadoop (surprise suprise) need winutils.exe. Without this it's not working, but if you not set the hadoop directory You simply get
15/02/26 00:03:16 ERROR Shell: Failed to locate the winutils binary in the hadoop binary path java.io.IOException: Could not locate executable null\bin\winutils.exe in the Hadoop binaries. b0c1 ---------------------------------------------------------------------------------------------------------------------------------- Skype: boci13, Hangout: boci.b...@gmail.com On Wed, Feb 25, 2015 at 11:50 PM, Sean Owen <so...@cloudera.com> wrote: > Spark and Hadoop should be listed as 'provided' dependency in your > Maven or SBT build. But that should make it available at compile time. > > On Wed, Feb 25, 2015 at 10:42 PM, boci <boci.b...@gmail.com> wrote: > > Hi, > > > > I have a little question. I want to develop a spark based application, > but > > spark depend to hadoop-client library. I think it's not necessary (spark > > standalone) so I excluded from sbt file.. the result is interesting. My > > trait where I create the spark context not compiled. > > > > The error: > > ... > > scala.reflect.internal.Types$TypeError: bad symbolic reference. A > signature > > in SparkContext.class refers to term mapred > > [error] in package org.apache.hadoop which is not available. > > [error] It may be completely missing from the current classpath, or the > > version on > > [error] the classpath might be incompatible with the version used when > > compiling SparkContext.class. > > ... > > > > I used this class for integration test. I'm using windows and I don't > want > > to using hadoop for integration test. How can I solve this? > > > > Thanks > > Janos > > >