You can just use the Maven build for now, even for Spark 1.0.0. Matei
On Jun 2, 2014, at 5:30 PM, Mohit Nayak <wiza...@gmail.com> wrote: > Hey, > Yup that fixed it. Thanks so much! > > Is this the only solution, or could this be resolved in future versions of > Spark ? > > > On Mon, Jun 2, 2014 at 5:14 PM, Sean Owen <so...@cloudera.com> wrote: > If it's the SBT build, I suspect you are hitting > https://issues.apache.org/jira/browse/SPARK-1949 > > Can you try to apply the excludes you see at > https://github.com/apache/spark/pull/906/files to your build to see if > it resolves it? > > If so I think this could be helpful to commit. > > On Tue, Jun 3, 2014 at 1:01 AM, Mohit Nayak <wiza...@gmail.com> wrote: > > Hey, > > Thanks for the reply. > > > > I am using SBT. Here is a list of my dependancies: > > val sparkCore = "org.apache.spark" % "spark-core_2.10" % V.spark > > val hadoopCore = "org.apache.hadoop" % "hadoop-core" % > > V.hadoop % "provided" > > val jodaTime = "com.github.nscala-time" %% "nscala-time" % > > "0.8.0" > > val scalaUtil = "com.twitter" %% "util-collection" % > > V.util > > val logback = "ch.qos.logback" % "logback-classic" % "1.0.6" % > > "runtime" > > var openCsv = "net.sf.opencsv" % "opencsv" % "2.1" > > var scalaTest = "org.scalatest" % "scalatest_2.10" % "2.1.0" % "test" > > var scalaIOCore = "com.github.scala-incubator.io" %% "scala-io-core" % > > V.scalaIO > > var scalaIOFile = "com.github.scala-incubator.io" %% "scala-io-file" % > > V.scalaIO > > var kryo = "com.esotericsoftware.kryo" % "kryo" % "2.16" > > var spray = "io.spray" %% "spray-json" % "1.2.5" > > var scala_reflect = "org.scala-lang" % "scala-reflect" % "2.10.3" > > > > > > > > On Mon, Jun 2, 2014 at 4:23 PM, Sean Owen <so...@cloudera.com> wrote: > >> > >> This ultimately means you have a couple copies of the servlet APIs in > >> the build. What is your build like (SBT? Maven?) and what exactly are > >> you depending on? > >> > >> On Tue, Jun 3, 2014 at 12:21 AM, Mohit Nayak <wiza...@gmail.com> wrote: > >> > Hi, > >> > I've upgraded to Spark 1.0.0. I'm not able to run any tests. They throw > >> > a > >> > > >> > java.lang.SecurityException: class "javax.servlet.FilterRegistration"'s > >> > signer information does not match signer information of other classes in > >> > the > >> > same package > >> > > >> > > >> > I'm using Hadoop-core 1.0.4 and running this locally. > >> > I noticed that there was an issue regarding this and was marked as > >> > resolved > >> > [https://issues.apache.org/jira/browse/SPARK-1693] > >> > Please guide.. > >> > > >> > -- > >> > -Mohit > >> > wiza...@gmail.com > >> > > >> > > >> > > >> > -- > >> > -Mohit > >> > wiza...@gmail.com > > > > > > > > > > -- > > -Mohit > > wiza...@gmail.com > > > > -- > -Mohit > wiza...@gmail.com