I don't understand why since there's no javax.servlet in my build.sbt :

scalaVersion := "2.10.4"

libraryDependencies ++= Seq(
  "org.apache.spark" %% "spark-core" % "1.1.0",
  "org.apache.spark" %% "spark-sql" % "1.1.0",
  "org.apache.spark" %% "spark-mllib" % "1.1.0",
  "org.apache.hadoop" % "hadoop-client" % "2.4.0",
  "org.apache.commons" % "commons-math3" % "3.3",
  "org.scalatest" % "scalatest_2.10" % "2.2.0" % "test"
)

resolvers += "Akka Repository" at "http://repo.akka.io/releases/";


On Tue, Nov 4, 2014 at 11:00 AM, Sean Owen <so...@cloudera.com> wrote:

> Generally this means you included some javax.servlet dependency in
> your project deps. You should exclude any of these as they conflict in
> this bad way with other copies of the servlet API from Spark.
>
> On Tue, Nov 4, 2014 at 7:55 AM, Jaonary Rabarisoa <jaon...@gmail.com>
> wrote:
> > Hi all,
> >
> > I have a spark job that I build with sbt and I can run without any
> problem
> > with sbt run. But when I run it inside IntelliJ Idea I got the following
> > error :
> >
> > Exception encountered when invoking run on a nested suite - class
> > "javax.servlet.FilterRegistration"'s signer information does not match
> > signer information of other classes in the same package
> > java.lang.SecurityException: class "javax.servlet.FilterRegistration"'s
> > signer information does not match signer information of other classes in
> the
> > same package
> > at java.lang.ClassLoader.checkCerts(ClassLoader.java:952)
> > at java.lang.ClassLoader.preDefineClass(ClassLoader.java:666)
> > at java.lang.ClassLoader.defineClass(ClassLoader.java:794)
> > at
> java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142)
> > at java.net.URLClassLoader.defineClass(URLClassLoader.java:449)
> > at java.net.URLClassLoader.access$100(URLClassLoader.java:71)
> > at java.net.URLClassLoader$1.run(URLClassLoader.java:361)
> > at java.net.URLClassLoader$1.run(URLClassLoader.java:355)
> > at java.security.AccessController.doPrivileged(Native Method)
> > at java.net.URLClassLoader.findClass(URLClassLoader.java:354)
> > at java.lang.ClassLoader.loadClass(ClassLoader.java:425)
> > at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308)
> > at java.lang.ClassLoader.loadClass(ClassLoader.java:358)
> > at
> >
> org.eclipse.jetty.servlet.ServletContextHandler.<init>(ServletContextHandler.java:136)
> > at
> >
> org.eclipse.jetty.servlet.ServletContextHandler.<init>(ServletContextHandler.java:129)
> > at
> >
> org.eclipse.jetty.servlet.ServletContextHandler.<init>(ServletContextHandler.java:98)
> > at
> org.apache.spark.ui.JettyUtils$.createServletHandler(JettyUtils.scala:98)
> > at
> org.apache.spark.ui.JettyUtils$.createServletHandler(JettyUtils.scala:89)
> > at org.apache.spark.ui.WebUI.attachPage(WebUI.scala:67)
> > at org.apache.spark.ui.WebUI$$anonfun$attachTab$1.apply(WebUI.scala:60)
> > at org.apache.spark.ui.WebUI$$anonfun$attachTab$1.apply(WebUI.scala:60)
> > at
> >
> scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
> > at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47)
> > at org.apache.spark.ui.WebUI.attachTab(WebUI.scala:60)
> > at org.apache.spark.ui.SparkUI.initialize(SparkUI.scala:66)
> > at org.apache.spark.ui.SparkUI.<init>(SparkUI.scala:60)
> > at org.apache.spark.ui.SparkUI.<init>(SparkUI.scala:42)
> > at org.apache.spark.SparkContext.<init>(SparkContext.scala:223)
> > at org.apache.spark.SparkContext.<init>(SparkContext.scala:98)
> >
> >
> > How can I solve this ?
> >
> >
> > Cheers,
> >
> >
> > Jao
> >
>

Reply via email to