try sbt clean command before build the app. or delete .ivy2 ans .sbt folders(not a good methode). Then try to rebuild the project.
On Thu, Jun 5, 2014 at 11:45 AM, Sean Owen <so...@cloudera.com> wrote: > I think this is SPARK-1949 again: https://github.com/apache/spark/pull/906 > I think this change fixed this issue for a few people using the SBT > build, worth committing? > > On Thu, Jun 5, 2014 at 6:40 AM, Shrikar archak <shrika...@gmail.com> > wrote: > > Hi All, > > Now that the Spark Version 1.0.0 is release there should not be any > problem > > with the local jars. > > Shrikars-MacBook-Pro:SimpleJob shrikar$ cat simple.sbt > > name := "Simple Project" > > > > version := "1.0" > > > > scalaVersion := "2.10.4" > > > > libraryDependencies ++= Seq("org.apache.spark" %% "spark-core" % "1.0.0", > > "org.apache.spark" %% "spark-streaming" % > > "1.0.0") > > > > resolvers += "Akka Repository" at "http://repo.akka.io/releases/" > > > > I am still having this issue > > [error] (run-main) java.lang.NoClassDefFoundError: > > javax/servlet/http/HttpServletResponse > > java.lang.NoClassDefFoundError: javax/servlet/http/HttpServletResponse > > at org.apache.spark.HttpServer.start(HttpServer.scala:54) > > at > > > org.apache.spark.broadcast.HttpBroadcast$.createServer(HttpBroadcast.scala:156) > > at > > > org.apache.spark.broadcast.HttpBroadcast$.initialize(HttpBroadcast.scala:127) > > at > > > org.apache.spark.broadcast.HttpBroadcastFactory.initialize(HttpBroadcastFactory.scala:31) > > at > > > org.apache.spark.broadcast.BroadcastManager.initialize(BroadcastManager.scala:48) > > at > > > org.apache.spark.broadcast.BroadcastManager.<init>(BroadcastManager.scala:35) > > at org.apache.spark.SparkEnv$.create(SparkEnv.scala:218) > > at org.apache.spark.SparkContext.<init>(SparkContext.scala:202) > > > > Any help would be greatly appreciated. > > > > Thanks, > > Shrikar > > > > > > On Fri, May 23, 2014 at 3:58 PM, Shrikar archak <shrika...@gmail.com> > wrote: > >> > >> Still the same error no change > >> > >> Thanks, > >> Shrikar > >> > >> > >> On Fri, May 23, 2014 at 2:38 PM, Jacek Laskowski <ja...@japila.pl> > wrote: > >>> > >>> Hi Shrikar, > >>> > >>> How did you build Spark 1.0.0-SNAPSHOT on your machine? My > >>> understanding is that `sbt publishLocal` is not enough and you really > >>> need `sbt assembly` instead. Give it a try and report back. > >>> > >>> As to your build.sbt, upgrade Scala to 2.10.4 and "org.apache.spark" > >>> %% "spark-streaming" % "1.0.0-SNAPSHOT" only that will pull down > >>> spark-core as a transitive dep. The resolver for Akka Repository is > >>> not needed. Your build.sbt should really look as follows: > >>> > >>> name := "Simple Project" > >>> > >>> version := "1.0" > >>> > >>> scalaVersion := "2.10.4" > >>> > >>> libraryDependencies += "org.apache.spark" %% "spark-streaming" % > >>> "1.0.0-SNAPSHOT" > >>> > >>> Jacek > >>> > >>> On Thu, May 22, 2014 at 11:27 PM, Shrikar archak <shrika...@gmail.com> > >>> wrote: > >>> > Hi All, > >>> > > >>> > I am trying to run the network count example as a seperate standalone > >>> > job > >>> > and running into some issues. > >>> > > >>> > Environment: > >>> > 1) Mac Mavericks > >>> > 2) Latest spark repo from Github. > >>> > > >>> > > >>> > I have a structure like this > >>> > > >>> > Shrikars-MacBook-Pro:SimpleJob shrikar$ find . > >>> > . > >>> > ./simple.sbt > >>> > ./src > >>> > ./src/main > >>> > ./src/main/scala > >>> > ./src/main/scala/NetworkWordCount.scala > >>> > ./src/main/scala/SimpleApp.scala.bk > >>> > > >>> > > >>> > simple.sbt > >>> > name := "Simple Project" > >>> > > >>> > version := "1.0" > >>> > > >>> > scalaVersion := "2.10.3" > >>> > > >>> > libraryDependencies ++= Seq("org.apache.spark" %% "spark-core" % > >>> > "1.0.0-SNAPSHOT", > >>> > "org.apache.spark" %% "spark-streaming" % > >>> > "1.0.0-SNAPSHOT") > >>> > > >>> > resolvers += "Akka Repository" at "http://repo.akka.io/releases/" > >>> > > >>> > > >>> > I am able to run the SimpleApp which is mentioned in the doc but > when I > >>> > try > >>> > to run the NetworkWordCount app I get error like this am I missing > >>> > something? > >>> > > >>> > [info] Running com.shrikar.sparkapps.NetworkWordCount > >>> > 14/05/22 14:26:47 INFO spark.SecurityManager: Changing view acls to: > >>> > shrikar > >>> > 14/05/22 14:26:47 INFO spark.SecurityManager: SecurityManager: > >>> > authentication disabled; ui acls disabled; users with view > permissions: > >>> > Set(shrikar) > >>> > 14/05/22 14:26:48 INFO slf4j.Slf4jLogger: Slf4jLogger started > >>> > 14/05/22 14:26:48 INFO Remoting: Starting remoting > >>> > 14/05/22 14:26:48 INFO Remoting: Remoting started; listening on > >>> > addresses > >>> > :[akka.tcp://spark@192.168.10.88:49963] > >>> > 14/05/22 14:26:48 INFO Remoting: Remoting now listens on addresses: > >>> > [akka.tcp://spark@192.168.10.88:49963] > >>> > 14/05/22 14:26:48 INFO spark.SparkEnv: Registering MapOutputTracker > >>> > 14/05/22 14:26:48 INFO spark.SparkEnv: Registering BlockManagerMaster > >>> > 14/05/22 14:26:48 INFO storage.DiskBlockManager: Created local > >>> > directory at > >>> > > >>> > > /var/folders/r2/mbj08pb55n5d_9p8588xk5b00000gn/T/spark-local-20140522142648-0a14 > >>> > 14/05/22 14:26:48 INFO storage.MemoryStore: MemoryStore started with > >>> > capacity 911.6 MB. > >>> > 14/05/22 14:26:48 INFO network.ConnectionManager: Bound socket to > port > >>> > 49964 > >>> > with id = ConnectionManagerId(192.168.10.88,49964) > >>> > 14/05/22 14:26:48 INFO storage.BlockManagerMaster: Trying to register > >>> > BlockManager > >>> > 14/05/22 14:26:48 INFO storage.BlockManagerInfo: Registering block > >>> > manager > >>> > 192.168.10.88:49964 with 911.6 MB RAM > >>> > 14/05/22 14:26:48 INFO storage.BlockManagerMaster: Registered > >>> > BlockManager > >>> > 14/05/22 14:26:48 INFO spark.HttpServer: Starting HTTP Server > >>> > [error] (run-main) java.lang.NoClassDefFoundError: > >>> > javax/servlet/http/HttpServletResponse > >>> > java.lang.NoClassDefFoundError: > javax/servlet/http/HttpServletResponse > >>> > at org.apache.spark.HttpServer.start(HttpServer.scala:54) > >>> > at > >>> > > >>> > > org.apache.spark.broadcast.HttpBroadcast$.createServer(HttpBroadcast.scala:156) > >>> > at > >>> > > >>> > > org.apache.spark.broadcast.HttpBroadcast$.initialize(HttpBroadcast.scala:127) > >>> > at > >>> > > >>> > > org.apache.spark.broadcast.HttpBroadcastFactory.initialize(HttpBroadcastFactory.scala:31) > >>> > at > >>> > > >>> > > org.apache.spark.broadcast.BroadcastManager.initialize(BroadcastManager.scala:48) > >>> > at > >>> > > >>> > > org.apache.spark.broadcast.BroadcastManager.<init>(BroadcastManager.scala:35) > >>> > at org.apache.spark.SparkEnv$.create(SparkEnv.scala:218) > >>> > at org.apache.spark.SparkContext.<init>(SparkContext.scala:202) > >>> > at > >>> > > >>> > > org.apache.spark.streaming.StreamingContext$.createNewSparkContext(StreamingContext.scala:549) > >>> > at > >>> > > >>> > > org.apache.spark.streaming.StreamingContext$.createNewSparkContext(StreamingContext.scala:561) > >>> > at > >>> > > >>> > > org.apache.spark.streaming.StreamingContext.<init>(StreamingContext.scala:91) > >>> > at > >>> > > com.shrikar.sparkapps.NetworkWordCount$.main(NetworkWordCount.scala:39) > >>> > at > com.shrikar.sparkapps.NetworkWordCount.main(NetworkWordCount.scala) > >>> > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > >>> > at > >>> > > >>> > > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39) > >>> > at > >>> > > >>> > > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) > >>> > at java.lang.reflect.Method.invoke(Method.java:597) > >>> > > >>> > > >>> > Thanks, > >>> > Shrikar > >>> > > >>> > >>> > >>> > >>> -- > >>> Jacek Laskowski | http://blog.japila.pl > >>> "Never discourage anyone who continually makes progress, no matter how > >>> slow." Plato > >> > >> > > >