I think I found the answer -> http://apache-spark-user-list.1001560.n3.nabble.com/Error-while-running-example-scala-application-using-spark-submit-td10056.html
Do I have no way of running this in Windows locally? On Mon, May 18, 2015 at 10:44 PM, Justin Pihony <justin.pih...@gmail.com> wrote: > I'm not 100% sure that is causing a problem, though. The stream still > starts, but is giving blank output. I checked the environment variables in > the ui and it is running local[*], so there should be no bottleneck there. > > On Mon, May 18, 2015 at 10:08 PM, Justin Pihony <justin.pih...@gmail.com> > wrote: > >> I am trying to print a basic twitter stream and receiving the following >> error: >> >> >> 15/05/18 22:03:14 INFO Executor: Fetching >> http://192.168.56.1:49752/jars/twitter4j-media-support-3.0.3.jar with >> timestamp 1432000973058 >> 15/05/18 22:03:14 INFO Utils: Fetching >> http://192.168.56.1:49752/jars/twitter4j-media-support-3.0.3.jar to >> C:\Users\Justin\AppData\Local\Temp\spark-4a37d3 >> >> e9-34a2-40d4-b09b-6399931f527d\userFiles-65ee748e-4721-4e16-9fe6-65933651fec1\fetchFileTemp8970201232303518432.tmp >> 15/05/18 22:03:14 ERROR Executor: Exception in task 0.0 in stage 0.0 (TID >> 0) >> java.lang.NullPointerException >> at java.lang.ProcessBuilder.start(ProcessBuilder.java:1012) >> at org.apache.hadoop.util.Shell.runCommand(Shell.java:482) >> at org.apache.hadoop.util.Shell.run(Shell.java:455) >> at >> org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.ja >> va:715) >> at org.apache.hadoop.fs.FileUtil.chmod(FileUtil.java:873) >> at org.apache.hadoop.fs.FileUtil.chmod(FileUtil.java:853) >> at org.apache.spark.util.Utils$.fetchFile(Utils.scala:443) >> at >> >> org.apache.spark.executor.Executor$$anonfun$org$apache$spark$executor$Executor$$updateDependencies$5.apply(Executor.scala:374) >> at >> >> org.apache.spark.executor.Executor$$anonfun$org$apache$spark$executor$Executor$$updateDependencies$5.apply(Executor.scala:366) >> at >> >> scala.collection.TraversableLike$WithFilter$$anonfun$foreach$1.apply(TraversableLike.scala:772) >> at >> >> scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:98) >> at >> >> scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:98) >> at >> scala.collection.mutable.HashTable$class.foreachEntry(HashTable.scala:226) >> at scala.collection.mutable.HashMap.foreachEntry(HashMap.scala:39) >> at scala.collection.mutable.HashMap.foreach(HashMap.scala:98) >> at >> >> scala.collection.TraversableLike$WithFilter.foreach(TraversableLike.scala:771) >> at >> org.apache.spark.executor.Executor.org >> $apache$spark$executor$Executor$$updateDependencies(Executor.scala:366) >> at >> org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:184) >> at >> >> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) >> at >> >> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) >> at java.lang.Thread.run(Thread.java:744) >> >> >> Code is: >> >> spark-shell --jars >> >> \Spark\lib\spark-streaming-twitter_2.10-1.3.1.jar,\Spark\lib\twitter4j-async-3.0.3.jar,\Spark\lib\twitter4j-core-3.0.3.jar,\Spark\lib\twitter4j-media-support-3.0.3.jar,\Spark\lib\twitter4j-stream-3.0.3.jar >> >> import org.apache.spark.streaming.twitter._ >> import org.apache.spark.streaming._ >> >> System.setProperty("twitter4j.oauth.consumerKey","*****") >> System.setProperty("twitter4j.oauth.consumerSecret","*****") >> System.setProperty("twitter4j.oauth.accessToken","*****") >> System.setProperty("twitter4j.oauth.accessTokenSecret","*****") >> >> val ssc = new StreamingContext(sc, Seconds(10)) >> val stream = TwitterUtils.createStream(ssc, None) >> stream.print >> ssc.start >> >> >> This seems to be happening at FileUtil.chmod(targetFile.getAbsolutePath, >> "a+x") but Im not sure why... >> >> >> >> -- >> View this message in context: >> http://apache-spark-user-list.1001560.n3.nabble.com/TwitterUtils-on-Windows-tp22939.html >> Sent from the Apache Spark User List mailing list archive at Nabble.com. >> >> --------------------------------------------------------------------- >> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org >> For additional commands, e-mail: user-h...@spark.apache.org >> >> >