I'm not 100% sure that is causing a problem, though. The stream still starts, but is giving blank output. I checked the environment variables in the ui and it is running local[*], so there should be no bottleneck there.
On Mon, May 18, 2015 at 10:08 PM, Justin Pihony <justin.pih...@gmail.com> wrote: > I am trying to print a basic twitter stream and receiving the following > error: > > > 15/05/18 22:03:14 INFO Executor: Fetching > http://192.168.56.1:49752/jars/twitter4j-media-support-3.0.3.jar with > timestamp 1432000973058 > 15/05/18 22:03:14 INFO Utils: Fetching > http://192.168.56.1:49752/jars/twitter4j-media-support-3.0.3.jar to > C:\Users\Justin\AppData\Local\Temp\spark-4a37d3 > > e9-34a2-40d4-b09b-6399931f527d\userFiles-65ee748e-4721-4e16-9fe6-65933651fec1\fetchFileTemp8970201232303518432.tmp > 15/05/18 22:03:14 ERROR Executor: Exception in task 0.0 in stage 0.0 (TID > 0) > java.lang.NullPointerException > at java.lang.ProcessBuilder.start(ProcessBuilder.java:1012) > at org.apache.hadoop.util.Shell.runCommand(Shell.java:482) > at org.apache.hadoop.util.Shell.run(Shell.java:455) > at > org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.ja > va:715) > at org.apache.hadoop.fs.FileUtil.chmod(FileUtil.java:873) > at org.apache.hadoop.fs.FileUtil.chmod(FileUtil.java:853) > at org.apache.spark.util.Utils$.fetchFile(Utils.scala:443) > at > > org.apache.spark.executor.Executor$$anonfun$org$apache$spark$executor$Executor$$updateDependencies$5.apply(Executor.scala:374) > at > > org.apache.spark.executor.Executor$$anonfun$org$apache$spark$executor$Executor$$updateDependencies$5.apply(Executor.scala:366) > at > > scala.collection.TraversableLike$WithFilter$$anonfun$foreach$1.apply(TraversableLike.scala:772) > at > scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:98) > at > scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:98) > at > scala.collection.mutable.HashTable$class.foreachEntry(HashTable.scala:226) > at scala.collection.mutable.HashMap.foreachEntry(HashMap.scala:39) > at scala.collection.mutable.HashMap.foreach(HashMap.scala:98) > at > > scala.collection.TraversableLike$WithFilter.foreach(TraversableLike.scala:771) > at > org.apache.spark.executor.Executor.org > $apache$spark$executor$Executor$$updateDependencies(Executor.scala:366) > at > org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:184) > at > > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) > at > > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) > at java.lang.Thread.run(Thread.java:744) > > > Code is: > > spark-shell --jars > > \Spark\lib\spark-streaming-twitter_2.10-1.3.1.jar,\Spark\lib\twitter4j-async-3.0.3.jar,\Spark\lib\twitter4j-core-3.0.3.jar,\Spark\lib\twitter4j-media-support-3.0.3.jar,\Spark\lib\twitter4j-stream-3.0.3.jar > > import org.apache.spark.streaming.twitter._ > import org.apache.spark.streaming._ > > System.setProperty("twitter4j.oauth.consumerKey","*****") > System.setProperty("twitter4j.oauth.consumerSecret","*****") > System.setProperty("twitter4j.oauth.accessToken","*****") > System.setProperty("twitter4j.oauth.accessTokenSecret","*****") > > val ssc = new StreamingContext(sc, Seconds(10)) > val stream = TwitterUtils.createStream(ssc, None) > stream.print > ssc.start > > > This seems to be happening at FileUtil.chmod(targetFile.getAbsolutePath, > "a+x") but Im not sure why... > > > > -- > View this message in context: > http://apache-spark-user-list.1001560.n3.nabble.com/TwitterUtils-on-Windows-tp22939.html > Sent from the Apache Spark User List mailing list archive at Nabble.com. > > --------------------------------------------------------------------- > To unsubscribe, e-mail: user-unsubscr...@spark.apache.org > For additional commands, e-mail: user-h...@spark.apache.org > >