> On 19 May 2015, at 03:08, Justin Pihony <justin.pih...@gmail.com> wrote:
> 
> 
> 15/05/18 22:03:14 INFO Executor: Fetching
> http://192.168.56.1:49752/jars/twitter4j-media-support-3.0.3.jar with
> timestamp 1432000973058
> 15/05/18 22:03:14 INFO Utils: Fetching
> http://192.168.56.1:49752/jars/twitter4j-media-support-3.0.3.jar to
> C:\Users\Justin\AppData\Local\Temp\spark-4a37d3
> e9-34a2-40d4-b09b-6399931f527d\userFiles-65ee748e-4721-4e16-9fe6-65933651fec1\fetchFileTemp8970201232303518432.tmp
> 15/05/18 22:03:14 ERROR Executor: Exception in task 0.0 in stage 0.0 (TID 0)
> java.lang.NullPointerException
>        at java.lang.ProcessBuilder.start(ProcessBuilder.java:1012)
>        at org.apache.hadoop.util.Shell.runCommand(Shell.java:482)
>        at org.apache.hadoop.util.Shell.run(Shell.java:455)
>        at
> org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.ja
> va:715)
>        at org.apache.hadoop.fs.FileUtil.chmod(FileUtil.java:873)
>        at org.apache.hadoop.fs.FileUtil.chmod(FileUtil.java:853)
>        at org.apache.spark.util.Utils$.fetchFile(Utils.scala:443)
>        at

you're going to need to set up Hadoop on your system enough for to execute the 
chmod operation via the winutils.exe

one tactic: grab the hortonworks windows version, install it (including setting 
up HADOOP_HOME). You don't need to run any of the hadoop services, you just 
need the binaries in the right place.

other: 

1. grab the copy of the relevant binaries which I've stuck up online
        
https://github.com/steveloughran/clusterconfigs/tree/master/clusters/morzine/hadoop_home/bin
2. install to some directory hadoop/bin
3. set the env variable HADOOP_HOME to the hadoopp dir (not the bin one)
4. set PATH=%PATH%;%HADOOP_HOME%/bin

---------------------------------------------------------------------
To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
For additional commands, e-mail: user-h...@spark.apache.org

Reply via email to