Is spark compatible with Hadoop 1.0.4 ?

I built spark with the variable -
SPARK_HADOOP_VERSION=1.2.1 sbt/sbt assembly

But I got the following jar -
spark-assembly-0.9.0-incubating-hadoop1.0.4.jar

When I ran a spark job, I'm  running  into this issue-

 ./pyspark ~/test/test_spark/spark.script
Found multiple Spark assembly jars in
/home/suhash/git_repos/spark-apache-v0.9.0-incubating/assembly/target/scala-2.10:
Traceback (most recent call last):
  File "/home/ssatish/test/test_spark/spark.script", line 2, in <module>
    sc=SparkContext('local', 'App Name')
  File
"/home/suhash/git_repos/spark-apache-v0.9.0-incubating/python/pyspark/context.py",
line 83, in __init__
    SparkContext._ensure_initialized(self)
  File
"/home/suhash/git_repos/spark-apache-v0.9.0-incubating/python/pyspark/context.py",
line 159, in _ensure_initialized
    SparkContext._gateway = launch_gateway()
  File
"/home/suhash/git_repos/spark-apache-v0.9.0-incubating/python/pyspark/java_gateway.py",
line 46, in launch_gateway
    port = int(proc.stdout.readline())
ValueError: invalid literal for int() with base 10:
'spark-assembly-0.9.0-incubating-hadoop1.0.4.jar\n'
Exception AttributeError: "'SparkContext' object has no attribute '_jsc'"
in <bound method SparkContext.__del__ of <pyspark.context.SparkContext
object at 0x15eaa90>> ignored




Thanks,
Suhas.

Reply via email to