Is spark compatible with Hadoop 1.0.4 ?
I built spark with the variable -
SPARK_HADOOP_VERSION=1.2.1 sbt/sbt assembly
But I got the following jar -
spark-assembly-0.9.0-incubating-hadoop1.0.4.jar
When I ran a spark job, I'm running into this issue-
./pyspark ~/test/test_spark/spark.script
Found multiple Spark assembly jars in
/home/suhash/git_repos/spark-apache-v0.9.0-incubating/assembly/target/scala-2.10:
Traceback (most recent call last):
File "/home/ssatish/test/test_spark/spark.script", line 2, in <module>
sc=SparkContext('local', 'App Name')
File
"/home/suhash/git_repos/spark-apache-v0.9.0-incubating/python/pyspark/context.py",
line 83, in __init__
SparkContext._ensure_initialized(self)
File
"/home/suhash/git_repos/spark-apache-v0.9.0-incubating/python/pyspark/context.py",
line 159, in _ensure_initialized
SparkContext._gateway = launch_gateway()
File
"/home/suhash/git_repos/spark-apache-v0.9.0-incubating/python/pyspark/java_gateway.py",
line 46, in launch_gateway
port = int(proc.stdout.readline())
ValueError: invalid literal for int() with base 10:
'spark-assembly-0.9.0-incubating-hadoop1.0.4.jar\n'
Exception AttributeError: "'SparkContext' object has no attribute '_jsc'"
in <bound method SparkContext.__del__ of <pyspark.context.SparkContext
object at 0x15eaa90>> ignored
Thanks,
Suhas.