I just tried to compile version 0.8.1 against CDH-4.5.0 and it failed just 
the same.

Le mardi 17 décembre 2013 20:56:24 UTC+1, Debasish Das a écrit :
>
> Thanks Matei. 
>
> We will wait for the release candidate of Spark 0.8.1 and whether it could 
> be run against latest CDH/HDP YARN.
>
> On Monday, December 16, 2013 10:02:41 PM UTC-8, Matei Zaharia wrote:
>>
>> Ah, this is because of a YARN API update in CDH 4.5.0 (as well as Apache 
>> Hadoop 2.2). You’ll need to wait for Spark 0.8.1 to compile against that. 
>> There is a release candidate posted on our Apache mailing list: 
>> http://spark.incubator.apache.org/mailing-lists.html<http://www.google.com/url?q=http%3A%2F%2Fspark.incubator.apache.org%2Fmailing-lists.html&sa=D&sntz=1&usg=AFQjCNEHrWNSNaUVApa1nquVgxPLGinJSQ>
>> .
>>
>> Matei
>>
>> On Dec 16, 2013, at 4:51 PM, Debasish Das <[email protected]> wrote:
>>
>> Hi Patrick,
>>
>> With the following configs:
>>
>> export SPARK_HADOOP_VERSION=2.0.0-cdh4.5.0
>> export SPARK_YARN=true
>>
>> Inside the project yarn, the errors are as follows:
>>
>> [warn] 
>> /home/debasish/sag_spark/yarn/src/main/scala/org/apache/spark/deploy/yarn/Client.scala:59:
>>  
>> Treating numbers with a leading zero as octal is deprecated.
>> [warn]   val STAGING_DIR_PERMISSION: FsPermission = 
>> FsPermission.createImmutable(0700:Short)
>> [warn]                                                                   
>>         ^
>> [warn] 
>> /home/debasish/sag_spark/yarn/src/main/scala/org/apache/spark/deploy/yarn/Client.scala:62:
>>  
>> Treating numbers with a leading zero as octal is deprecated.
>> [warn]   val APP_FILE_PERMISSION: FsPermission = 
>> FsPermission.createImmutable(0644:Short) 
>> [warn]                                                                   
>>      ^
>> [error] 
>> /home/debasish/sag_spark/yarn/src/main/scala/org/apache/spark/deploy/yarn/YarnAllocationHandler.scala:36:
>>  
>> object AMResponse is not a member of package 
>> org.apache.hadoop.yarn.api.records
>> [error] import org.apache.hadoop.yarn.api.records.{AMResponse, 
>> ApplicationAttemptId}
>> [error]        ^
>> [error] 
>> /home/debasish/sag_spark/yarn/src/main/scala/org/apache/spark/deploy/yarn/YarnAllocationHandler.scala:105:
>>  
>> value getAMResponse is not a member of 
>> org.apache.hadoop.yarn.api.protocolrecords.AllocateResponse
>> [error]     val amResp = 
>> allocateWorkerResources(workersToRequest).getAMResponse
>> [error]                                                            ^
>> [warn] two warnings found
>> [error] two errors found
>> [error] (yarn/compile:compile) Compilation failed
>> [error] Total time: 15 s, completed Dec 16, 2013 7:47:03 PM
>>
>> Note that I can run the code against cdh4.5.0 mr1 client but we need the 
>> YARN jar for deployment.
>>
>> Thanks.
>> Deb
>>
>> On Friday, December 13, 2013 11:03:32 AM UTC-8, Patrick Wendell wrote:
>>>
>>> What errors are you getting in this case? Is it the same errors as 
>>> before or something else?
>>>
>>>
>>> On Thu, Dec 12, 2013 at 11:54 PM, Debasish Das <[email protected]>wrote:
>>>
>>>> Thanks TD. sbt clean helped.
>>>>
>>>> With these configs I could get the jar file and it runs fine on the 
>>>> standalone spark cluster:
>>>>
>>>> export SPARK_HADOOP_VERSION=2.0.0-mr1-cdh4.5.0
>>>> export SPARK_YARN=false
>>>>
>>>> If I try to generate the deployment jar for YARN with the following 
>>>> configs, I am getting errors.
>>>>
>>>> export SPARK_HADOOP_VERSION=2.0.0-cdh4.5.0
>>>> export SPARK_YARN=true
>>>>
>>>> Thanks.
>>>> Deb 
>>>>
>>>>
>>>> On Thursday, December 12, 2013 10:39:05 PM UTC-8, TD wrote:
>>>>
>>>>> Can you try doing a "sbt clean" before building? I have seen this 
>>>>> error once and a clean build helped. 
>>>>>
>>>>>
>>>>> On Thu, Dec 12, 2013 at 10:37 PM, Debasish Das 
>>>>> <[email protected]>wrote:
>>>>>
>>>>>> Hi,
>>>>>>
>>>>>> I could compile Spark with CDH 4.2.0 but when I tried to access hdfs 
>>>>>> it failed. 
>>>>>>
>>>>>> I looked for the old post on Spark user group and found that Spark 
>>>>>> should be compiled with the exact hadoop client version of the cluster. 
>>>>>>
>>>>>> Our cluster is at CDH 4.5.0. I put the following configs for the 
>>>>>> compilation on the master branch:
>>>>>>
>>>>>> export SPARK_HADOOP_VERSION=2.0.0-cdh4.5.0
>>>>>> export SPARK_YARN=true
>>>>>>
>>>>>> I also tried to see if I can build against the client only
>>>>>>
>>>>>> export SPARK_HADOOP_VERSION=2.0.0-mr1-cdh4.5.0
>>>>>> export SPARK_YARN=false
>>>>>>
>>>>>> I am getting 43 compilation errors from spark-streaming project. 
>>>>>>
>>>>>> I have attached few msgs.
>>>>>>
>>>>>> [error] /home/debasish/sag_spark/streaming/src/main/scala/org/
>>>>>> apache/spark/streaming/api/java/JavaPairDStream.scala:51: type 
>>>>>> mismatch;
>>>>>> [error]  found   : org.apache.spark.streaming.DStream[(K, V)]
>>>>>> [error]  required: org.apache.spark.streaming.
>>>>>> api.java.JavaPairDStream[K,V]
>>>>>> [error]  Note: implicit method fromPairDStream is not applicable here 
>>>>>> because it comes after the application point and it lacks an explicit 
>>>>>> result type
>>>>>> [error]     dstream.filter((x => f(x).booleanValue()))
>>>>>> [error]                   ^
>>>>>> [error] /home/debasish/sag_spark/streaming/src/main/scala/org/
>>>>>> apache/spark/streaming/api/java/JavaPairDStream.scala:54: type 
>>>>>> mismatch;
>>>>>> [error]  found   : org.apache.spark.streaming.DStream[(K, V)]
>>>>>> [error]  required: org.apache.spark.streaming.
>>>>>> api.java.JavaPairDStream[K,V]
>>>>>> [error]  Note: implicit method fromPairDStream is not applicable here 
>>>>>> because it comes after the application point and it lacks an explicit 
>>>>>> result type
>>>>>> [error]   def cache(): JavaPairDStream[K, V] = dstream.cache()
>>>>>> [error]                                                     ^
>>>>>> [error] /home/debasish/sag_spark/streaming/src/main/scala/org/
>>>>>> apache/spark/streaming/api/java/JavaPairDStream.scala:57: type 
>>>>>> mismatch;
>>>>>> [error]  found   : org.apache.spark.streaming.DStream[(K, V)]
>>>>>> [error]  required: org.apache.spark.streaming.
>>>>>> api.java.JavaPairDStream[K,V]
>>>>>> [error]  Note: implicit method fromPairDStream is not applicable here 
>>>>>> because it comes after the application point and it lacks an explicit 
>>>>>> result type
>>>>>> [error]   def persist(): JavaPairDStream[K, V] = dstream.persist()
>>>>>> [error]                                                         ^
>>>>>> [error] /home/debasish/sag_spark/streaming/src/main/scala/org/
>>>>>> apache/spark/streaming/api/java/JavaPairDStream.scala:60: type 
>>>>>> mismatch;
>>>>>> [error]  found   : org.apache.spark.streaming.DStream[(K, V)]
>>>>>> [error]  required: org.apache.spark.streaming.
>>>>>> api.java.JavaPairDStream[K,V]
>>>>>> [error]  Note: implicit method fromPairDStream is not applicable here 
>>>>>> because it comes after the application point and it lacks an explicit 
>>>>>> result type
>>>>>> [error]   def persist(storageLevel: StorageLevel): JavaPairDStream[K, 
>>>>>> V] = dstream.persist(storageLevel)
>>>>>> [error]                                                               
>>>>>>                     ^
>>>>>> [error] /home/debasish/sag_spark/streaming/src/main/scala/org/
>>>>>> apache/spark/streaming/api/java/JavaPairDStream.scala:66: type 
>>>>>> mismatch;
>>>>>> [error]  found   : org.apache.spark.streaming.DStream[(K, V)]
>>>>>> [error]  required: org.apache.spark.streaming.
>>>>>> api.java.JavaPairDStream[K,V]
>>>>>> [error]  Note: implicit method fromPairDStream is not applicable here 
>>>>>> because it comes after the application point and it lacks an explicit 
>>>>>> result type
>>>>>> [error]   def repartition(numPartitions: Int): JavaPairDStream[K, V] 
>>>>>> = dstream.repartition(numPartitions)
>>>>>> [error]                                                               
>>>>>>                     ^
>>>>>> [error] /home/debasish/sag_spark/streaming/src/main/scala/org/
>>>>>> apache/spark/streaming/api/java/JavaPairDStream.scala:83: type 
>>>>>> mismatch;
>>>>>> [error]  found   : org.apache.spark.streaming.DStream[(K, V)]
>>>>>> [error]  required: org.apache.spark.streaming.
>>>>>> api.java.JavaPairDStream[K,V]
>>>>>> [error]  Note: implicit method fromPairDStream is not applicable here 
>>>>>> because it comes after the application point and it lacks an explicit 
>>>>>> result type
>>>>>> [error]     dstream.window(windowDuration)
>>>>>> [error]                   ^
>>>>>>
>>>>>> Note that the project compiled fine with CDH 4.2.0 but I could not 
>>>>>> access our HDFS data.
>>>>>>
>>>>>>  Thanks.
>>>>>> Deb
>>>>>>
>>>>>>
>>>>>> -- 
>>>>>> You received this message because you are subscribed to the Google 
>>>>>> Groups "Spark Users" group.
>>>>>> To unsubscribe from this group and stop receiving emails from it, 
>>>>>> send an email to [email protected].
>>>>>>
>>>>>> For more options, visit https://groups.google.com/groups/opt_out.
>>>>>>
>>>>>
>>>>>
>>>> -- 
>>>> You received this message because you are subscribed to the Google 
>>>> Groups "Spark Users" group.
>>>> To unsubscribe from this group and stop receiving emails from it, send 
>>>> an email to [email protected].
>>>> For more options, visit https://groups.google.com/groups/opt_out.
>>>>
>>>
>>>
>> -- 
>> You received this message because you are subscribed to the Google Groups 
>> "Spark Users" group.
>> To unsubscribe from this group and stop receiving emails from it, send an 
>> email to [email protected].
>> For more options, visit https://groups.google.com/groups/opt_out.
>>
>>
>>

-- 
You received this message because you are subscribed to the Google Groups 
"Spark Users" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
For more options, visit https://groups.google.com/groups/opt_out.

Reply via email to