I meant yarn and hadoop defaulted to 1.0.4 so the yarn build fails since 1.0.4 
doesn’t exist for yarn...

Thanks,
Ron

On Aug 4, 2014, at 10:01 AM, Ron's Yahoo! <zlgonza...@yahoo.com> wrote:

> That failed since it defaulted the versions for yarn and hadoop 
> I’ll give it a try with just 2.4.0 for both yarn and hadoop…
> 
> Thanks,
> Ron
> 
> On Aug 4, 2014, at 9:44 AM, Patrick Wendell <pwend...@gmail.com> wrote:
> 
>> Can you try building without any of the special `hadoop.version` flags and 
>> just building only with -Phadoop-2.4? In the past users have reported issues 
>> trying to build random spot versions... I think HW is supposed to be 
>> compatible with the normal 2.4.0 build.
>> 
>> 
>> On Mon, Aug 4, 2014 at 8:35 AM, Ron's Yahoo! <zlgonza...@yahoo.com.invalid> 
>> wrote:
>> Thanks, I ensured that $SPARK_HOME/pom.xml had the HDP repository under the 
>> repositories element. I also confirmed that if the build couldn’t find the 
>> version, it would fail fast so it seems as if it’s able to get the versions 
>> it needs to build the distribution.
>> I ran the following (generated from make-distribution.sh), but it did not 
>> address the problem, while building with an older version 
>> (2.4.0.2.1.2.0-402) worked. Any other thing I can try?
>> 
>> mvn clean package -Phadoop-2.4 -Phive -Pyarn 
>> -Dyarn.version=2.4.0.2.1.2.0-563 -Dhadoop.version=2.4.0.2.1.3.0-563 
>> -DskipTests
>> 
>> 
>> Thanks,
>> Ron
>> 
>> 
>> On Aug 4, 2014, at 7:13 AM, Steve Nunez <snu...@hortonworks.com> wrote:
>> 
>>> Provided you¹ve got the HWX repo in your pom.xml, you can build with this
>>> line:
>>> 
>>> mvn -Pyarn -Phive -Phadoop-2.4 -Dhadoop.version=2.4.0.2.1.1.0-385
>>> -DskipTests clean package
>>> 
>>> I haven¹t tried building a distro, but it should be similar.
>>> 
>>> 
>>>     - SteveN
>>> 
>>> On 8/4/14, 1:25, "Sean Owen" <so...@cloudera.com> wrote:
>>> 
>>>> For any Hadoop 2.4 distro, yes, set hadoop.version but also set
>>>> -Phadoop-2.4. http://spark.apache.org/docs/latest/building-with-maven.html
>>>> 
>>>> On Mon, Aug 4, 2014 at 9:15 AM, Patrick Wendell <pwend...@gmail.com>
>>>> wrote:
>>>>> For hortonworks, I believe it should work to just link against the
>>>>> corresponding upstream version. I.e. just set the Hadoop version to
>>>>> "2.4.0"
>>>>> 
>>>>> Does that work?
>>>>> 
>>>>> - Patrick
>>>>> 
>>>>> 
>>>>> On Mon, Aug 4, 2014 at 12:13 AM, Ron's Yahoo!
>>>>> <zlgonza...@yahoo.com.invalid>
>>>>> wrote:
>>>>>> 
>>>>>> Hi,
>>>>>>  Not sure whose issue this is, but if I run make-distribution using
>>>>>> HDP
>>>>>> 2.4.0.2.1.3.0-563 as the hadoop version (replacing it in
>>>>>> make-distribution.sh), I get a strange error with the exception below.
>>>>>> If I
>>>>>> use a slightly older version of HDP (2.4.0.2.1.2.0-402) with
>>>>>> make-distribution, using the generated assembly all works fine for me.
>>>>>> Either 1.0.0 or 1.0.1 will work fine.
>>>>>> 
>>>>>>  Should I file a JIRA or is this a known issue?
>>>>>> 
>>>>>> Thanks,
>>>>>> Ron
>>>>>> 
>>>>>> Exception in thread "main" org.apache.spark.SparkException: Job aborted
>>>>>> due to stage failure: Task 0.0:0 failed 1 times, most recent failure:
>>>>>> Exception failure in TID 0 on host localhost:
>>>>>> java.lang.IncompatibleClassChangeError: Found interface
>>>>>> org.apache.hadoop.mapreduce.TaskAttemptContext, but class was expected
>>>>>> 
>>>>>> 
>>>>>> org.apache.avro.mapreduce.AvroKeyInputFormat.createRecordReader(AvroKeyI
>>>>>> nputFormat.java:47)
>>>>>> 
>>>>>> 
>>>>>> org.apache.spark.rdd.NewHadoopRDD$$anon$1.<init>(NewHadoopRDD.scala:111)
>>>>>> 
>>>>>> org.apache.spark.rdd.NewHadoopRDD.compute(NewHadoopRDD.scala:99)
>>>>>> 
>>>>>> org.apache.spark.rdd.NewHadoopRDD.compute(NewHadoopRDD.scala:61)
>>>>>>        org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:262)
>>>>>>        org.apache.spark.rdd.RDD.iterator(RDD.scala:229)
>>>>>>        org.apache.spark.rdd.MappedRDD.compute(MappedRDD.scala:31)
>>>>>>        org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:262)
>>>>>> 
>>>>>> org.apache.spark.CacheManager.getOrCompute(CacheManager.scala:77)
>>>>>>        org.apache.spark.rdd.RDD.iterator(RDD.scala:227)
>>>>>>        org.apache.spark.rdd.MappedRDD.compute(MappedRDD.scala:31)
>>>>>>        org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:262)
>>>>>>        org.apache.spark.rdd.RDD.iterator(RDD.scala:229)
>>>>>> 
>>>>>> org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:111)
>>>>>>        org.apache.spark.scheduler.Task.run(Task.scala:51)
>>>>>> 
>>>>>> org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:187)
>>>>>> 
>>>>>> 
>>>>>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.jav
>>>>>> a:1145)
>>>>>> 
>>>>>> 
>>>>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.ja
>>>>>> va:615)
>>>>>>        java.lang.Thread.run(Thread.java:745)
>>>>> 
>>>>> 
>>>> 
>>>> ---------------------------------------------------------------------
>>>> To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org
>>>> For additional commands, e-mail: dev-h...@spark.apache.org
>>>> 
>>> 
>>> 
>>> 
>>> -- 
>>> CONFIDENTIALITY NOTICE
>>> NOTICE: This message is intended for the use of the individual or entity to 
>>> which it is addressed and may contain information that is confidential, 
>>> privileged and exempt from disclosure under applicable law. If the reader 
>>> of this message is not the intended recipient, you are hereby notified that 
>>> any printing, copying, dissemination, distribution, disclosure or 
>>> forwarding of this communication is strictly prohibited. If you have 
>>> received this communication in error, please contact the sender immediately 
>>> and delete it from your system. Thank You.
>>> 
>>> ---------------------------------------------------------------------
>>> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
>>> For additional commands, e-mail: user-h...@spark.apache.org
>> 
>> 
> 

Reply via email to