I currently don't have plans to work on that.

-Sandy

> On Apr 22, 2014, at 8:06 PM, Gordon Wang <gw...@gopivotal.com> wrote:
> 
> Thanks I see. Do you guys have plan to port this to sbt?
> 
> 
>> On Wed, Apr 23, 2014 at 10:24 AM, Sandy Ryza <sandy.r...@cloudera.com> wrote:
>> Right, it only works for Maven
>> 
>> 
>>> On Tue, Apr 22, 2014 at 6:23 PM, Gordon Wang <gw...@gopivotal.com> wrote:
>>> Hi Sandy,
>>> 
>>> Thanks for your reply !
>>> 
>>> Does this work for sbt ?
>>> 
>>> I checked the commit, looks like only maven build has such option.
>>> 
>>> 
>>> 
>>>> On Wed, Apr 23, 2014 at 12:38 AM, Sandy Ryza <sandy.r...@cloudera.com> 
>>>> wrote:
>>>> Hi Gordon,
>>>> 
>>>> We recently handled this in SPARK-1064.  As of 1.0.0, you'll be able to 
>>>> pass -Phadoop-provided to Maven and avoid including Hadoop and its 
>>>> dependencies in the assembly jar.
>>>> 
>>>> -Sandy
>>>> 
>>>> 
>>>>> On Tue, Apr 22, 2014 at 2:43 AM, Gordon Wang <gw...@gopivotal.com> wrote:
>>>>> In this page http://spark.apache.org/docs/0.9.0/running-on-yarn.html
>>>>> 
>>>>> We have to use spark assembly to submit spark apps to yarn cluster.
>>>>> And I checked the assembly jars of spark. It contains some yarn classes 
>>>>> which are added during compile time. The yarn classes are not what I 
>>>>> want. 
>>>>> 
>>>>> My question is that is it possible to use other jars to submit spark app 
>>>>> to yarn cluster. 
>>>>> I do not want to use the assembly jar because it has yarn classes which 
>>>>> may overwrite the yarn class in HADOOP_CLASSPATH. If the yarn cluster is 
>>>>> upgraded, even if the YARN apis are same, spark has to be recompiled 
>>>>> against to the new version of yarn.
>>>>> 
>>>>> Any help is appreciated ! Thanks.
>>>>> 
>>>>> -- 
>>>>> Regards
>>>>> Gordon Wang
>>> 
>>> 
>>> 
>>> -- 
>>> Regards
>>> Gordon Wang
> 
> 
> 
> -- 
> Regards
> Gordon Wang

Reply via email to