There're 3 jars under lib_managed/jars directory with and
without -Dscala-2.11 flag.

Difference between scala-2.10 and scala-2.11 profiles is that scala-2.10
profile has the following:
      <modules>
        <module>external/kafka</module>
      </modules>

FYI

On Sat, Jan 17, 2015 at 4:07 PM, Ted Yu <yuzhih...@gmail.com> wrote:

> I did the following:
>  1655  dev/change-version-to-2.11.sh
>  1657  mvn -DHADOOP_PROFILE=hadoop-2.4 -Pyarn,hive -Phadoop-2.4
> -Dscala-2.11 -DskipTests clean package
>
> And mvn command passed.
>
> Did you see any cross-compilation errors ?
>
> Cheers
>
> BTW the two links you mentioned are consistent in terms of building for
> Scala 2.11
>
> On Sat, Jan 17, 2015 at 3:43 PM, Walrus theCat <walrusthe...@gmail.com>
> wrote:
>
>> Hi,
>>
>> When I run this:
>>
>> dev/change-version-to-2.11.sh
>> mvn -Pyarn -Phadoop-2.4 -Dscala-2.11 -DskipTests clean package
>>
>> as per here
>> <https://spark.apache.org/docs/latest/building-spark.html#building-for-scala-211>,
>> maven doesn't build Spark's dependencies.
>>
>> Only when I run:
>>
>> dev/change-version-to-2.11.sh
>> sbt/sbt -Pyarn -Phadoop-2.4 -Dscala-2.11 -DskipTests  clean package
>>
>> as gathered from here 
>> <https://github.com/ScrapCodes/spark-1/blob/patch-3/docs/building-spark.md>, 
>> do I get Spark's dependencies built without any cross-compilation errors.
>>
>> *Question*:
>>
>> - How can I make maven do this?
>>
>> - How can I specify the use of Scala 2.11 in my own .pom files?
>>
>> Thanks
>>
>>
>

Reply via email to