Now that Kafka 0.8.2.0 has been released, adding external/kafka module
works.

FYI

On Sun, Jan 18, 2015 at 7:36 PM, Ted Yu <yuzhih...@gmail.com> wrote:

> bq. there was no 2.11 Kafka available
>
> That's right. Adding external/kafka module resulted in:
>
> [ERROR] Failed to execute goal on project spark-streaming-kafka_2.11:
> Could not resolve dependencies for project
> org.apache.spark:spark-streaming-kafka_2.11:jar:1.3.0-SNAPSHOT: Could not
> find artifact org.apache.kafka:kafka_2.11:jar:0.8.0 in central (
> https://repo1.maven.org/maven2) -> [Help 1]
>
> Cheers
>
> On Sun, Jan 18, 2015 at 10:41 AM, Sean Owen <so...@cloudera.com> wrote:
>
>> I could be wrong, but I thought this was on purpose. At the time it
>> was set up, there was no 2.11 Kafka available? or one of its
>> dependencies wouldn't work with 2.11?
>>
>> But I'm not sure what the OP means by "maven doesn't build Spark's
>> dependencies" because Ted indicates it does, and of course you can see
>> that these artifacts are published.
>>
>> On Sun, Jan 18, 2015 at 2:46 AM, Ted Yu <yuzhih...@gmail.com> wrote:
>> > There're 3 jars under lib_managed/jars directory with and without
>> > -Dscala-2.11 flag.
>> >
>> > Difference between scala-2.10 and scala-2.11 profiles is that scala-2.10
>> > profile has the following:
>> >       <modules>
>> >         <module>external/kafka</module>
>> >       </modules>
>> >
>> > FYI
>> >
>> > On Sat, Jan 17, 2015 at 4:07 PM, Ted Yu <yuzhih...@gmail.com> wrote:
>> >>
>> >> I did the following:
>> >>  1655  dev/change-version-to-2.11.sh
>> >>  1657  mvn -DHADOOP_PROFILE=hadoop-2.4 -Pyarn,hive -Phadoop-2.4
>> >> -Dscala-2.11 -DskipTests clean package
>> >>
>> >> And mvn command passed.
>> >>
>> >> Did you see any cross-compilation errors ?
>> >>
>> >> Cheers
>> >>
>> >> BTW the two links you mentioned are consistent in terms of building for
>> >> Scala 2.11
>> >>
>> >> On Sat, Jan 17, 2015 at 3:43 PM, Walrus theCat <walrusthe...@gmail.com
>> >
>> >> wrote:
>> >>>
>> >>> Hi,
>> >>>
>> >>> When I run this:
>> >>>
>> >>> dev/change-version-to-2.11.sh
>> >>> mvn -Pyarn -Phadoop-2.4 -Dscala-2.11 -DskipTests clean package
>> >>>
>> >>> as per here, maven doesn't build Spark's dependencies.
>> >>>
>> >>> Only when I run:
>> >>>
>> >>> dev/change-version-to-2.11.sh
>> >>> sbt/sbt -Pyarn -Phadoop-2.4 -Dscala-2.11 -DskipTests  clean package
>> >>>
>> >>> as gathered from here, do I get Spark's dependencies built without any
>> >>> cross-compilation errors.
>> >>>
>> >>> Question:
>> >>>
>> >>> - How can I make maven do this?
>> >>>
>> >>> - How can I specify the use of Scala 2.11 in my own .pom files?
>> >>>
>> >>> Thanks
>> >>
>> >>
>> >
>>
>
>

Reply via email to