Thanks Jeff, I'm planning to use it in standalone mode, OK will use hadoop
2.4 package. Chao!



On Wed, Mar 18, 2015 at 10:56 AM, Jeffrey Jedele <jeffrey.jed...@gmail.com>
wrote:

> What you call "sub-category" are packages pre-built to run on certain
> Hadoop environments. It really depends on where you want to run Spark. As
> far as I know, this is mainly about the included HDFS binding - so if you
> just want to play around with Spark, any of the packages should be fine. I
> wouldn't use source though, because you'd have to compile it yourself.
>
> PS: Make sure to use "Reply to all". If you're not including the mailing
> list in the response, I'm the only one who will get your message.
>
> Regards,
> Jeff
>
> 2015-03-18 10:49 GMT+01:00 James King <jakwebin...@gmail.com>:
>
>> Any sub-category recommendations hadoop, MapR, CDH?
>>
>> On Wed, Mar 18, 2015 at 10:48 AM, James King <jakwebin...@gmail.com>
>> wrote:
>>
>>> Many thanks Jeff will give it a go.
>>>
>>> On Wed, Mar 18, 2015 at 10:47 AM, Jeffrey Jedele <
>>> jeffrey.jed...@gmail.com> wrote:
>>>
>>>> Probably 1.3.0 - it has some improvements in the included Kafka
>>>> receiver for streaming.
>>>>
>>>> https://spark.apache.org/releases/spark-release-1-3-0.html
>>>>
>>>> Regards,
>>>> Jeff
>>>>
>>>> 2015-03-18 10:38 GMT+01:00 James King <jakwebin...@gmail.com>:
>>>>
>>>>> Hi All,
>>>>>
>>>>> Which build of Spark is best when using Kafka?
>>>>>
>>>>> Regards
>>>>> jk
>>>>>
>>>>
>>>>
>>>
>>
>

Reply via email to