Thanks Jeff, I'm planning to use it in standalone mode, OK will use hadoop 2.4 package. Chao!
On Wed, Mar 18, 2015 at 10:56 AM, Jeffrey Jedele <jeffrey.jed...@gmail.com> wrote: > What you call "sub-category" are packages pre-built to run on certain > Hadoop environments. It really depends on where you want to run Spark. As > far as I know, this is mainly about the included HDFS binding - so if you > just want to play around with Spark, any of the packages should be fine. I > wouldn't use source though, because you'd have to compile it yourself. > > PS: Make sure to use "Reply to all". If you're not including the mailing > list in the response, I'm the only one who will get your message. > > Regards, > Jeff > > 2015-03-18 10:49 GMT+01:00 James King <jakwebin...@gmail.com>: > >> Any sub-category recommendations hadoop, MapR, CDH? >> >> On Wed, Mar 18, 2015 at 10:48 AM, James King <jakwebin...@gmail.com> >> wrote: >> >>> Many thanks Jeff will give it a go. >>> >>> On Wed, Mar 18, 2015 at 10:47 AM, Jeffrey Jedele < >>> jeffrey.jed...@gmail.com> wrote: >>> >>>> Probably 1.3.0 - it has some improvements in the included Kafka >>>> receiver for streaming. >>>> >>>> https://spark.apache.org/releases/spark-release-1-3-0.html >>>> >>>> Regards, >>>> Jeff >>>> >>>> 2015-03-18 10:38 GMT+01:00 James King <jakwebin...@gmail.com>: >>>> >>>>> Hi All, >>>>> >>>>> Which build of Spark is best when using Kafka? >>>>> >>>>> Regards >>>>> jk >>>>> >>>> >>>> >>> >> >