Thanks, Eustache.

There's the link in the second reply to an article I wrote for DZone.

Best regards,

-chanwit

--
Chanwit Kaewkasi
linkedin.com/in/chanwit


On Thu, Mar 20, 2014 at 9:39 PM, Eustache DIEMERT <eusta...@diemert.fr> wrote:
> Hey, do you have a blog post or url I can share ?
>
> This is a quite cool experiment !
>
> E/
>
>
> 2014-03-20 15:01 GMT+01:00 Chanwit Kaewkasi <chan...@gmail.com>:
>
>> Hi Chester,
>>
>> It is on our todo-list but it doesn't work at the moment. The
>> Parallela cores can not be utilized by the JVM. So, Spark will just
>> use its ARM cores. We'll be looking at Parallela again when the JVM
>> supports it.
>>
>> Best regards,
>>
>> -chanwit
>>
>> --
>> Chanwit Kaewkasi
>> linkedin.com/in/chanwit
>>
>>
>> On Thu, Mar 20, 2014 at 8:52 PM, Chester <chesterxgc...@yahoo.com> wrote:
>> > I am curious  to see if you have tried on Parallela supercomputer (16 or
>> > 64 cores) cluster, run spark on that should be fun.
>> >
>> > Chester
>> >
>> > Sent from my iPad
>> >
>> > On Mar 19, 2014, at 9:18 AM, Chanwit Kaewkasi <chan...@gmail.com> wrote:
>> >
>> >> Hi Koert,
>> >>
>> >> There's some NAND flash built-in each node. We mount the NAND flash as
>> >> a local directory for Spark to spill data out.
>> >> A DZone article, also written by me, will tell more about the cluster.
>> >> We really appreciate the design of Spark's RDD done by the Spark team.
>> >> It turned out to be perfect for ARM clusters.
>> >>
>> >> http://www.dzone.com/articles/big-data-processing-arm-0
>> >>
>> >> Another great thing is that our cluster can operate at the room
>> >> temperature (25C / 77F) too.
>> >>
>> >> The board is Cubieboard here it is:
>> >> https://en.wikipedia.org/wiki/Cubieboard#Specification
>> >>
>> >> Best regards,
>> >>
>> >> -chanwit
>> >>
>> >> --
>> >> Chanwit Kaewkasi
>> >> linkedin.com/in/chanwit
>> >>
>> >>
>> >> On Wed, Mar 19, 2014 at 9:43 PM, Koert Kuipers <ko...@tresata.com>
>> >> wrote:
>> >>> i dont know anything about arm clusters.... but it looks great. what
>> >>> are the
>> >>> specs? the nodes have no local disk at all?
>> >>>
>> >>>
>> >>> On Tue, Mar 18, 2014 at 10:36 PM, Chanwit Kaewkasi <chan...@gmail.com>
>> >>> wrote:
>> >>>>
>> >>>> Hi all,
>> >>>>
>> >>>> We are a small team doing a research on low-power (and low-cost) ARM
>> >>>> clusters. We built a 20-node ARM cluster that be able to start
>> >>>> Hadoop.
>> >>>> But as all of you've known, Hadoop is performing on-disk operations,
>> >>>> so it's not suitable for a constraint machine powered by ARM.
>> >>>>
>> >>>> We then switched to Spark and had to say wow!!
>> >>>>
>> >>>> Spark / HDFS enables us to crush Wikipedia articles (of year 2012) of
>> >>>> size 34GB in 1h50m. We have identified the bottleneck and it's our
>> >>>> 100M network.
>> >>>>
>> >>>> Here's the cluster:
>> >>>>
>> >>>> https://dl.dropboxusercontent.com/u/381580/aiyara_cluster/Mk-I_SSD.png
>> >>>>
>> >>>> And this is what we got from Spark's shell:
>> >>>>
>> >>>> https://dl.dropboxusercontent.com/u/381580/aiyara_cluster/result_00.png
>> >>>>
>> >>>> I think it's the first ARM cluster that can process a non-trivial
>> >>>> size
>> >>>> of Big Data.
>> >>>> (Please correct me if I'm wrong)
>> >>>> I really want to thank the Spark team that makes this possible !!
>> >>>>
>> >>>> Best regards,
>> >>>>
>> >>>> -chanwit
>> >>>>
>> >>>> --
>> >>>> Chanwit Kaewkasi
>> >>>> linkedin.com/in/chanwit
>> >>>
>> >>>
>
>

Reply via email to