Hadoop might be the wrong technology for you.
Map Reduce is a batch processing mechanism. Also HDFS might be
critical since to access your data you need to close the file - means
you might have many small file, a situation where hdfs is not very
strong (namespace is hold in memory).
Hbase might be an interesting tool for you, also zookeeper if you want
to do something home grown...
On Jun 23, 2008, at 11:31 PM, Vadim Zaliva wrote:
Hi!
I am considering using Hadoop for (almost) realime data processing. I
have data coming every second and I would like to use hadoop cluster
to process
it as fast as possible. I need to be able to maintain some guaranteed
max. processing time, for example under 3 minutes.
Does anybody have experience with using Hadoop in such manner? I will
appreciate if you can share your experience or give me pointers
to some articles or pages on the subject.
Vadim
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
101tec Inc.
Menlo Park, California, USA
http://www.101tec.com