Ari, thanks for your note.

Like to understand more how Chukwa group log entries ... If I have appA running 
in machine X, Y and appB running in machine Y, Z.  Each of them calling the 
Chukwa log API.

Do I have all entries going in the same HDFS file ?  or 4 separated HDFS files 
based on the App/Machine combination ?

If the answer of first Q is "yes", then what if appA and appB has different 
format of log entries ?
If the answer of second Q is "yes", then are all these HDFS files cut at the 
same time boundary ?

Looks like in Chukwa, application first log to a daemon, which buffer-write the 
log entries into a local file.  And there is a separate process to ship these 
data to a remote collector daemon which issue the actual HDFS write.  I observe 
the following overhead ...

1) The overhead of extra write to local disk and ship the data over to the 
collector.  If HDFS supports append, the application can then go directly to 
the HDFS

2) The centralized collector establish a bottleneck to the otherwise perfectly 
parallel HDFS architecture.

Am I missing something here ?

Rgds,
Ricky

-----Original Message-----
From: Ariel Rabkin [mailto:asrab...@gmail.com] 
Sent: Monday, April 13, 2009 7:38 AM
To: core-user@hadoop.apache.org
Subject: Re: HDFS as a logfile ??

Chukwa is a Hadoop subproject aiming to do something similar, though
particularly for the case of Hadoop logs.  You may find it useful.

Hadoop unfortunately does not support concurrent appends.  As a
result, the Chukwa project found itself creating a whole new demon,
the chukwa collector, precisely to merge the event streams and write
it out, just once. We're set to do a release within the next week or
two, but in the meantime you can check it out from SVN at
https://svn.apache.org/repos/asf/hadoop/chukwa/trunk

--Ari

On Fri, Apr 10, 2009 at 12:06 AM, Ricky Ho <r...@adobe.com> wrote:
> I want to analyze the traffic pattern and statistics of a distributed 
> application.  I am thinking of having the application write the events as log 
> entries into HDFS and then later I can use a Map/Reduce task to do the 
> analysis in parallel.  Is this a good approach ?
>
> In this case, does HDFS support concurrent write (append) to a file ?  
> Another question is whether the write API thread-safe ?
>
> Rgds,
> Ricky
>



-- 
Ari Rabkin asrab...@gmail.com
UC Berkeley Computer Science Department

Reply via email to