Hej again,
You can get task data via the completion events call from your client:
http://hadoop.apache.org/common/docs/r0.20.1/api/org/apache/hadoop/mapreduce/Job.html#getTaskCompletionEvents(int)
They should carry the data you seek.
On Mon, Dec 12, 2011 at 4:47 AM, W.P. McNeill wrote:
> You ca
You are out of luck if you don't want to use NFS, and yet want redundancy
for the NN. Even the new "NN HA" work being done by the community will
require NFS ... and the NFS itself needs to be HA.
But if you use a Netapp, then the likelihood of the Netapp crashing is
lower than the likelihood of a
Hi Arun,
You can enable rack awareness for your hadoop cluster by configuring
the "topology.script.file.name" property.
Please go through this link for more details about rack awareness.
http://hadoop.apache.org/common/docs/r0.19.2/cluster_setup.html#Hadoop+Rack+
Awareness
Devaraj K
This project is integrated into Hadoop Common (JUN 2011).
Hadoop-Snappy can be used as an add-on for recent (released) versions of Hadoop
that do not provide Snappy Codec support yet.
Hadoop-Snappy is being kept in synch with Hadoop Common.
what's this meaning
then how to enable snappy on
Hi Arun,
You can get the Counters object from the job after completion and
then find the counter whichever you want in Counters object using
findCounter api. Please find the sample snippet for accessing the counter
after Job completion
Configuration conf = new Configuration();
Cluster
You need to format the namenode before you start its daemon.
Try "*./hadoop namenode -format* " and then run start-dfs.sh
On Sun, Dec 11, 2011 at 6:25 PM, W.P. McNeill wrote:
> I am trying to run Hadoop as single-node cluster on OS X 10.7 (Lion),
> Hadoop 0.20.203. The namenode does not start a
I am trying to run Hadoop as single-node cluster on OS X 10.7 (Lion),
Hadoop 0.20.203. The namenode does not start and gives no indication
of what is wrong.
> start-dfs.sh
starting namenode, logging to
/tmp/hadoop/logs/hadoop-williammcneill-namenode-William-McNeills-MacBook.local.out
localhost: st
You can read counter values from the Job.getCounters API
(http://hadoop.apache.org/common/docs/r0.20.1/api/org/apache/hadoop/mapreduce/Job.html).
I'm not sure about the other information like execution times. I've
been wondering that myself.
On 12/10/11, ArunKumar wrote:
> Hai guys !
>
> Can i ac
have you looked at kafka. it provides a streaming view of data stream.
flume at the moment is getting rewritten as flume ng
On Dec 6, 2011 4:28 PM, "Praveen Sripati" wrote:
> If the requirement is for real time data processing, using Flume
> will not suffice as there is a time lag between the col