Hello everybody,

I've been playing with Hadoop for a few days, and I'm only starting to
explore it's beauty.

In an attempt to learn on the Grep Example, I ended up wondering whether you
can actually extract from within a map, on which file you are currently
running. 
e.g. Suppose I want to grep through a set of files, and instead of having
only a global response, I need an output per file as well.

> ./bin/hadoop jar hadoop-0.16.1-examples.jar grep input output "au[a-c]"

> input/file1.txt 3 aua
> input/file1.txt 2 aub
> input/file1.txt 1 auc

> input/file2.txt 1 aua
> input/file2.txt 2 aub
> input/file2.txt 3 auc

> 4 aua
> 4 aub
> 4 auc


now this could be really easy to do (just hit the right variable in the
JobConf?) or it could be absolutely impossible, since its hadoop's innate
goal to extract from file-related stuff - I'd really appreciate a hint or a
link to read about this.

regards,
Brensch
-- 
View this message in context: 
http://www.nabble.com/Howto-%3A-Monitor-File-Job-allocation-tp16297900p16297900.html
Sent from the Hadoop core-user mailing list archive at Nabble.com.

Reply via email to