Jignesh
        Don't worry on the old API at this point. Start off with the samples 
try understanding and implementing them. Once you have a firm grip with map 
reduce porting your code from old API to new one would just take a few mins. If 
you are starting on Map reduce I'd suggest you to take the following steps
-understand hadoop, hdfs and mapreduce
-understand the word count example code you already ran
-understand each and every parameters mentioned in the Driver Class along with 
Map and Reduce class, get a grip on the parameters used in map and reduce 
methods and why.
-understand the anotomy and internal working flow of end to end map reduce job.
-look at various types of input and output formats other than default 
TextInputFormat ans where they are to be used
-Look at Writables and usage

With this you should be good enough to code basic map reduce. Get used to the 
common exceptions and its work arounds.  After you get comfortable with this 
try other advance concepts like
-Distributed Cache
-Joins
-Counter
-Reporter 
-cluster admin commands
-fine tune map reduce jobs etc

You would learn these once you do real time map reduce programming.

If you are planing to use Hbase in Map reduce. Better understanding MR first 
and then moving on to HBase.

By the way, Hadoop Definite Guide by Tom White is also an awesome book.
Also to play around you'll get some simple samples at
http://kickstarthadoop.blogspot.com

Hope it helps!.

------Original Message------
From: Jignesh Patel
To: common-user@hadoop.apache.org
ReplyTo: common-user@hadoop.apache.org
Subject: hadoop knowledge gaining
Sent: Oct 7, 2011 19:55

Guys,
I am able to deploy the first program word count using hadoop. I am interesting 
exploring more about hadoop and Hbase and don't know which is the best way to 
grasp both of them.

I have hadoop in action but it has older api. I do also have Hbase definitive 
guide which I have not started exploring.

-Jignesh




Regards
Bejoy K S

Reply via email to