Hello Team,

I have one question regarding putting data into hdfs and running mapreduce
on data present in hdfs.

   1. hdfs is file system and so to interact with it what kind of clients
   are available? also where do we need to install those client?
   2. regarding pig, hive and mapreduce, where do we install them on hadoop
   cluster and from where do we run all scripts and how does it internally
   know that it needs to run on node 1, node2 or node 3?

any inputs here would really helpful.

Thanks, Andy.

Reply via email to