Hello Team, I have one question regarding putting data into hdfs and running mapreduce on data present in hdfs.
1. hdfs is file system and so to interact with it what kind of clients are available? also where do we need to install those client? 2. regarding pig, hive and mapreduce, where do we install them on hadoop cluster and from where do we run all scripts and how does it internally know that it needs to run on node 1, node2 or node 3? any inputs here would really helpful. Thanks, Andy.