Hi all, I have a maybe naive question on providing input to a mapreduce program: how can I specify the input with respect to the hdfs path?
right now I can specify a input file from my local directory, say, hadoop trunk I can also specify an absolute path for a dfs file using where it is actually stored on my local node, eg/, /usr/username/tmp/xxxxx How can I do something like hdfs://inputdata/myinputdata.txt? I always got a "cannot find file" kind of error Furthermore, maybe the input files can already be some sharded outputs from another mapreduce, e.g., myinputdata-0001.txt, myinputdata-0002.txt? Thanks a lot!