Hi all,

I have a maybe naive question on providing input to a mapreduce program:
   how can I specify the input with respect to the hdfs path?

right now I can specify a input file from my local directory, say, hadoop
trunk
I can also specify an absolute path for a dfs file using where it is
actually stored on my local node, eg/, /usr/username/tmp/xxxxx

How can I do something like hdfs://inputdata/myinputdata.txt? I always got a
"cannot find file" kind of error
Furthermore, maybe the input files can already be some sharded outputs from
another mapreduce, e.g., myinputdata-0001.txt, myinputdata-0002.txt?

Thanks a lot!

Reply via email to