Yes, Sadak, Within this API, you'll copy your files into Hadoop HDFS as you do when writing to an OutputStream. It will be replicated in your cluster's HDFS then.
Cheers. 2011/10/4 visioner sadak <visioner.sa...@gmail.com>: > Hey thanks wellington just a thought will my data be replicated as well coz > i thought tht mapper does the job of breaking data in to pieces and > distribution and reducer will do the joining and combining while fetching > data back thts why was confused to use a MR..can i use this API for > uploading a large number of small files as well thru my application or > should i use sequence file class for that...because i saw the small file > problem in hadoop as well as mentioned in below link > > http://www.cloudera.com/blog/2009/02/the-small-files-problem/ > > On Wed, Oct 5, 2011 at 12:54 AM, Wellington Chevreuil > <wellington.chevre...@gmail.com> wrote: >> >> Hey Sadak, >> >> you don't need to write a MR job for that. You can make your java >> program use Hadoop Java API for that. You would need to use FileSystem >> >> (http://hadoop.apache.org/common/docs/current/api/org/apache/hadoop/fs/FileSystem.html) >> and Path >> (http://hadoop.apache.org/common/docs/current/api/index.html?org/apache/hadoop/fs/Path.html) >> classes for that. >> >> Cheers, >> Wellington. >> >> 2011/10/4 visioner sadak <visioner.sa...@gmail.com>: >> > Hello guys, >> > >> > I would like to know how to do file uploads in HDFS using >> > java,is it to be done using map reduce what if i have a large number of >> > small files should i use sequence file along with map reduce???,It will >> > be >> > great if you can provide some sort of information... > >