For #1, please take a look at hadoop-hdfs-project/hadoop-hdfs-client/src/main/java/org/apache/hadoop/hdfs/DFSClient.java
e.g. the following methods: public DFSInputStream open(String src) throws IOException { public HdfsDataOutputStream append(final String src, final int buffersize, EnumSet<CreateFlag> flag, final Progressable progress, final FileSystem.Statistics statistics) throws IOException { Cheers On Wed, Feb 17, 2016 at 3:40 PM, Arun Patel <arunp.bigd...@gmail.com> wrote: > I would like to store large documents (over 100 MB) on HDFS and insert > metadata in HBase. > > 1) Users will use HBase REST API for PUT and GET requests for storing and > retrieving documents. In this case, how to PUT and GET documents to/from > HDFS?What are the recommended ways for storing and accessing document > to/from HDFS that provides optimum performance? > > Can you please share any sample code? or a Github project? > > 2) What are the performance issues I need to know? > > Regards, > Arun >