Sitamaran,

That was my mistake, you are right it should be "datanode" at the end of
the path. The full path should be:
/home/user/hadoop/data2.2/hdfs/datanode. Sorry about that.


2013/12/28 Sitaraman Vilayannur <vrsitaramanietfli...@gmail.com>

> Hi Diego,
>  You have mentioned where my datanode directory is:
> /home/user/hadoop/data2.2/hdfs/namenode.
> it should be datanode ?
> Sitaraman
>
> On 12/29/13, Sitaraman Vilayannur <vrsitaramanietfli...@gmail.com> wrote:
> > Thanks Diego, will use the copyToLocal.
> > Sitaraman
> >
> > On 12/29/13, Diego Gutierrez <diego.gutier...@ucsp.edu.pe> wrote:
> >> Hi,
> >>
> >> For the first question, the name of the file containing the response of
> >> MR
> >> process is like that:  part-r-00000. This file is inside the output
> >> directory.
> >>
> >> The second question, to locate the file in your linux file system you
> >> have
> >> to go to your datanode directory. This directory path is in your
> >> hdfs-site.xml file in {HADOOP_INSTALLATION}/etc/hadoop. In my case the
> >> full
> >> path of the MR response file is:
> >>
> /home/user/hadoop/data2.2/hdfs/datanode/current/BP-1422783414-127.0.1.1-1387647958374/current/finalized.
> >>
> >> where my datanode directory is: /home/user/hadoop/data2.2/hdfs/namenode.
> >>
> >> In the /finalized directory, there are files named like that:
> >> blk_1073741836. In one of those files is the MR response file. This
> >> aproach
> >> is not good if you want to open the response with vi for example,
> instead
> >> you can copy the response file into your linux file system using
> >> *copyToLocal* command(
> >>
> https://hadoop.apache.org/docs/current2/hadoop-project-dist/hadoop-common/FileSystemShell.html#copyToLocal
> ).
> >> The syntax could be like that: bin/hadoop dfs -copyToLocal
> >> /output/part-r-00000 /home/user/
> >>
> >> Hope this help
> >>
> >>
> >> 2013/12/28 Sitaraman Vilayannur <vrsitaramanietfli...@gmail.com>
> >>
> >>> Hi all,
> >>>   Thanks much for the pointers. I ran word count it appears,
> >>> successfully. Currently in inspected the output like so,
> >>> /usr/local/Software/hadoop-2.2.0/bin/hadoop fs -cat
> >>> /user/sitaraman/output/*
> >>> Java HotSpot(TM) 64-Bit Server VM warning: You have loaded library
> >>> /usr/local/Software/hadoop-2.2.0/lib/native/libhadoop.so.1.0.0 which
> >>> might have disabled stack guard. The VM will try to fix the stack
> >>> guard now.
> >>> It's highly recommended that you fix the library with 'execstack -c
> >>> <libfile>', or link it with '-z noexecstack'.
> >>> 13/12/28 15:13:06 WARN util.NativeCodeLoader: Unable to load
> >>> native-hadoop library for your platform... using builtin-java classes
> >>> where applicable
> >>> as      2
> >>> do      1
> >>> do:     1
> >>> i       2
> >>> not     1
> >>> say     1
> >>>
> >>> What is the name of the file in the output directory?  Can i read it
> >>> say using emacs or vi where will the directory /user/sitaraman/output
> >>> be in the linux fs, that is how to locate where it is....
> >>> Thanks again for all the help.
> >>> Sitaraman
> >>>
> >>> On 12/28/13, Hardik Pandya <smarty.ju...@gmail.com> wrote:
> >>> > I recently blogged about it - hope it helps
> >>> >
> >>> >
> >>>
> http://letsdobigdata.wordpress.com/2013/12/07/running-hadoop-mapreduce-application-from-eclipse-kepler/
> >>> >
> >>> > Regards,
> >>> > Hardik
> >>> >
> >>> > On Fri, Dec 27, 2013 at 6:53 AM, Sitaraman Vilayannur <
> >>> > vrsitaramanietfli...@gmail.com> wrote:
> >>> >
> >>> >> Hi,
> >>> >>  Would much appreciate a pointer to a mapreduce tutorial which
> >>> >> explains
> >>> >> how i can run a simulated cluster of mapreduce nodes on a single PC
> >>> >> and
> >>> >> write a Java program with the MapReduce Paradigm.
> >>> >>  Thanks very much.
> >>> >> Sitaraman
> >>> >>
> >>> >
> >>>
> >>
> >
>

Reply via email to