Your Hadoop and Hbase cluster is OK. But the point is below
FileNotFoundException.

>> 03/04/21 19:51:58 WARN mapred.JobClient: Use genericOptions for the option 
>> -libjars java.io.FileNotFoundException: File ~/HAMA/hama-0.1.0-dev.jar does 
>> not exist.

When the submit M/R job to hadoop cluster, JobClient will copy the
job's jar and configuration to the map-reduce system directory on the
distributed file-system.

Anyway, It should be a absolute path as a
"/home/you/hama/hama-0.x.x-dev.jar" instead of "~/". BTW, $HAMA_HOME
and, $CLASSPATH will be automatically setted up by ./bin/hama script
with hama-env.sh. Could you show the full configuration contents to
me?

On Fri, Oct 30, 2009 at 1:53 PM,  <[email protected]> wrote:
> Hi,
> I think the paths that I have set up are correct. In fact I went to the Hbase
> home directory, invoked the shell and I typed the list command. This is the 
> O/P
> that I got.
>
> hbase(main):001:0> list
>
> SparseMatrix_randjlmkz
> SparseMatrix_randljaul
> SparseMatrix_randltgca
> SparseMatrix_randmabpy
> SparseMatrix_randmpcgl
> SparseMatrix_randozrcg
> SparseMatrix_randslptv
> SparseMatrix_randugfev
> SparseMatrix_randusion
> SparseMatrix_randxsdbg
> SparseMatrix_randzrgjs
> 13 row(s) in 0.1992 seconds
> hbase(main):002:0>
>
> Similarly I tried to see all the files in my HDFS database
>
> had...@zeus:~/hadoop-0.19.1$ bin/hadoop dfs -ls
>
> Found 4 items
> drwxr-xr-x - hadoop supergroup 0 2003-04-21 17:45
> /user/hadoop/SparseMatrix_TMP_dir
> drwxr-xr-x - hadoop supergroup 0 2003-04-21 17:28
> /user/hadoop/etexts-output
> drwxr-xr-x - hadoop supergroup 0 2003-04-21 06:55 /user/hadoop/wordcount
> drwxr-xr-x - hadoop supergroup 0 2003-04-21 17:27 /user/hadoop/wordcount-1
> had...@zeus:~/hadoop-0.19.1$
>
> So this makes me feel that the paths I have set in the hama-env.sh are 
> correct.
> What do you think ?
> Also can you tell me where are the log directory is created ?
>
> Abhishek Agrawal
>
> SUNY- Buffalo
> (716-435-7122)
>
> On Fri 10/30/09 12:43 AM , "Edward J. Yoon" [email protected] sent:
>> > Should I copy the hadoop, hbase folders to the
>> location /usr/local/src
>> nope, that's just a example. You'll need to set up the path of your
>> hadoop/hbase/jdk.
>>
>> On Fri, Oct 30, 2009 at 1:19 PM,  <aa...@buffa
>> lo.edu> wrote:> Hi,
>> >    I am sorry but my
>> /usr/local/src/ folder is empty. I have stored the folders> for Hadoop and
> Hbase in the ~/ directory.
>> Consequently my hama-site.xml file> looks like this.
>> >
>> >
>> > # The java implementation to use.
>>  Required.> export
>> JAVA_HOME=/usr/lib/jvm/java-6-sun>
>> > # hadoop conf dir. to find the mapreduce
>> cluster.> export
>> HADOOP_CONF_DIR=/home/hadoop/hadoop-0.19.1/conf>
>> > # hbase conf dir. to find the hbase
>> cluster.> export
>> HBASE_CONF_DIR=/home/hadoop/hbase-0.19.3/conf>
>> > # Extra Java CLASSPATH elements.
>>  Optional.> export
>> HAMA_CLASSPATH=$HADOOP_CONF_DIR:$HBASE_CONF_DIR>
>> > # The maximum amount of heap to use, in MB.
>> Default is 1000.> export HBASE_HEAPSIZE=1000
>> >
>> > There is no other reason for making the variable
>> HADOOP_CONF_DIR point to> /usr/local/src/hadoop-0.19.0/conf other than the
>> fact that your hadoop folder is> at the location /usr/local/src.
>> > Should I copy the hadoop, hbase folders to the
>> location /usr/local/src>
>> > Thank You
>> >
>> > Abhishek Agrawal
>> >
>> > SUNY- Buffalo
>> > (716-435-7122)
>> >
>> > On Thu 10/29/09  9:58 PM ,
>> "Edward J. Yoon" edwardy
>> [email protected] sent:>> > had...@zeus:~
>> >> /HAMA$ echo $CLASSPATH>
>> >> >
>> >>
>> /home/hadoop/hadoop-0.19.1/hadoop-0.19.1-core.jar:/home/hadoop/hbase-0.19.3
>> >>
>> /hbase-0.19.3.jar:/home/hadoop/hbase-0.19.3/lib/commons-logging-1.0.4.jar:/
>> >>
>> home/hadoop/hbase-0.19.3/lib/log4j-1.2.13:/home/hadoop/weka.jar:/home/hadoo
>> >>
>> p/HAMA/hama-0.1.0-dev.jar:/home/hadoop/HAMA/hama-0.1.0-dev-examples.jar:.>> 
>> You
> don't need to directly set up classpath.
>> Instead, please change>> the conf/hama-env.sh as described
>> below.>>
>> >> ----
>> >> # Set environment variables here.
>> >>
>> >> # The java implementation to use.
>>  Required.>> export
>> JAVA_HOME=/usr/lib/j2sdk1.5-sun>>
>> >> # hadoop conf dir. to find the mapreduce
>> cluster.>> export
>> HADOOP_CONF_DIR=/usr/local/src/hadoop-0.19.0/conf>>
>> >> # hbase conf dir. to find the hbase
>> cluster.>> export
>> HBASE_CONF_DIR=/usr/local/src/hbase-0.19.0/conf>>
>> >> # Extra Java CLASSPATH elements.
>>  Optional.>> export
>> HAMA_CLASSPATH=$HADOOP_CONF_DIR:$HBASE_CONF_DIR>>
>> >> # The maximum amount of heap to use, in MB.
>> Default is 1000.>> export HBASE_HEAPSIZE=1000
>> >>
>> >> Thanks.
>> >>
>> >> On Fri, Oct 30, 2009 at 5:00 AM,
>>  <aa...@buffa>> lo.edu> wrote:> Hi all,
>> >> > Ã�Â
>> �  �  Okay I>> made the changes suggested by you. Now
>> another problem has come>> up.> I execute the command
>> >> > bin/hama examples rand -m 10 -r 10 100
>> 100 35%>> matA and I get the following
>> O/p>>> > had...@zeus:~
>> >> /HAMA$ bin/hama examples rand -m 10 -r 10
>> 100 100 35% matA>>> > 03/04/21 19:51:51 INFO
>> hama.AbstractMatrix:>> Initializing the matrix storage.>
>> 03/04/21 19:51:57 INFO hama.AbstractMatrix:>> Create Matrix 
>> SparseMatrix_randjlmkz>
>> 03/04/21 19:51:57 INFO hama.AbstractMatrix:>> Create the 100 * 100 random 
>> matrix :>
>> SparseMatrix_randjlmkz>> > Wrote input for Map #0
>> >> > Wrote input for Map #1
>> >> > Wrote input for Map #2
>> >> > Wrote input for Map #3
>> >> > Wrote input for Map #4
>> >> > Wrote input for Map #5
>> >> > Wrote input for Map #6
>> >> > Wrote input for Map #7
>> >> > Wrote input for Map #8
>> >> > Wrote input for Map #9
>> >> > 03/04/21 19:51:58 WARN
>> mapred.JobClient: Use>> GenericOptionsParser for parsing the>
>> arguments. Applications should implement Tool>> for the same.> 03/04/21
> 19:51:58 WARN
>> mapred.JobClient: Use>> genericOptions for the option -libjars>
>> java.io.FileNotFoundException: File>> ~/HAMA/hama-0.1.0-dev.jar does not
>> exist.> Ã�  Ã�Â
>> �>> � at>
>> >>
>> org.apache.hadoop.fs.RawLocalFileSystem.getFileStatus(RawLocalFileSystem.ja
>> >> va:420)> Ã�Â
>> �  �>> � at>
>> >>
>> org.apache.hadoop.fs.FilterFileSystem.getFileStatus(FilterFileSystem.java:2
>> >> 44)> Ã�Â
>> �  �>> � at
>> >>
>> org.apache.hadoop.fs.FileUtil.copy(FileUtil.java:192)>
>> �  �  �>> � at
>> >>
>> org.apache.hadoop.fs.FileSystem.copyFromLocalFile(FileSystem.java:1187)>
>> �  �  �>> � at
>> >>
>> org.apache.hadoop.fs.FileSystem.copyFromLocalFile(FileSystem.java:1163)>
>> �  �  �>> � at
>> >>
>> org.apache.hadoop.fs.FileSystem.copyFromLocalFile(FileSystem.java:1135)>
>> �  �  �>> � at>
>> >>
>> org.apache.hadoop.mapred.JobClient.configureCommandLineOptions(JobClient.ja
>> >> va:693)> Ã�Â
>> �  �>> � at
>> >>
>> org.apache.hadoop.mapred.JobClient.submitJob(JobClient.java:788)>
>> �  �  �>> � at
>> >>
>> org.apache.hadoop.mapred.JobClient.runJob(JobClient.java:1142)>
>> �  �  �>> � at
>> >>
>> org.apache.hama.SparseMatrix.random_mapred(SparseMatrix.java:160)>
>> �  �  �>> � at
>> >>
>> org.apache.hama.examples.RandomMatrix.main(RandomMatrix.java:49)>
>> �  �  �>> � at
>> sun.reflect.NativeMethodAccessorImpl.invoke0(Native>> Method)> Ã�Â
>> �  �>> � at>
>> >>
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:3
>> >> 9)> Ã�Â
>> �  �>> � at>
>> >>
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImp
>> >> l.java:25)> Ã�Â
>> �  �>> � at
>> java.lang.reflect.Method.invoke(Method.java:597)> Ã�Â
>> �  �>> � at>
>> >>
>> org.apache.hadoop.util.ProgramDriver$ProgramDescription.invoke(ProgramDrive
>> >> r.java:68)> Ã�Â
>> �  �>> � at
>> >>
>> org.apache.hadoop.util.ProgramDriver.driver(ProgramDriver.java:141)>
>> �  �  �>> � at
>> >>
>> org.apache.hama.examples.ExampleDriver.main(ExampleDriver.java:34)>
>> �  �  �>> � at
>> sun.reflect.NativeMethodAccessorImpl.invoke0(Native>> Method)> Ã�Â
>> �  �>> � at>
>> >>
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:3
>> >> 9)> Ã�Â
>> �  �>> � at>
>> >>
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImp
>> >> l.java:25)> Ã�Â
>> �  �>> � at
>> java.lang.reflect.Method.invoke(Method.java:597)> Ã�Â
>> �  �>> � at
>> >>
>> org.apache.hadoop.util.RunJar.main(RunJar.java:165)>
>> �  �  �>> � at
>> >>
>> org.apache.hadoop.mapred.JobShell.run(JobShell.java:54)>
>> �  �  �>> � at
>> >>
>> org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65)>
>> �  �  �>> � at
>> >>
>> org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:79)>
>> �  �  �>> � at
>> >>
>> org.apache.hadoop.mapred.JobShell.main(JobShell.java:68)>>> > This is a
> snapshot of the
>> ~/HAMA>> folder>
>> >> > had...@zeus:~
>> >> /HAMA$ ls>
>> >> > bin � build
>> � build.xml>> � conf
>> � hama-0.1.0-dev-examples.jar>> � hama-0.1.0-dev.jar>
>> hama-0.1.0-dev-test.jar � lib>> � src
>> � src-gen>>> > This is the classpath that I have
>> set>> up>
>> >> > had...@zeus:~
>> >> /HAMA$ echo $CLASSPATH>
>> >> >
>> >>
>> /home/hadoop/hadoop-0.19.1/hadoop-0.19.1-core.jar:/home/hadoop/hbase-0.19.3
>> >>
>> /hbase-0.19.3.jar:/home/hadoop/hbase-0.19.3/lib/commons-logging-1.0.4.jar:/
>> >>
>> home/hadoop/hbase-0.19.3/lib/log4j-1.2.13:/home/hadoop/weka.jar:/home/hadoo
>> >>
>> p/HAMA/hama-0.1.0-dev.jar:/home/hadoop/HAMA/hama-0.1.0-dev-examples.jar:.&g
>> t;>> > Now I cannot understand the reason why
>> it cannot>> find ~/HAMA/hama-0.1.0-dev.jar .> It is
>> in the classpath as well in the concerned>> folder.>
>> >> > Thank You
>> >> >
>> >> > Abhishek Agrawal
>> >> >
>> >> > SUNY- Buffalo
>> >> > (716-435-7122)
>> >> >
>> >> > On Wed 10/28/09
>> � 9:06 PM ,>> "Edward J. Yoon"
>> edwardy>> [email protected]
>>  sent:>> > Ã�  Ok..I shall obtain
>> the>> source code from this>>
>> link>> http://svn.apache.org/viewvc/incubator/hama/branches/. This is the>
> correct link>> right ?
>> >> >>
>> >> >> Yes.
>> >> >>
>> >> >> > http://wiki.apache.org/hama/GettingStarted.>> Probably>> a change
> should be made
>> there.>> >> Thanks, I'll update that
>> page.>> >>
>> >> >> On Thu, Oct 29, 2009 at 10:02
>> AM,>> � <aa...@buffa&g
>> t;> lo.edu> wrote:> Hi,>> >> >
>> ���  Ok..I
>> shall>> obtain the source code>> from this
>> link>> http://svn.apache.org/viewvc/incubator/hama/branches/. This is> the
> correct link>> right ?I read the HAMA
>> >> getting started guide on>>
>> wikipedia.> http://wiki.apache.org/hama/GettingStarted.>> Probably>> a change
> should be
>> made>> there.>>> > Thank You
>> >> >> >
>> >> >> > Abhishek Agrawal
>> >> >> >
>> >> >> > SUNY- Buffalo
>> >> >> > (716-435-7122)
>> >> >> >
>> >> >> > On Wed 10/28/09
>> >>
>> ��� 8:44 PM ,>>
>> "Edward J. Yoon">> edwardy>> [email protected]
>> >>  sent:>> Hi.>>
>> >>>> >> >> you should use the
>> hama-0.19 branch>> instead>> of trunk. or update
>> hadoop>> &>> hbase to 0.20.x version.>>
>> >>>> >> >> On Thu, Oct 29, 2009 at
>> 8:05>> AM,>>
>> ��� <aa...@buffa&a
>> mp;g>> t;> lo.edu> wrote:> Hellow,>>
>> >> >>>
>> ���
>> ��>>>>
>> ���
>> ���>>
>> ���
>> ���  The>>
>> version>> of Hadoop used by me is>>
>> hadoop-0.19.1 and the version of>>>> hbase> used by me is hbase-0.19.3. I
>> have>> put the files>> hadoop-0.19.1-core.jar>> and>>> hbase-0.19.3.jar in 
>> the
>> lib>> >>> folder in hama.>> >>
>> >>> >> >> > I am getting the
>> following>> exception>> when I try>> to
>> create a random>> matrix. The> exception>> is first
>> caught in>> >>> HColumnDescriptor.java. I am
>> attaching>> that file>> for your convenience. As>> you can> see on>> line 
>> 197 the
>> HColumnDescriptor>>>> constructor is called. On line 201
>> this>>> constructor call a method>>>> isLegalFamilyName(familyName). If name
>> does>> not end> with : an exception is>> being
>> >> thrown>> >> >
>> >> >> >> > Also where are the
>> log files>> stored>> ?>> >
>> >> >> >> > had...@zeus:~
>> >> >> >> /HAMA$ bin/hama examples
>> rand -m 10>> -r 10>> 2000 2000
>> 30.5%>>>> matrixA>>> >> >
>> >> java.lang.IllegalArgumentException:>>
>> Family names>> must end in a colon:>> path>>>
>> >>
>> ���
>> ���>>
>> ���
>> ���>>
>> ���
>> >>>>
>> ���
>> ��� at>>>
>> >>>> >>
>> >>
>> org.apache.hadoop.hbase.HColumnDescriptor.isLegalFamilyName(HColumnDescript
>> >> >> >>
>> or.java:236)>>>
>> ���
>> ��>>>>
>> ���
>> ���>>
>> ���
>> >>>>
>> ���
>> ��� at>>>
>> >>>> >>
>> >>
>> org.apache.hadoop.hbase.HColumnDescriptor.(HColumnDescriptor.java:201)>&
>> gt;>>>
>> ���
>> ���>>
>> ���
>> ���>>
>> ���
>> >>>>
>> ���
>> ��� at>>>
>> >>>> >>
>> >>
>> org.apache.hadoop.hbase.HColumnDescriptor.(HColumnDescriptor.java:157)>&
>> gt;>>>
>> ���
>> ���>>
>> ���
>> ���>>
>> ���
>> >>>>
>> ���
>> ��� at>>>
>> >>>> >>
>> >>
>> org.apache.hadoop.hbase.HColumnDescriptor.(HColumnDescriptor.java:146)>&
>> gt;>>>
>> ���
>> ���>>
>> ���
>> ���>>
>> ���
>> >>>>
>> ���
>> ��� at>>
>> >>>> >>
>> >>
>> org.apache.hama.HamaAdminImpl.initialJob(HamaAdminImpl.java:80)>>>
>> >>
>> ���
>> ���>>
>> ���
>> ���>>
>> ���
>> >>>>
>> ���
>> ��� at>>
>> >>>> >>
>> >>
>> org.apache.hama.HamaAdminImpl.(HamaAdminImpl.java:70)>>>>>
>> ���
>> ���>>
>> ���
>> ���>>
>> ���
>> >>>>
>> ���
>> ��� at>>>
>> >>>> >>
>> >>
>> org.apache.hama.matrix.AbstractMatrix.setConfiguration(AbstractMatrix.java:
>> >> >> >> 105)>
>> >>
>> ���
>> ��>>>>
>> ���
>> ���>>
>> ���
>> >>>>
>> ���
>> ��� at>>
>> >>>> >>
>> >>
>> org.apache.hama.matrix.SparseMatrix.(SparseMatrix.java:60)>>>>>
>> ���
>> ���>>
>> ���
>> ���>>
>> ���
>> >>>>
>> ���
>> ��� at>>
>> >>>> >>
>> >>
>> org.apache.hama.matrix.SparseMatrix.random_mapred(SparseMatrix.java:120)>
>> ;>> ;>> ;
>> >>
>> ���
>> ���>>
>> ���
>> ���>>
>> ���
>> >>>>
>> ���
>> ��� at>>
>> >>>> >>
>> >>
>> org.apache.hama.examples.RandomMatrix.main(RandomMatrix.java:49)>
>> >> >>
>> >>
>> ���
>> ���>>
>> ���
>> ���>>
>> ���
>> >>>>
>> ���
>> ��� at>>>>
>> sun.reflect.NativeMethodAccessorImpl.invoke0(Native>>
>> Method)>>>
>> ���
>> ��>>>>
>> ���
>> ���>>
>> ���
>> >>>>
>> ���
>> ��� at>>>
>> >>>> >>
>> >>
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:3
>> >> >> >> 9)>
>> >>
>> ���
>> ��>>>>
>> ���
>> ���>>
>> ���
>> >>>>
>> ���
>> ��� at>>>
>> >>>> >>
>> >>
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImp
>> >> >> >> l.java:25)>
>> >>
>> ���
>> ��>>>>
>> ���
>> ���>>
>> ���
>> >>>>
>> ���
>> ��� at>>>>
>> java.lang.reflect.Method.invoke(Method.java:597)>>>
>> ���
>> ��>>>>
>> ���
>> ���>>
>> ���
>> >>>>
>> ���
>> ��� at>>>
>> >>>> >>
>> >>
>> org.apache.hadoop.util.ProgramDriver$ProgramDescription.invoke(ProgramDrive
>> >> >> >> r.java:68)>
>> >>
>> ���
>> ��>>>>
>> ���
>> ���>>
>> ���
>> >>>>
>> ���
>> ��� at>>
>> >>>> >>
>> >>
>> org.apache.hadoop.util.ProgramDriver.driver(ProgramDriver.java:141)>>
>> >>>
>> ���
>> ���>>
>> ���
>> ���>>
>> ���
>> >>>>
>> ���
>> ��� at>>
>> >>>> >>
>> >>
>> org.apache.hama.examples.ExampleDriver.main(ExampleDriver.java:34)>>&
>> gt;> java.lang.IllegalArgumentException:
>> >> Family>> names>> must end in a
>> colon:>> column>>>
>> >>
>> ���
>> ���>>
>> ���
>> ���>>
>> ���
>> >>>>
>> ���
>> ��� at>>>
>> >>>> >>
>> >>
>> org.apache.hadoop.hbase.HColumnDescriptor.isLegalFamilyName(HColumnDescript
>> >> >> >>
>> or.java:236)>>>
>> ���
>> ��>>>>
>> ���
>> ���>>
>> ���
>> >>>>
>> ���
>> ��� at>>>
>> >>>> >>
>> >>
>> org.apache.hadoop.hbase.HColumnDescriptor.(HColumnDescriptor.java:201)>&
>> gt;>>>
>> ���
>> ���>>
>> ���
>> ���>>
>> ���
>> >>>>
>> ���
>> ��� at>>>
>> >>>> >>
>> >>
>> org.apache.hadoop.hbase.HColumnDescriptor.(HColumnDescriptor.java:157)>&
>> gt;>>>
>> ���
>> ���>>
>> ���
>> ���>>
>> ���
>> >>>>
>> ���
>> ��� at>>
>> >>>> >>
>> >>
>> org.apache.hama.matrix.AbstractMatrix.create(AbstractMatrix.java:144)>&g
>> t;>>>
>> ���
>> ���>>
>> ���
>> ���>>
>> ���
>> >>>>
>> ���
>> ��� at>>>
>> >>>> >>
>> >>
>> org.apache.hama.matrix.AbstractMatrix.tryToCreateTable(AbstractMatrix.java:
>> >> >> >> 122)>
>> >>
>> ���
>> ��>>>>
>> ���
>> ���>>
>> ���
>> >>>>
>> ���
>> ��� at>>
>> >>>> >>
>> >>
>> org.apache.hama.matrix.SparseMatrix.(SparseMatrix.java:62)>>>>>
>> ���
>> ���>>
>> ���
>> ���>>
>> ���
>> >>>>
>> ���
>> ��� at>>
>> >>>> >>
>> >>
>> org.apache.hama.matrix.SparseMatrix.random_mapred(SparseMatrix.java:120)>
>> ;>> ;>> ;
>> >>
>> ���
>> ���>>
>> ���
>> ���>>
>> ���
>> >>>>
>> ���
>> ��� at>>
>> >>>> >>
>> >>
>> org.apache.hama.examples.RandomMatrix.main(RandomMatrix.java:49)>
>> >> >>
>> >>
>> ���
>> ���>>
>> ���
>> ���>>
>> ���
>> >>>>
>> ���
>> ��� at>>>>
>> sun.reflect.NativeMethodAccessorImpl.invoke0(Native>>
>> Method)>>>
>> ���
>> ��>>>>
>> ���
>> ���>>
>> ���
>> >>>>
>> ���
>> ��� at>>>
>> >>>> >>
>> >>
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:3
>> >> >> >> 9)>
>> >>
>> ���
>> ��>>>>
>> ���
>> ���>>
>> ���
>> >>>>
>> ���
>> ��� at>>>
>> >>>> >>
>> >>
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImp
>> >> >> >> l.java:25)>
>> >>
>> ���
>> ��>>>>
>> ���
>> ���>>
>> ���
>> >>>>
>> ���
>> ��� at>>>>
>> java.lang.reflect.Method.invoke(Method.java:597)>>>
>> ���
>> ��>>>>
>> ���
>> ���>>
>> ���
>> >>>>
>> ���
>> ��� at>>>
>> >>>> >>
>> >>
>> org.apache.hadoop.util.ProgramDriver$ProgramDescription.invoke(ProgramDrive
>> >> >> >> r.java:68)>
>> >>
>> ���
>> ��>>>>
>> ���
>> ���>>
>> ���
>> >>>>
>> ���
>> ��� at>>
>> >>>> >>
>> >>
>> org.apache.hadoop.util.ProgramDriver.driver(ProgramDriver.java:141)>>
>> >>>
>> ���
>> ���>>
>> ���
>> ���>>
>> ���
>> >>>>
>> ���
>> ��� at>>
>> >>>> >>
>> >>
>> org.apache.hama.examples.ExampleDriver.main(ExampleDriver.java:34)>>&
>> amp;>> gt; >>> >> >
>> >> >> >> >
>> >> >> >> > Abhishek
>> Agrawal>> >> >> >
>> >> >> >> > SUNY-
>> Buffalo>> >> >> >
>> (716-435-7122)>> >> >> >
>> >> >> >> >
>> >> >> >> >
>> >> >> >> >
>> >> >> >> >
>> >> >> >> >
>> >> >> >>
>> >> >> >>
>> >> >> >>
>> >> >> >> --
>> >> >> >> Best Regards, Edward J.
>> Yoon @>> NHN,>> corp.>>
>> edwardy>> >> >> [email protected]http://blog.udanax.org>>>>
>> >>>> >> >>
>> >> >> >
>> >> >> >
>> >> >>
>> >> >>
>> >> >>
>> >> >> --
>> >> >> Best Regards, Edward J. Yoon @
>> NHN,>> corp.>> edwardy
>> >> >> [email protected]http://blog.udanax.org>>>> >>
>> >> >>
>> >> >
>> >> >
>> >>
>> >>
>> >>
>> >> --
>> >> Best Regards, Edward J. Yoon @ NHN,
>> corp.>> edwardy
>> >> [email protected]http://blog.udanax.org>>
>> >>
>> >>
>> >
>> >
>>
>>
>>
>> --
>> Best Regards, Edward J. Yoon @ NHN, corp.
>> edwardy
>> [email protected]http://blog.udanax.org
>>
>>
>>
>
>



-- 
Best Regards, Edward J. Yoon @ NHN, corp.
[email protected]
http://blog.udanax.org

Reply via email to