where to find the log info

2011-07-28 Thread Daniel,Wu
Hi everyone, I am new to it, and want to do some debug/log. I'd like to check what the value is for each mapper execution. If I add the following code in bold, where can I find the log info? If I can't do it in this way, how should I do? public void map(Object key, Text value, Context

Re: where to find the log info

2011-07-28 Thread Harsh J
Task logs are written to userlogs directory on the TT nodes. You can view task logs on the JobTracker/TaskTracker web UI for each task at: http://machine:50030/taskdetails.jsp?jobid=JOBIDtipid=TASKID All of syslogs, stdout and stderr logs are available in the links to logs off that page.

Re: File System Counters.

2011-07-28 Thread Harsh J
Raj, There is no overlap. Data read from HDFS FileSystem instances go to HDFS_BYTES_READ, and data read from Local FileSystem instances go to FILE_BYTES_READ. These are two different FileSystems, and have no overlap at all. On Thu, Jul 28, 2011 at 5:56 AM, R V cattiv...@yahoo.com wrote: Hello

RE: where to find the log info

2011-07-28 Thread Devaraj K
Daniel, You can find those std out statements in {LOG Directory}/userlogs/{task attemp id}/stdout file. Same way you can find std err statements in {LOG Directory}/userlogs/{task attemp id}/stderr and log statements in {LOG Directory}/userlogs/{task attemp id}/syslog. Devaraj K -Original

Re: Replication and failure

2011-07-28 Thread Harsh J
Mohit, I believe Tom's book (Hadoop: The Definitive Guide) covers this precisely well. Perhaps others too. Replication is a best-effort sort of thing. If 2 nodes are all that is available, then two replicas are written and one is left to the replica monitor service to replicate later as possible

Reader/Writer problem in HDFS

2011-07-28 Thread Meghana
Hi, We have a job where the map tasks are given the path to an output folder. Each map task writes a single file to that folder. There is no reduce phase. There is another thread, which constantly looks for new files in the output folder. If found, it persists the contents to index, and deletes

Why hadoop 0.20.203 unit test failed

2011-07-28 Thread Yu Li
Hi all, I'm trying to compile and unit testing hadoop 0.20.203, but met with almost the same problem with previous discussion in the mailing list( http://mail-archives.apache.org/mod_mbox/hadoop-general/201105.mbox/%3CBANLkTim68H=8ngbfzmsvrqob9pmy7fv...@mail.gmail.com%3E). Even after setting

Hadoop output contains __temporary

2011-07-28 Thread 刘鎏
Hi, all In my recent work in hadoop, I find that the output dir contains: both _SUCCESS and __temporary. And then the next job would be failed because the input path contains _temporary. How does this happen? And How to avoid this? Thanks for your replies. liuliu --

RE: Reader/Writer problem in HDFS

2011-07-28 Thread Laxman
One approach can be use some .tmp extension while writing. Once the write is completed rename back to original file name. Also, reducer has to filter out .tmp files. This will ensure reducer will not pickup the partial files. We do have the similar scenario where the a/m approach resolved the

Re: Reader/Writer problem in HDFS

2011-07-28 Thread Meghana
Thanks Laxman! That would definitely help things. :) Is there a better FileSystem/other method call to create a file in one go (i.e. atomic i guess?), without having to call create() and then write to the stream? ..meghana On 28 July 2011 16:12, Laxman lakshman...@huawei.com wrote: One

Fwd: HBase Mapreduce cannot find Map class

2011-07-28 Thread air
-- Forwarded message -- From: air cnwe...@gmail.com Date: 2011/7/28 Subject: HBase Mapreduce cannot find Map class To: CDH Users cdh-u...@cloudera.org import java.io.IOException; import java.text.ParseException; import java.text.SimpleDateFormat; import java.util.Date; import

Error in 9000 and 9001 port in hadoop-0.20.2

2011-07-28 Thread Doan Ninh
Hi, I run Hadoop in 4 Ubuntu 11.04 on VirtualBox. On the master node (192.168.1.101), I configure fs.default.name = hdfs:// 127.0.0.1:9000. Then i configure everything on 3 other node When i start the cluster by entering $HADOOP_HOME/bin/start-all.sh on the master node Everything is ok, but the

Re: Error in 9000 and 9001 port in hadoop-0.20.2

2011-07-28 Thread madhu phatak
I had issue using IP address in XML files . You can try to use host names in the place of IP address . On Thu, Jul 28, 2011 at 5:22 PM, Doan Ninh uitnetw...@gmail.com wrote: Hi, I run Hadoop in 4 Ubuntu 11.04 on VirtualBox. On the master node (192.168.1.101), I configure fs.default.name =

Re: Error in 9000 and 9001 port in hadoop-0.20.2

2011-07-28 Thread Doan Ninh
In the first time, i use *hadoop-cluster-1* for 192.168.1.101. That is the hostname of the master node. But, the same error occurs How can i fix it? On Thu, Jul 28, 2011 at 7:07 PM, madhu phatak phatak@gmail.com wrote: I had issue using IP address in XML files . You can try to use host

Re: Hadoop Question

2011-07-28 Thread Joey Echeverria
How about having the slave write to temp file first, then move it to the file the master is monitoring for after they close it? -Joey On Jul 27, 2011, at 22:51, Nitin Khandelwal nitin.khandel...@germinait.com wrote: Hi All, How can I determine if a file is being written to (by any

next gen map reduce

2011-07-28 Thread real great..
In which Hadoop version is next gen introduced? -- Regards, R.V.

Re: Error in 9000 and 9001 port in hadoop-0.20.2

2011-07-28 Thread Nitin Khandelwal
Plz change ur* fs.default.name* to hdfs://192.168.1.101:9000 Thanks, Nitin On 28 July 2011 17:46, Doan Ninh uitnetw...@gmail.com wrote: In the first time, i use *hadoop-cluster-1* for 192.168.1.101. That is the hostname of the master node. But, the same error occurs How can i fix it? On

Re: next gen map reduce

2011-07-28 Thread Thomas Graves
Its currently still on the MR279 branch - http://svn.apache.org/viewvc/hadoop/common/branches/MR-279/. It is planned to be merged to trunk soon. Tom On 7/28/11 7:31 AM, real great.. greatness.hardn...@gmail.com wrote: In which Hadoop version is next gen introduced?

Re: Error in 9000 and 9001 port in hadoop-0.20.2

2011-07-28 Thread Doan Ninh
I changed fs.default.name to hdfs://192.168.1.101:9000. But, the same error as before. I need a help On Thu, Jul 28, 2011 at 7:45 PM, Nitin Khandelwal nitin.khandel...@germinait.com wrote: Plz change ur* fs.default.name* to hdfs://192.168.1.101:9000 Thanks, Nitin On 28 July 2011 17:46,

/tmp/hadoop-oracle/dfs/name is in an inconsistent state

2011-07-28 Thread Daniel,Wu
When I started hadoop, the namenode failed to startup because of the following error. The strange thing is that it says/tmp/hadoop-oracle/dfs/name isinconsistent, but I don't think I have configured it as /tmp/hadoop-oracle/dfs/name. Where should I check for the related configuration?

Re: /tmp/hadoop-oracle/dfs/name is in an inconsistent state

2011-07-28 Thread Uma Maheswara Rao G 72686
Hi, Before starting, you need to format the namenode. ./hdfs namenode -format then this directories will be created. respective configuration is 'dfs.namenode.name.dir' default configurations will exist in hdfs-default.xml. If you want to configure your own directory path, you can add the

Re: next gen map reduce

2011-07-28 Thread Robert Evans
It has not been introduced yet. If you are referring to MRV2. It is targeted to go into the 0.23 release of Hadoop, but is currently on the MR-279 branch. Which should hopefully be merged to trunk in abut a week. --Bobby On 7/28/11 7:31 AM, real great.. greatness.hardn...@gmail.com wrote:

Re: Hadoop-streaming using binary executable c program

2011-07-28 Thread Robert Evans
I am not completely sure what you are getting at. It looks like the output of your c program is (And this is just a guess) NOTE: \t stands for the tab character and in streaming it is used to separate the key from the value \n stands for carriage return and is used to separate individual

RE: Error in 9000 and 9001 port in hadoop-0.20.2

2011-07-28 Thread Laxman
Start the namenode[set fs.default.name to hdfs://192.168.1.101:9000] and check your netstat report [netstat -nlp] to check which port and IP it is binding. Ideally, 9000 should be bound to 192.168.1.101. If yes, configure the same IP in slaves as well. Otw, we may need to revisit your configs

Class loading problem

2011-07-28 Thread Kumar, Ranjan
I have a class to define data I am reading from a MySQL database. According to online tutorials I created a class called MyRecord and extended it from Writable, DBWritable. While running it with hadoop I get a NoSuchMethodException: dataTest$MyRecord.init() I am using 0.21.0 Thanks for your

Re: Class loading problem

2011-07-28 Thread John Armstrong
On Thu, 28 Jul 2011 10:05:57 -0400, Kumar, Ranjan ranjan.kum...@morganstanleysmithbarney.com wrote: I have a class to define data I am reading from a MySQL database. According to online tutorials I created a class called MyRecord and extended it from Writable, DBWritable. While running it with

Re: HBase Mapreduce cannot find Map class

2011-07-28 Thread Stack
See http://hbase.apache.org/apidocs/org/apache/hadoop/hbase/mapreduce/package-summary.html#package_description for some help. St.Ack On Thu, Jul 28, 2011 at 4:04 AM, air cnwe...@gmail.com wrote: -- Forwarded message -- From: air cnwe...@gmail.com Date: 2011/7/28 Subject:

Re: Replication and failure

2011-07-28 Thread Mohit Anchlia
On Thu, Jul 28, 2011 at 12:17 AM, Harsh J ha...@cloudera.com wrote: Mohit, I believe Tom's book (Hadoop: The Definitive Guide) covers this precisely well. Perhaps others too. Replication is a best-effort sort of thing. If 2 nodes are all that is available, then two replicas are written and

Unit testing strategy for map/reduce methods

2011-07-28 Thread W.P. McNeill
I've been playing with unit testing strategies for my Hadoop work. A discussion of techniques and a link to example code here: http://cornercases.wordpress.com/2011/07/28/unit-testing-mapreduce-with-overridden-write-methods/ .

Exporting From Hive

2011-07-28 Thread Bale, Michael
Hi, I was wondering if anyone could help me? Does anyone know if it is possible to include the column headers in an output from a Hive Query? I've had a look through the internet but can't seem to find an answer. If not, is it possible to export the result from a describe table query? If so I

Re: cygwin not connecting to Hadoop server

2011-07-28 Thread Uma Maheswara Rao G 72686
Hi A Df, see inline at :: - Original Message - From: A Df abbey_dragonfor...@yahoo.com Date: Wednesday, July 27, 2011 10:31 pm Subject: Re: cygwin not connecting to Hadoop server To: common-user@hadoop.apache.org common-user@hadoop.apache.org See inline at **. More questions and

Re: OSX starting hadoop error

2011-07-28 Thread Bryan Keller
I am also seeing this error upon startup. I am guessing you are using OS X Lion? It started happening to me after I upgraded to 10.7. Hadoop seems to function properly despite this error showing up, though it is annoying. On Jul 27, 2011, at 12:37 PM, Ben Cuthbert wrote: All When starting

Re: OSX starting hadoop error

2011-07-28 Thread Bryan Keller
FYI, I logged a bug for this: https://issues.apache.org/jira/browse/HADOOP-7489 On Jul 28, 2011, at 11:36 AM, Bryan Keller wrote: I am also seeing this error upon startup. I am guessing you are using OS X Lion? It started happening to me after I upgraded to 10.7. Hadoop seems to function

Re: cygwin not connecting to Hadoop server

2011-07-28 Thread Uma Maheswara Rao G 72686
Hi A Df, see inline at :: - Original Message - From: A Df abbey_dragonfor...@yahoo.com Date: Wednesday, July 27, 2011 10:31 pm Subject: Re: cygwin not connecting to Hadoop server To: common-user@hadoop.apache.org common-user@hadoop.apache.org See inline at **. More questions and

Re: Exporting From Hive

2011-07-28 Thread Ayon Sinha
This is for CLI Use this: set hive.cli.print.header=true; Instead of doing this on the prompt everytime you can change your hive start command to: hive -hiveconf hive.cli.print.header=true But be careful with this setting as quite a few commands stop working with NPE with this on. I

Re: File System Counters.

2011-07-28 Thread R V
Harsh   If this is the case I don't understand something. If I see FILE_BYTES_READ to be non zero for a map, the only thing I can assume is that it came  from a spill during sort phase.   I have a 10 node cluster, and I ran TeraSort with size 100,000 Bytes ( 1000 records).    My io.sort.mb is

RE: next gen map reduce

2011-07-28 Thread Aaron Baff
Does this mean 0.22.0 has reached stable and will be released as the stable version soon? --Aaron -Original Message- From: Robert Evans [mailto:ev...@yahoo-inc.com] Sent: Thursday, July 28, 2011 6:39 AM To: common-user@hadoop.apache.org Subject: Re: next gen map reduce It has not been

Re: Hadoop Question

2011-07-28 Thread George Datskos
Nitin, On 2011/07/28 14:51, Nitin Khandelwal wrote: How can I determine if a file is being written to (by any thread) in HDFS. That information is exposed by the NameNode http servlet. You can obtain it with the fsck tool (hadoop fsck /path/to/dir -openforwrite) or you can do an http get

TestDFSIO error: libhdfs.so.1 does not exist

2011-07-28 Thread Yang Xiaoliang
Hi all, I am benchmarking a Hadoop Cluster with the hadoop-*-test.jar TestDFSIO but the following error returns: File /usr/hadoop-0.20.2/libhdfs/libhdfs.so.1 does not exist. How to solve this problem? Thanks!