Re: Multiple data centre in Hadoop

2012-04-12 Thread Manu S
Hi Abhishek,

1. Use multiple directories for *dfs.name.dir*  *dfs.data.dir* etc
* Recommendation: write to *two local directories on different
physical volumes*, and to an *NFS-mounted* directory
– Data will be preserved even in the event of a total failure of the
NameNode machines
* Recommendation: *soft-mount the NFS* directory
– If the NFS mount goes offline, this will not cause the NameNode
to fail

2. *Rack awareness*
https://issues.apache.org/jira/secure/attachment/12345251/Rack_aware_HDFS_proposal.pdf

On Thu, Apr 12, 2012 at 2:18 AM, Abhishek Pratap Singh
manu.i...@gmail.comwrote:

 Thanks Robert.
 Is there a best practice or design than can address the High Availability
 to certain extent?

 ~Abhishek

 On Wed, Apr 11, 2012 at 12:32 PM, Robert Evans ev...@yahoo-inc.com
 wrote:

  No it does not. Sorry
 
 
  On 4/11/12 1:44 PM, Abhishek Pratap Singh manu.i...@gmail.com wrote:
 
  Hi All,
 
  Just wanted if hadoop supports more than one data centre. This is
 basically
  for DR purposes and High Availability where one centre goes down other
 can
  bring up.
 
 
  Regards,
  Abhishek
 
 




-- 
Thanks  Regards

*Manu S*
SI Engineer - OpenSource  HPC
Wipro Infotech
Mob: +91 8861302855Skype: manuspkd
www.opensourcetalk.co.in


setting client retry

2012-04-12 Thread Rita
In the hdfs-site.xml file what argument do I need to set for client
retries? Also, what is the default parameter?

-- 
--- Get your facts first, then you can distort them as you please.--


Re: setting client retry

2012-04-12 Thread John George
There are several different types of 'client retries'. The following are
some that I know of.

My guess is that you meant the following one. If so, it is defined in
core-site.xml
ipc.client.connect.max.retries (default value: 10) - Indicates the
number of retries a client will make to establish a server connection.

The other type of retries that I can think of on hdfs side:
dfs.client.block.write.retries (default value: 3) - As the name
suggests, this is the number of times a DFS client retries write to the
DataNodes.

dfs.client.block.write.locateFollowingBlock.retries (default value: 5) -
On certain exceptions, the client might retry when trying to get an
additional block from NN and this configuration controls that.


There might be more. Feel free to let me know if you meant something else.

Regards,
John George

-Original Message-
From: Rita rmorgan...@gmail.com
Reply-To: common-user@hadoop.apache.org common-user@hadoop.apache.org
Date: Thu, 12 Apr 2012 07:35:43 -0500
To: common-user@hadoop.apache.org common-user@hadoop.apache.org
Subject: setting client retry

In the hdfs-site.xml file what argument do I need to set for client
retries? Also, what is the default parameter?

-- 
--- Get your facts first, then you can distort them as you please.--



Anybody have experience using NetApp Direct Access Storage (DAS) in a Hadoop cluster?

2012-04-12 Thread kent.stevenson
Netapp has some marketing information about their NetApp Open Solution for 
Hadoop that provides external direct access storage.  (See 
http://www.netapp.com/us/solutions/infrastructure/hadoop.html and the PDF white 
paper http://www.netapp.com/us/library/supportdocs/ds-3237.html)  Anybody have 
any experience with this type of approach vs. the use of traditional internal 
storage?

Thanks,
Kent


UDF compiling

2012-04-12 Thread Barry, Sean F
I am trying to compiling a customized WordCount UDF but I get this cannot
find symbol error when I compile. And I'm not sure how to resolve this
issue.

hduser@master:~ javac -classpath
/usr/lib/hadoop/hadoop-core-0.20.2-cdh3u3.jar WordCount.java
WordCount.java:24: error: cannot find symbol
conf.setMapperClass(WordMapper.class);
^
  symbol:   class WordMapper
  location: class WordCount
WordCount.java:25: error: cannot find symbol
conf.setReducerClass(SumReducer.class);
 ^
  symbol:   class SumReducer
  location: class WordCount
2 errors



hduser@master:~ ls
  SumReducer.classWordMapper.class
  SumReducer.java  WordCount.java  WordMapper.java
hduser@master:~




Re: UDF compiling

2012-04-12 Thread Harsh J
The right command for javac usage in your case is:

javac -classpath /usr/lib/hadoop/hadoop-core-0.20.2-cdh3u3.jar:. WordCount.java

You need . for javac to find the other class files/sources within
the same dir.

On Fri, Apr 13, 2012 at 12:12 AM, Barry, Sean F sean.f.ba...@intel.com wrote:
 I am trying to compiling a customized WordCount UDF but I get this cannot
 find symbol error when I compile. And I'm not sure how to resolve this
 issue.

 hduser@master:~ javac -classpath
 /usr/lib/hadoop/hadoop-core-0.20.2-cdh3u3.jar WordCount.java
 WordCount.java:24: error: cannot find symbol
                conf.setMapperClass(WordMapper.class);
                                    ^
  symbol:   class WordMapper
  location: class WordCount
 WordCount.java:25: error: cannot find symbol
                conf.setReducerClass(SumReducer.class);
                                     ^
  symbol:   class SumReducer
  location: class WordCount
 2 errors



 hduser@master:~ ls
  SumReducer.class        WordMapper.class
  SumReducer.java          WordCount.java  WordMapper.java
 hduser@master:~





-- 
Harsh J


Re: UDF compiling

2012-04-12 Thread Stephen Boesch
HI
   try adding  the directory in which WordCount.class was placed to the
-classpath
4/12 Barry, Sean F sean.f.ba...@intel.com

 I am trying to compiling a customized WordCount UDF but I get this cannot
 find symbol error when I compile. And I'm not sure how to resolve this
 issue.

 hduser@master:~ javac -classpath
 /usr/lib/hadoop/hadoop-core-0.20.2-cdh3u3.jar WordCount.java
 WordCount.java:24: error: cannot find symbol
conf.setMapperClass(WordMapper.class);
^
  symbol:   class WordMapper
  location: class WordCount
 WordCount.java:25: error: cannot find symbol
conf.setReducerClass(SumReducer.class);
 ^
  symbol:   class SumReducer
  location: class WordCount
 2 errors



 hduser@master:~ ls
  SumReducer.classWordMapper.class
  SumReducer.java  WordCount.java  WordMapper.java
 hduser@master:~





Yuan Jin is out of the office.

2012-04-12 Thread Yuan Jin

I will be out of the office starting  04/13/2012 and will not return until
04/16/2012.

I am out of office, and will reply you when I am back.

For HAMSTER related things, you can contact Jason(Deng Peng Zhou/China/IBM)
or Anthony(Fei Xiong/China/IBM)
For CFM related things, you can contact Daniel(Liang SH Su/China/Contr/IBM)
For TMB related things, you can contact Flora(Jun Ying Li/China/IBM)
For TWB related things, you can contact Kim(Yuan SH Jin/China/IBM)
For others, I will reply you when I am back.


Re: Yuan Jin is out of the office.

2012-04-12 Thread Chen He
who cares?

On Thu, Apr 12, 2012 at 6:09 PM, Yuan Jin jiny...@cn.ibm.com wrote:


 I will be out of the office starting  04/13/2012 and will not return until
 04/16/2012.

 I am out of office, and will reply you when I am back.

 For HAMSTER related things, you can contact Jason(Deng Peng Zhou/China/IBM)
 or Anthony(Fei Xiong/China/IBM)
 For CFM related things, you can contact Daniel(Liang SH Su/China/Contr/IBM)
 For TMB related things, you can contact Flora(Jun Ying Li/China/IBM)
 For TWB related things, you can contact Kim(Yuan SH Jin/China/IBM)
 For others, I will reply you when I am back.



Re: Yuan Jin is out of the office.

2012-04-12 Thread Chen He
This is the second time. Pure junk email. Could you avoid sending email to
public mail-list, Ms/Mr. Yuan Jin?

On Thu, Apr 12, 2012 at 6:22 PM, Chen He airb...@gmail.com wrote:

 who cares?


 On Thu, Apr 12, 2012 at 6:09 PM, Yuan Jin jiny...@cn.ibm.com wrote:


 I will be out of the office starting  04/13/2012 and will not return until
 04/16/2012.

 I am out of office, and will reply you when I am back.

 For HAMSTER related things, you can contact Jason(Deng Peng
 Zhou/China/IBM)
 or Anthony(Fei Xiong/China/IBM)
 For CFM related things, you can contact Daniel(Liang SH
 Su/China/Contr/IBM)
 For TMB related things, you can contact Flora(Jun Ying Li/China/IBM)
 For TWB related things, you can contact Kim(Yuan SH Jin/China/IBM)
 For others, I will reply you when I am back.





Re: getting UnknownHostException

2012-04-12 Thread madhu phatak
Please check contents of /etc/hosts for the hostname and ipaddress mapping.

On Thu, Apr 12, 2012 at 11:11 PM, Sujit Dhamale sujitdhamal...@gmail.comwrote:

 Hi Friends ,
 i am getting UnknownHostException while executing Hadoop Word count program

 getting below details from job tracker Web page

 *User:* sujit
 *Job Name:* word count
 *Job File:*

 hdfs://localhost:54310/app/hadoop/tmp/mapred/staging/sujit/.staging/job_201204112234_0002/job.xml
 http://localhost:50030/jobconf.jsp?jobid=job_201204112234_0002
 *Submit Host:* sujit.(null)
 *Submit Host Address:* 127.0.1.1
 *Job-ACLs: All users are allowed*
 *Job Setup:*None
 *Status:* Failed
 *Failure Info:*Job initialization failed: java.net.UnknownHostException:
 sujit.(null) is not a valid Inet address at org.apache.hadoop.net.
 NetUtils.verifyHostnames(NetUtils.java:569) at
 org.apache.hadoop.mapred.JobInProgress.initTasks(JobInProgress.java:711) at
 org.apache.hadoop.mapred.JobTracker.initJob(JobTracker.java:4207) at

 org.apache.hadoop.mapred.EagerTaskInitializationListener$InitJob.run(EagerTaskInitializationListener.java:79)
 at

 java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
 at

 java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
 at java.lang.Thread.run(Thread.java:662)
 *Started at:* Wed Apr 11 22:36:46 IST 2012
 *Failed at:* Wed Apr 11 22:36:47 IST 2012
 *Failed in:* 0sec
 *Job Cleanup:*None




 Can some one help me how to resolve this issue .
 i tried with : http://wiki.apache.org/hadoop/UnknownHost

 but still not able to resolve issue ,
 please help me out .


 Hadoop Version: hadoop-1.0.1.tar.gz
 java version 1.6.0_30
 Operating System : Ubuntu 11.10


 *Note *: All node were up before starting execution of Program

 Kind Regards
 Sujit Dhamale
 http://wiki.apache.org/hadoop/UnknownHost




-- 
https://github.com/zinnia-phatak-dev/Nectar


Re: hadoop debugging in eclipse

2012-04-12 Thread Tim.Wu
google remote debug hadoop in eclipse“. It is super easy.

On Fri, Apr 13, 2012 at 10:42 AM, kewen wang wangkewen...@gmail.com wrote:

 Hello,
  Can you give me some advice about how to debug hadoop src codes in
 eclipse if I want to modify some src codes in hadoop?
 Thanks.

 WangKewen




-- 
Best,
WU Pengcheng ( Tim )


Anyone successfully run Hadoop in pseudo or cluster model under Cygwin?

2012-04-12 Thread Tim.Wu
If yes. Could u send me an email?  Your prompt reply will be appreciated,
because I asked two questions in this mailing list in March, but no one
reply me.

Questions are listed in
http://mail-archives.apache.org/mod_mbox/hadoop-common-user/201203.mbox/%3CCA+2n-oGLocSo3YMUiYzhbtOPzO=11g1rl_b45+y-tggyvzk...@mail.gmail.com%3E

and
http://mail-archives.apache.org/mod_mbox/hadoop-common-user/201203.mbox/%3CCA+2n-oERYM15BKb4dc9KETpjNfucqzd-=coecf2sgclj5u7...@mail.gmail.com%3E


-- 
Best,
WU Pengcheng ( Tim )