Re: A question about Hadoop 1 job user id used for group mapping, which could lead to performance degradatioin

2014-01-08 Thread Jian Fang
performance in a relatively big cluster. This behavior is very easy to reproduce by simply running terasort on a cluster. Any suggestion to fix this problem? On Wed, Jan 8, 2014 at 11:18 AM, Jian Fang wrote: > Thanks Vinod for your quick response. It is running in non-secure mode. > > I still

Re: A question about Hadoop 1 job user id used for group mapping, which could lead to performance degradatioin

2014-01-08 Thread Jian Fang
hould do so as the user, not the job-id. It is not just logging, I'd be > surprised if jobs pass. Do you have permissions enabled on HDFS? > > Oh, or is this in non-secure mode (i.e. without kerberos)? > > +Vinod > > > On Jan 7, 2014, at 5:14 PM, Jian Fang > wrote:

A question about Hadoop 1 job user id used for group mapping, which could lead to performance degradatioin

2014-01-07 Thread Jian Fang
Hi, I looked at Hadoop 1.X source code and found some logic that I could not understand. In the org.apache.hadoop.mapred.Child class, there were two UGIs defined as follows. UserGroupInformation current = UserGroupInformation.getCurrentUser(); current.addToken(jt); UserGroupInformat

Re: Why my tests shows Yarn is worse than MRv1 for terasort?

2013-10-22 Thread Jian Fang
Output Format Counters Bytes Written=1 Thanks, John On Tue, Oct 22, 2013 at 2:44 PM, Jian Fang wrote: > Hi, > > I have the same problem. I compared Hadoop 2.2.0 with Hadoop 1.0.3 and it > turned out that the terasort for MR2 is 2 times slower than th

Re: Why my tests shows Yarn is worse than MRv1 for terasort?

2013-10-22 Thread Jian Fang
Hi, I have the same problem. I compared Hadoop 2.2.0 with Hadoop 1.0.3 and it turned out that the terasort for MR2 is 2 times slower than that in MR1. I cannot really believe it. The cluster has 20 nodes with 19 data nodes. My Hadoop 2.2.0 cluster configurations are as follows. mapreduc

Re: HTTP ERROR 500 when call application master proxy URL in Hadoop 2.1.0-beta

2013-09-11 Thread Jian Fang
4 PM, Jian Fang wrote: > Ok, seems there is a jira for this issue. > > https://issues.apache.org/jira/browse/YARN-800 > > > On Mon, Sep 9, 2013 at 3:39 PM, Jian Fang > wrote: > >> Hi, >> >> I need to use the web services in application master, for ex

Re: HTTP ERROR 500 when call application master proxy URL in Hadoop 2.1.0-beta

2013-09-10 Thread Jian Fang
Ok, seems there is a jira for this issue. https://issues.apache.org/jira/browse/YARN-800 On Mon, Sep 9, 2013 at 3:39 PM, Jian Fang wrote: > Hi, > > I need to use the web services in application master, for example, > > curl > http://10.6.179.230:9026/proxy/application_137876

Re: "bad interpreter: Text file busy" and other errors in Hadoop 2.1.0-beta

2013-09-09 Thread Jian Fang
John > > This exception should indicate error from the container process. If the > container process exits with non-zero exit code, it will be logged. > In case of such errors, you'd better look at the per-container log see > what's happening there. > > Jian >

HTTP ERROR 500 when call application master proxy URL in Hadoop 2.1.0-beta

2013-09-09 Thread Jian Fang
Hi, I need to use the web services in application master, for example, curl http://10.6.179.230:9026/proxy/application_1378761541170_0003/ws/v1/mapreduce/info This used to work fine, but after I upgraded to Hadoop 2.1.0-beta, I got the following errors. Error 500 Connection refused HTTP ER

"bad interpreter: Text file busy" and other errors in Hadoop 2.1.0-beta

2013-08-30 Thread Jian Fang
Hi, I upgraded to Hadoop 2.1.0-beta and suddenly I started to see error messages as follows. Exception from container-launch: org.apache.hadoop.util.Shell$ExitCodeException: bash: /var/lib/hadoop/tmp/nm-local-dir/usercache/hadoop/appcache/application_1377823589199_0002/container_1377823589199_000

Re: ClientFinalizer shutdown hook in FileSystem caused problems

2013-04-11 Thread Jian Fang
Ok, I figured out the solution. I need to make sure my own shutdown hook has higher priority than the one in FileSystem. On Thu, Apr 11, 2013 at 12:08 PM, Jian Fang wrote: > Hi, > > We have some customized code to flush data during hadoop shutdown process, > but we run into t

ClientFinalizer shutdown hook in FileSystem caused problems

2013-04-11 Thread Jian Fang
Hi, We have some customized code to flush data during hadoop shutdown process, but we run into the following errors with Hadoop 2.0.3-alpha Exception in thread "Thread-3" java.lang.ExceptionInInitializerError at org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2323)

Re: HDFS-4354 and HDFS-4416

2013-04-09 Thread Jian Fang
Thanks Harsh for your quick response. On Tue, Apr 9, 2013 at 11:45 AM, Harsh J wrote: > These are part of the HDFS-347 branch which should be merged to trunk > soon (and then hopefully, onto one of the upcoming 2.x releases, via a > backport). > > On Tue, Apr 9, 2013 at 10:0

HDFS-4354 and HDFS-4416

2013-04-09 Thread Jian Fang
Hi, I saw HDFS-4354 and HDFS-4416 have been used in Cloudera's CDH4.2 to improve short circuit read. Is there any plan to apply them to Hadoop 2.0.X branches? Seems the short circuit read implementation in current Hadoop 2.0.x branches are quite different from HDFS-4354 and HDFS-4416. Thanks in a

Re: hftp in Hadoop 0.20.2

2012-08-15 Thread Jian Fang
instead on the fuller > dfs.http.address in 1.x and dfs.namenode.http.address in 2.x. > > On Sat, Aug 11, 2012 at 3:45 AM, Jian Fang > wrote: > > Thanks Joey for the clarification. I will ask our hadoop admin to change > > that. > > But it would be great if this could

Re: hftp in Hadoop 0.20.2

2012-08-10 Thread Jian Fang
for > HFTP. > > You should make sure that your settings for dfs.http.address and > dfs.info.port are in sync. So change one of those to match the port > number of the other. > > -Joey > > On Fri, Aug 10, 2012 at 5:41 PM, Jian Fang > wrote: > > Hi Joey, > >

Re: hftp in Hadoop 0.20.2

2012-08-10 Thread Jian Fang
tp.address? Thanks, Jian On Fri, Aug 10, 2012 at 5:08 PM, Joey Echeverria wrote: > Can you post your NN logs? It looks like the NN is not actually > started or is listening on another port for HTTP. > > -Joey > > On Fri, Aug 10, 2012 at 2:38 PM, Jian Fang > wrote: >

Re: hftp in Hadoop 0.20.2

2012-08-10 Thread Jian Fang
Already did that. Connection was rejected. On Fri, Aug 10, 2012 at 2:24 PM, Joey Echeverria wrote: > Try: > > $ telnet pnjhadoopnn01 50070 > > -Joey > > On Fri, Aug 10, 2012 at 1:10 PM, Jian Fang > wrote: > > Here is the property in hdfs-site.xml > >

Re: hftp in Hadoop 0.20.2

2012-08-10 Thread Jian Fang
r what its > dfs.http.address may really be using for its port? > > On Fri, Aug 10, 2012 at 7:48 PM, Jian Fang > wrote: > > Hi Harsh, > > > > Seems the -p requires the root privilege, which I don't have. I run > > "netstat -a | grep 50070", but d

Re: hftp in Hadoop 0.20.2

2012-08-10 Thread Jian Fang
i, Aug 10, 2012 at 1:50 AM, Harsh J wrote: > Jian, > > From your NN, can you get us the output "netstat -anp | grep 50070"? > > On Fri, Aug 10, 2012 at 9:29 AM, Jian Fang > wrote: > > Thanks Harsh. But there is no firewall there, the two clusters are on the >

Re: hftp in Hadoop 0.20.2

2012-08-09 Thread Jian Fang
ll isn't the cause of the connection refused on port 50070 on > the NN and ports 50075 on the DNs here? > > On Fri, Aug 10, 2012 at 1:47 AM, Jian Fang > wrote: > > Hi, > > > > We have a hadoop cluster of version 0.20.2 in production. Now we have > > another n

hftp in Hadoop 0.20.2

2012-08-09 Thread Jian Fang
Hi, We have a hadoop cluster of version 0.20.2 in production. Now we have another new Hadoop cluster using cloudera's CDH3U4. We like to run distcp to copy files between the two clusters. Since the hadoop versions are different, we have to use hftp protocol to copy files based on the hadoop docume