patch format

2014-04-01 Thread Avinash Kujur
hello everyone, Is there any kind of specific format of the patch in which we need to write the patch file. i submitted one patch but it got failed. i saw some of the patches which were submitted previously. Is there sny kind of format for patch? if it is, then please give me the link for format.

Re: patch format

2014-04-01 Thread Azuryy Yu
There is no special patch format, just using svn diff or git diff to generate the patch, which depeds on your code repository. On Tue, Apr 1, 2014 at 5:14 PM, Avinash Kujur wrote: > > hello everyone, > > Is there any kind of specific format of the patch in which we need to > write the patch fil

Re: patch format

2014-04-01 Thread Azuryy Yu
There is no special patch format, just using svn diff or git diff to generate the patch, which depeds on your code repository. On Tue, Apr 1, 2014 at 5:14 PM, Avinash Kujur wrote: > > hello everyone, > > Is there any kind of specific format of the patch in which we need to > write the patch fil

how to make patch?

2014-04-01 Thread Avinash Kujur
hi, how can i make patch from my updated file. please provide me any link if possible for the procedure. when i submitted the .patch file that gave the error like : MAPREDUCE-5742 patch is being downloaded at Tue Apr 1 13:51:32 UTC 2014 fromhttp://issues.apache.org/jira/secure/attachment/126380

Re: how to make patch?

2014-04-01 Thread Ted Yu
Can you generate patch from root of your workspace (/home/cloudera/hadoop-common) ? $ p1 mapreduce-5742.patch can't find file to patch at input line 6 Perhaps you used the wrong -p or --strip option? The text leading up to this was: -- |diff --git a/home/cloudera/hadoop-com

number of map tasks on yarn

2014-04-01 Thread Libo Yu
Hi all, I pretty much use the default yarn setting to run a word count example on a 3 node cluster. Here are my settings: yarn.nodemanager.resource.memory-mb 8192 yarn.scheduler.minimum-allocation-mb 1024 yarn.scheduler.maximum-allocation-vcores 32 I would expect to see 8192/1024 * 3 = 24 map tas

How to open .gv file for Yarn event model

2014-04-01 Thread Azuryy Yu
Hi, I compiled Yarn event model using maven, but how to open .gv file to view it? Thanks.

how to customize hadoop configuration for a job?

2014-04-01 Thread Libo Yu
Hi all, According to the book "Hadoop; The Definitive Guide", it is possible to use "-D property=value" to override any default or site property in the configuration.   I gave it shot and it is true. The property specified with "-D" is ignored. Then I put the property in an xml file and use "

Re: how to customize hadoop configuration for a job?

2014-04-01 Thread Mingjiang Shi
Hi Libo, Could you post the exact command you run? Meanwhile, could you try to remove the space between -D and the "property=value", i.e. use -Dproperty=value instead of "-D property=value". On Wed, Apr 2, 2014 at 9:48 AM, Libo Yu wrote: > Hi all, > > According to the book "Hadoop; The Definiti

Re: number of map tasks on yarn

2014-04-01 Thread Stanley Shi
map task number is not decided by the resources you need. It's decided by something else. Regards, *Stanley Shi,* On Wed, Apr 2, 2014 at 9:08 AM, Libo Yu wrote: > Hi all, > > I pretty much use the default yarn setting to run a word count example on > a 3 node cluster. Here are my settings: >

Re: number of map tasks on yarn

2014-04-01 Thread Wangda Tan
More specifically, Number of map tasks for each job is depended on InputFormat.getSplits(...). The number of map tasks is as same as number of splits returned by InputFormat.getSplits(...). You can read source code of FileInputFormat to get more understanding about this. Regards, Wangda Tan On

Re: number of map tasks on yarn

2014-04-01 Thread Mingjiang Shi
+1 for Wangda's comment. My 2 cents: There are 2 aspect of the problem: 1. How many maps task in a job. 2. How many map tasks can be run concurrently. For #1, see Wangda's comments. For #2, it depends on the cluster resource. In your case, the cluster will only be able to run 24 map tasks concur

Re: how to customize hadoop configuration for a job?

2014-04-01 Thread unmesha sreeveni
Hi Libo, You can implement your driver code using ToolRunner.So that you can pass your extra configuration through command line instead of editing your code all the time. Driver code public class WordCount extends Configured implements Tool { public static void main(String[] a

why copy from localhost.localdomain failed and throw UnknownHostException

2014-04-01 Thread Li Li
the reducer hangs and the log is: 2014-04-02 14:03:57,235 WARN org.apache.hadoop.mapred.ReduceTask: attempt_201402271518_0260_r_00_0 copy failed: attempt_201402271518_0260_m_01_1 from localhost.localdomain 2014-04-02 14:03:57,235 WARN org.apache.hadoop.mapred.ReduceTask: java.net.UnknownHo

dfs.client.domain.socket.data.traffic

2014-04-01 Thread 宾莉金
If we configure the following conf, is all normal data traffic pass over UNIX domain sockets? dfs.client.read.shortcircuit true dfs.domain.socket.path /var/lib/hadoop/dn_socket dfs.client.domain.socket.data.traffic true -- *Best Regards,* lijin bin

Re: dfs.client.domain.socket.data.traffic

2014-04-01 Thread Mingjiang Shi
How do you define "all normal data traffic"? The short circuit is only possible in cases where the client is co-located with the data. On Wed, Apr 2, 2014 at 2:27 PM, 宾莉金 wrote: > If we configure the following conf, is all normal data traffic pass over > UNIX domain sockets? > > dfs.client.

Re: dfs.client.domain.socket.data.traffic

2014-04-01 Thread 宾莉金
My mistake, if the client is co-located with the data, was the client get the data through UNIX domain sockets? 2014-04-02 14:35 GMT+08:00 Mingjiang Shi : > How do you define "all normal data traffic"? The short circuit is only > possible in cases where the client is co-located with the data. >