RegionServer many socket fds are in CLOSE_WAIT and not getting cleared

2014-07-23 Thread Shankar hiremath
Dear All, Observation: I have HBase cluster with Kerberos enabled, when the Region Server startups then we observed some 20-30 socket fds are in CLOSE_WAIT state, After that when the Region Server starts opening then the socket fds in CLOSE_WAIT starts increasing gradually (almost same as number

Re: how to reduce delay in HDFS restart

2014-07-23 Thread Stanley Shi
Do you have a secondary namenode running? Secondary NN is used for this purpose; Also, if you have HDFS HA enabled, this problem will also not occur. Regards, *Stanley Shi,* On Tue, Jul 22, 2014 at 7:24 AM, Anfernee Xu wrote: > Hi, > > For some reason, all PIDs file are missing in my cluster,

Re: Decommissioning a data node and problems bringing it back online

2014-07-23 Thread Stanley Shi
which distribution are you using? Regards, *Stanley Shi,* On Thu, Jul 24, 2014 at 4:38 AM, andrew touchet wrote: > I should have added this in my first email but I do get an error in the > data node's log file > > '2014-07-12 19:39:58,027 INFO > org.apache.hadoop.hdfs.server.datanode.DataNode

Re: Decommissioning a data node and problems bringing it back online

2014-07-23 Thread andrew touchet
I should have added this in my first email but I do get an error in the data node's log file '2014-07-12 19:39:58,027 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: BlockReport of 0 blocks got processed in 1 msecs' On Wed, Jul 23, 2014 at 3:18 PM, andrew touchet wrote: > Hello, > > I a

Decommissioning a data node and problems bringing it back online

2014-07-23 Thread andrew touchet
Hello, I am Decommissioning data nodes for an OS upgrade on a HPC cluster . Currently, users can run jobs that use data stored on /hdfs. They are able to access all datanodes/compute nodes except the one being decommissioned. Is this safe to do? Will edited files affect the decommissioning node?

HDP hadoop 2.4.1 fails to run mapreduce app

2014-07-23 Thread MrAsanjar .
please help; i have verified mapre-site.xml => mapreduce.framework.name=yarn verified HADOOP_CLASS_PATH in hadoop-env.sh /usr/lib/hadoop/bin/hadoop jar /usr/lib/hadoop-mapreduce/hadoop-mapreduce-examples-*.jar pi 2 5 ... . . Wrote input for Map #0 Wrote inp

Re: Skippin those gost darn 0 byte diles

2014-07-23 Thread Edward Capriolo
Anyway, a solution (seen in Flume if I remember correctly) is having a good file name strategy. For exemple, all new files should end in ".open" and only when they are finished the suffix is removed. Then for processing, you only target the latter. I am not sure this will help. The sequence file r

More than one map task in a node - Hadoop 2x

2014-07-23 Thread rab ra
Hello, I am trying to successfully configure hadoop 2.4.0 to run more than one map task in a node. I have done this in hadoop 1x and I found it was straightforward. But in Hadoop 2x, with yarn coming in, I found bit difficult to follow the documentation. Can someone give me the link or share some

Hadoop streaming - Class not found

2014-07-23 Thread rab ra
Hello, I am trying to run an executable using hadoop streaming 2.4 My executable is my mapper which is a groovy script. This script uses a class from a jar file which I am sending via -libjars argument. The hadoop streaming is made to span maps via an input file, each line feeds to one map. The

Re: Problem starting kerberos secure cluster with different server principals

2014-07-23 Thread Wellington Chevreuil
I believe principal should be the same, something like "hdfs/HOST@YOUR_REALM", as all these processes run as "hdfs" user. On 23 Jul 2014, at 12:42, Ramgopal Nali wrote: > Hi, > > I am using hadoop-2.4.1 > > I have configured different principals for the server processes and tried > start

Problem starting kerberos secure cluster with different server principals

2014-07-23 Thread Ramgopal Nali
Hi, I am using hadoop-2.4.1 I have configured different principals for the server processes and tried starting.(NN and JN principals are same,DN different) Server processes startup have failed. But if i am using the same principals for the process then cluster startup is success. Is it manda

Amended question - How to set up the conf folder

2014-07-23 Thread Chris MacKenzie
>My $HADOOP_CONF_DIR = /scratch/extra/cm469/hadoop-2.4.1/etc/hadoop/conf >Hi, > >Can anyone shed some light on this for me. Every time I attempt to set up >the conf directory, I run into a whole load of errors and ssh issues which >I don¹t see when my config files are in etc/hadoop > >I want to und

How to set up the conf folder

2014-07-23 Thread Chris MacKenzie
Hi, Can anyone shed some light on this for me. Every time I attempt to set up the conf directory, I run into a whole load of errors and ssh issues which I don¹t see when my config files are in etc/hadoop I want to understand how to use the conf directory. My ultimate goal is to use symbolic links

Re: Is it a good idea to delete / move the default configuration xml file ?

2014-07-23 Thread Chris MacKenzie
Hi thanks for that, much appreciated. I guess they are in the jar files then ;O) I was really surprised to see the default configs pulled in, especially considering I thought I was in full control, I did a file search on an installation and saw the files and jumped to the wrong conclusion. I feel

Redirecting to job history server fails in MapReduceV2

2014-07-23 Thread rohit sarewar
Hi I am facing an error during Java Map reduce code execution in using Hadoop 2.2 The FinalApplicationStatus is successful. I can find the output in desired location but It throws error while redirecting to job history server (please refer to red line in the logs below). *Explanation in detail:*