Dear All,
Observation:
I have HBase cluster with Kerberos enabled, when the Region Server startups
then we observed some 20-30 socket fds are in CLOSE_WAIT state,
After that when the Region Server starts opening then the socket fds in
CLOSE_WAIT starts increasing gradually (almost same as number
Do you have a secondary namenode running? Secondary NN is used for this
purpose;
Also, if you have HDFS HA enabled, this problem will also not occur.
Regards,
*Stanley Shi,*
On Tue, Jul 22, 2014 at 7:24 AM, Anfernee Xu wrote:
> Hi,
>
> For some reason, all PIDs file are missing in my cluster,
which distribution are you using?
Regards,
*Stanley Shi,*
On Thu, Jul 24, 2014 at 4:38 AM, andrew touchet wrote:
> I should have added this in my first email but I do get an error in the
> data node's log file
>
> '2014-07-12 19:39:58,027 INFO
> org.apache.hadoop.hdfs.server.datanode.DataNode
I should have added this in my first email but I do get an error in the
data node's log file
'2014-07-12 19:39:58,027 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: BlockReport of 0 blocks
got processed in 1 msecs'
On Wed, Jul 23, 2014 at 3:18 PM, andrew touchet wrote:
> Hello,
>
> I a
Hello,
I am Decommissioning data nodes for an OS upgrade on a HPC cluster .
Currently, users can run jobs that use data stored on /hdfs. They are able
to access all datanodes/compute nodes except the one being decommissioned.
Is this safe to do? Will edited files affect the decommissioning node?
please help;
i have verified mapre-site.xml => mapreduce.framework.name=yarn
verified HADOOP_CLASS_PATH in hadoop-env.sh
/usr/lib/hadoop/bin/hadoop jar
/usr/lib/hadoop-mapreduce/hadoop-mapreduce-examples-*.jar pi 2 5
...
.
.
Wrote input for Map #0
Wrote inp
Anyway, a solution (seen in Flume if I remember correctly) is having a good
file name strategy. For exemple, all new files should end in ".open" and
only when they are finished the suffix is removed. Then for processing, you
only target the latter.
I am not sure this will help. The sequence file r
Hello,
I am trying to successfully configure hadoop 2.4.0 to run more than one map
task in a node. I have done this in hadoop 1x and I found it was
straightforward. But in Hadoop 2x, with yarn coming in, I found bit
difficult to follow the documentation. Can someone give me the link or
share some
Hello,
I am trying to run an executable using hadoop streaming 2.4
My executable is my mapper which is a groovy script. This script uses a
class from a jar file which I am sending via -libjars argument.
The hadoop streaming is made to span maps via an input file, each line
feeds to one map.
The
I believe principal should be the same, something like "hdfs/HOST@YOUR_REALM",
as all these processes run as "hdfs" user.
On 23 Jul 2014, at 12:42, Ramgopal Nali wrote:
> Hi,
>
> I am using hadoop-2.4.1
>
> I have configured different principals for the server processes and tried
> start
Hi,
I am using hadoop-2.4.1
I have configured different principals for the server processes and tried
starting.(NN and JN principals are same,DN different)
Server processes startup have failed.
But if i am using the same principals for the process then cluster startup is
success.
Is it manda
>My $HADOOP_CONF_DIR = /scratch/extra/cm469/hadoop-2.4.1/etc/hadoop/conf
>Hi,
>
>Can anyone shed some light on this for me. Every time I attempt to set up
>the conf directory, I run into a whole load of errors and ssh issues which
>I don¹t see when my config files are in etc/hadoop
>
>I want to und
Hi,
Can anyone shed some light on this for me. Every time I attempt to set up
the conf directory, I run into a whole load of errors and ssh issues which
I don¹t see when my config files are in etc/hadoop
I want to understand how to use the conf directory. My ultimate goal is to
use symbolic links
Hi thanks for that, much appreciated.
I guess they are in the jar files then ;O) I was really surprised to see
the default configs pulled in, especially considering I thought I was in
full control, I did a file search on an installation and saw the files and
jumped to the wrong conclusion.
I feel
Hi
I am facing an error during Java Map reduce code execution in using Hadoop
2.2
The FinalApplicationStatus is successful. I can find the output in desired
location but It throws error while redirecting to job history server
(please refer to red line in the logs below).
*Explanation in detail:*
15 matches
Mail list logo