I was running the Hadoop in HA configuration with two Namenodes, three
Journalnodes and one Datanodes in fully distrubuted mode. I killed the java
process of all the nodes and when I again started the Namenode after
starting the Journalnode, I got the below error in both the Namenodes.
Please clari
When I call getFileBlockLocations() on a DFS, will it return the blocks for
currently-inactive nodes?
If so, how can I filter out the unavailable blocks?
Or more generally, how do I get the list of node status? Is that
ApplicationClientProtocol.getClusterNodes()?
Thanks,
John
If the node is dead, then its blocks are not included. If the node is
stale, its blocks are returned at the end of the array. That is, for a file
with 1 block replicated 3 times on D1, D2, D3 with D1 stale and D2 dead,
the list will be D3; D1.
On Fri, Oct 4, 2013 at 3:22 PM, John Lilley wrote:
>
I am testing scoop to load data from mysql. I am following the example in the
hadoop book. I am running the code in pseudo-distributed mode on Centos.
In the cluster/schedule, the ID is listed with state = ACCEPTED,
finalstatus=UNDEFINED, tracking UI = UNASSIGNED.
If I browse the HDFS file sys
Hi,
I am using hadoop 1.1.1. I want to test to see the snappy compression with
hadoop, but I have some problems to make it work on my Linux environment.
I am using opensuse 12.3 x86_64.
First, when I tried to enable snappy in hadoop 1.1.1 by:
conf.setBoolean("mapred.compress.map.outp
I kind of read the hadoop 1.1.1 source code for this, it is very strange for me
now.
>From the error, it looks like runtime JVM cannot find the native method of
>org/apache/hadoop/io/compress/snappy/SnappyCompressor.compressBytesDirect()I,
>that my guess from the error message, but from the log,