Sorry if I'm asking a silly question... Are you sure your RSs and Datanodes
are all up and running? Are you sure they are collocated?

> Datanode on l-hbase[26-31].data.cn8 and regionserver on
> l-hbase[25-31].data.cn8,

Could be that your only live RS is on l-hbase25.data.cn8, which would cause
that behavior... Btw, why 25th is not collocated with datanode?

Alex Baranau
--
http://cdap.io - open source framework to build and run data applications
on Hadoop & HBase

On Fri, May 15, 2015 at 8:12 PM, Louis Hust <louis.h...@gmail.com> wrote:

> Hi, Esteban,
>
> Hadoop Version 2.2.0, r1537062.
> So i do not know why it always write other datanode instead of local
> datanode,
> If there is some log for the hdfs write policy? And now the cluster is
> working not healthy,
> with heavy networking.
>
> 2015-05-15 1:28 GMT+08:00 Esteban Gutierrez <este...@cloudera.com>:
>
> > Hi Louis,
> >
> > Locality 0 is not right for a cluster of that size and having 3 replicas
> > per block unless all RS cannot connect to the local DN and somehow the
> > local DN to the RS is always excluded from the pipeline. In Hadoop
> > 2.0-alpha there was a bug (HDFS-3224) that caused the NN to report a DN
> as
> > live and dead if the storage ID was changed in a single volume (e.g.
> after
> > replacing one drive) and that caused fs.getFileBlockLocations() to report
> > less blocks for calculating the HDFS locality index. Unless your cluster
> is
> > using Hadoop 2.0-alpha I won't worry too much about that.
> >
> > Regarding the logs its odd that the JN is taking about 1.5 seconds just
> to
> > send less than 200 bytes. Perhaps some IO contention issue is going on in
> > your cluster?
> >
> > thanks,
> > esteban.
> >
> > --
> > Cloudera, Inc.
> >
> >
> > On Thu, May 14, 2015 at 5:48 AM, Louis Hust <louis.h...@gmail.com>
> wrote:
> >
> > > Hi, Esteban
> > >
> > > Each region server has about 122 regions, data is large. HDFS replica
> is
> > > defined as default 3,  and namenode have some WARN like below.
> > >
> > > {log}
> > > 2015-05-14 20:45:37,463 WARN
> > > org.apache.hadoop.hdfs.qjournal.client.QuorumJournalManager: Took
> 1503ms
> > to
> > > send a batch of 3 edits (179 bytes) to remote journal
> 192.168.44.29:8485
> > > {/log}
> > >
> > > Regionserver's log seems normal:
> > >
> > > {log}
> > > 2015-05-14 20:46:59,890 INFO  [Thread-15] regionserver.HRegion:
> Finished
> > > memstore flush of ~44.4 M/46586984, currentsize=0/0 for region
> > >
> > >
> >
> qmq_backup,0066485937885860620cb396a3e65c6c9de92cae9aa29,1412429632233.65684ef65f58cb3e27986ca38d397bee.
> > > in 3141ms, sequenceid=7493455453, compaction requested=true
> > > 2015-05-14 20:46:59,890 INFO
> > >  [regionserver60020-smallCompactions-1431462564717]
> regionserver.HRegion:
> > > Starting compaction on m in region
> > >
> > >
> >
> qmq_backup,0066485937885860620cb396a3e65c6c9de92cae9aa29,1412429632233.65684ef65f58cb3e27986ca38d397bee.
> > > {/log}
> > >
> > > Any idea?
> > >
> > >
> > >
> > > 2015-05-13 1:26 GMT+08:00 Esteban Gutierrez <este...@cloudera.com>:
> > >
> > > > Hi,
> > > >
> > > > How many regions you per RS? one possibility is that you have very
> > little
> > > > data in your cluster and regions have moved around and there are no
> > > blocks
> > > > in the local DN to the RS. Another possibility is that you have one
> > > replica
> > > > configured and regions moved too so that makes even harder to have
> some
> > > > local blocks in the DN to the RS. Lastly it could be some other
> problem
> > > > where the HDFS pipeline has excluded the local DN. Have you seen any
> > > > exception in the RSs or the NameNode that might be interesting?
> > > >
> > > > thanks,
> > > > esteban.
> > > >
> > > >
> > > >
> > > > --
> > > > Cloudera, Inc.
> > > >
> > > >
> > > > On Tue, May 12, 2015 at 2:59 AM, 娄帅 <louis.hust...@gmail.com> wrote:
> > > >
> > > > > Hi, all,
> > > > >
> > > > > I am maintaining an hbase 0.96.0 cluster, but from the web ui of
> > HBase
> > > > > regionserver,
> > > > > i saw Block locality is 0 for all regionserver.
> > > > >
> > > > > Datanode on l-hbase[26-31].data.cn8 and regionserver on
> > > > > l-hbase[25-31].data.cn8,
> > > > >
> > > > > Any idea?
> > > > >
> > > >
> > >
> >
>

Reply via email to