Thanks Stack for pointing us in the right direction. Indeed it was the
tcpNodeDelay setting. We set these to be true.
ipc.server.tcpnodelay ==> true
hbase.ipc.client.tcpnodelay ==> true
All reads that previously had the 40ms overhead are now between 2 and 3
ms like we would expect them to be.
t...
This:
java.nio.channels.SocketChannel[connected local=/10.128.204.225:52949
remote=/10.128.204.225:50010]
Seems to say that the error was between the datanode on the same box as the RS?
Nicolas
On Mon, Jul 30, 2012 at 6:43 PM, Jay T wrote:
A couple of our region servers (in a 16 node cluster) crash
A couple of our region servers (in a 16 node cluster) crashed due to
underlying Data Node errors. I am trying to understand how errors on
remote data nodes impact other region server processes.
*To briefly describe what happened:
*
1) Cluster was in operation. All 16 nodes were up, reads and w
ay 23, 2012 at 10:06 AM, Jay T
wrote:
We are currently on Hbase 0.90 (cdh3u3) and soon will be upgrading to
Hbase
0.94. Our application is written in Python and we use Thrift to connect
to HBase.
Looking at Thrift2 (hbase.thrift) I noticed that TScan struct does not
accept filterString as a para
We are currently on Hbase 0.90 (cdh3u3) and soon will be upgrading to Hbase
0.94. Our application is written in Python and we use Thrift to connect
to HBase.
Looking at Thrift2 (hbase.thrift) I noticed that TScan struct does not
accept filterString as a parameter. This was introduced in HBase 0.9