Hi,
Thanks a lot for a helping hand!!
As per the suggestion I have upgraded to 1.4.10. Now it took the option
tcp-smart-accept and tcp-smart-connect. But still i am seeing lots of
retries, and the processes running on remote server several times timeout
estabilishing http connection.
Following is my haproxy.cfg:
global
log 127.0.0.1 local0
maxconn 4096 # Total Max Connections. This is dependent on
ulimit
daemon
##nbproc 4 # Number of processing cores. Dual Dual-core Opteron
is 4 cores for example.
defaults
log global
mode http
clitimeout 60000
srvtimeout 30000
contimeout 4000
##option httpclose # Disable Keepalive
option forceclose
option redispatch
option tcp-smart-accept
option tcp-smart-connect
stats auth xxxxx:xxxx
stats uri /xxxx?xxxx
stats realm Get\ Lost
stats enable
listen http_proxy 10.14.146.190:80
mode http
acl invalid_src src xxxxxx xxxxxx xxxxxx
block if invalid_src
balance roundrobin # Load Balancing algorithm
option httpchk HEAD /check HTTP/1.0
option forwardfor # This sets X-Forwarded-For
cookie SERVERID insert nocache indirect
## Define your servers to balance
server tc1.gobindas.in 10.14.146.179:8080 cookie s1 check maxconn
512
server tc2.gobindas.in 10.14.146.176:8080 cookie s2 check maxconn
512
listen https_proxy 10.14.146.190:8443
mode tcp
option ssl-hello-chk
balance roundrobin
##reqadd FRONT_END_HTTPS:\ on
server tc1.gobindas.in 10.14.146.179:8443 weight 1 maxconn 512 check
server tc2.gobindas.in 10.14.146.176:8443 weight 1 maxconn 512 check
And following is the heartbeat configuration:
logfile /var/log/ha-log
logfacility local0
keepalive 2
deadtime 30
initdead 120
bcast eth0
udpport 694
mcast eth0 225.0.0.1 694 1 0
ucast eth0 10.14.146.169
udp eth0
auto_failback on
node lb1.gobindas.in
node lb2.gobindas.in
Now in new stats page I noticed one thing which was not in 1.3.22 is
LastChk, but I wonder tc1 is showing L7OK/302 in 324ms _and tc2 is showing
L7OK/302 in 104ms _ while currently haproxy is running on LB1 and there are
13 retries at TC2.
Also can this issue be due to time differences between cluster nodes? as I
have seen there is a time difference of around 2 minutes between physical
machine 1 vms and physical machine 2 vms.
Thanks,
Amit Nigam
----- Original Message -----
From: "Guillaume Bourque" <guillaume.bour...@gmail.com>
To: "Amit Nigam" <amitni...@gobindas.in>
Cc: <haproxy@formilux.org>
Sent: Thursday, December 23, 2010 8:01 PM
Subject: Re: node frequently goes down on another physical machine
Hi,
first upgrade to a more recent build of haproxy 1.4.x would be better.
Also you should send us more info of your hearbeat config and haproxy so
that we can understand better your setup.
Bye
Amit Nigam a écrit :
Hi, We have just recently migrated from apache to HAProxy. We are having
2 tomcat(TC1 and TC2) nodes along with 2 LB(LB1 LB2) nodes and other DB
nodes. We created 4 virtual servers out of one physical server. Using 2
servers we have 8 nodes.
We are running two HAProxy instances on LB1 and LB2 one each on a virtual
IP, which is being managed with heartbeat. The problem is when LB1
aquires the virtual IP TC2 shows frequent droppings, I can see thousands
of retries on stats page, while at that time TC1 runs fine. When LB2
aquires the IP, TC1 starts dropping the connections. Another problem is
stat page keeps changing stats related to retries. Even it shows number
of retries as per earlier times.
I have tried several things including as you suggested smart-tcp-connect,
foreclose and smart-tcp-connection, but my HAProxy build (1.3.16) says
unknow options for them.
Please help.
Thanks,
Amit
--
Guillaume Bourque, B.Sc.,
consultant, infrastructures technologiques libres !
Logisoft Technologies inc. http://www.logisoftech.com
514 576-7638, http://ca.linkedin.com/in/GuillaumeBourque/fr