Hello all,
We have a new cluster with 2 servers. Every server contains 2 instances of Tomcat 6.0.32. Every node has an Aapche 2.2.21 installed an mod_jk is configured. In front of the these cluster there is a hardware loadbalancer cluster for HA. Several times we get in the status manager of the mod_jk, that one or more worker get the state ERR, and then they get back. In the mod_jk log there are entries like (worker2) Tomcat is down. or (worker1) connection to Tomcat failed. But these worker are locally on the same server as the mod_jk! More stupid is, that the mod_jk of the other node don´t recordnized, that the node on the other physical server is down. Is there a big bug in my config? In server.xml of Node 1 / TC1 Membership Address:228.0.0.4 Bind:172.30.5.78 (local IP of the server) Port: 45564 Frequency:500 dropTime:3000 Receiver Address:172.30.5.78 Port:4000 autoBind=100 selectorTimeout:5000 maxThreads:6 In server.xml Node 1 / TC2 Membership Address:228.0.0.4 Bind:172.30.5.78 (local IP of the server) Port: 45574 Frequency:500 dropTime:3000 Receiver Address:172.30.5.78 Port:4000 autoBind=100 selectorTimeout:5000 maxThreads:6 In server.xml Node 2 / TC 1 Membership Address:228.0.0.4 Bind:172.30.5.77 (local IP of the server) Port: 45564 Frequency:500 dropTime:3000 Receiver Address:172.30.5.77 Port:4000 autoBind=100 selectorTimeout:5000 maxThreads:6 In server.xml / Node 2 / TC 2 Membership Address:228.0.0.4 Bind:172.30.5.77 (local IP of the server) Port: 45574 Frequency:500 dropTime:3000 Receiver Address:172.30.5.77 Port:4000 autoBind=100 selectorTimeout:5000 maxThreads:6 worker.properties: # List the workers name worker.list= loadbalancer,loadbalancertc,jkstatus,worker3,worker4,worker11,worker12 worker.maintain= 60 # ---------------- # First worker - LB # ---------------- worker.worker1.port=8010 worker.worker1.host=172.30.5.78 worker.worker1.type=ajp13 worker.worker1.lbfactor=100 worker.worker1.route=worker1 worker.worker1.connection_pool_timeout=600 worker.worker1.activation=active # ---------------- # Second worker - LB # ---------------- worker.worker2.port=8010 worker.worker2.host=172.30.5.77 worker.worker2.type=ajp13 worker.worker2.lbfactor=100 worker.worker2.route=worker2 worker.worker2.connection_pool_timeout=600 worker.worker2.activation=active # ---------------- # Third worker - Standalone # ---------------- worker.worker3.port=8010 worker.worker3.host=172.30.5.77 worker.worker3.type=ajp13 worker.worker3.lbfactor=100 worker.worker3.activation=active # ---------------- # fourth worker - Standalone # ---------------- worker.worker4.port=8010 worker.worker4.host=172.30.5.78 worker.worker4.type=ajp13 worker.worker4.lbfactor=100 worker.worker4.activation=active # ---------------- # sixth worker TC2010 - LB # ---------------- worker.worker6.port=8012 worker.worker6.host=172.30.5.78 worker.worker6.type=ajp13 worker.worker6.lbfactor=100 worker.worker6.activation=active worker.worker6.route=worker6 worker.worker6.connection_pool_timeout=600 # ---------------- # seventh worker TC3110 - LB # ---------------- worker.worker7.port=8012 worker.worker7.host=172.30.5.77 worker.worker7.type=ajp13 worker.worker7.lbfactor=100 worker.worker7.activation=active worker.worker7.route=worker7 worker.worker7.connection_pool_timeout=600 ##BBMAGK0 # ---------------- # eleventh worker TC2010 - Standalone # ---------------- worker.worker11.port=8012 worker.worker11.host=172.30.5.78 worker.worker11.type=ajp13 worker.worker11.lbfactor=100 worker.worker11.activation=active ##BBMAGK1 # ---------------- # twelfth worker TC2010 - Standalone # ---------------- worker.worker12.port=8012 worker.worker12.host=172.30.5.77 worker.worker12.type=ajp13 worker.worker12.lbfactor=100 worker.worker12.activation=active # ---------------------- # Load Balancer worker # ---------------------- worker.loadbalancer.type=lb worker.loadbalancer.balance_workers=worker1,worker2 worker.loadbalancer.sticky_session=true worker.loadbalancer.sticky_session_force=false worker.loadbalancer.method=Request worker.loadbalancer.retries=5 worker.loadbalancer.secret=t # ---------------------- # Load Balancer worker tc # ---------------------- worker.loadbalancertc.type=lb worker.loadbalancertc.balance_workers=worker6,worker7 worker.loadbalancertc.sticky_session=true worker.loadbalancertc.sticky_session_force=false worker.loadbalancertc.method=Request worker.loadbalancertc.retries=5 worker.loadbalancertc.secret=t # Define a 'jkstatus' worker using status worker.jkstatus.type=status Greetings Alexander
smime.p7s
Description: S/MIME cryptographic signature