Cody Burleson created JCR-3588: ---------------------------------- Summary: Response time higher on Node1 with load when Node2 has no load Key: JCR-3588 URL: https://issues.apache.org/jira/browse/JCR-3588 Project: Jackrabbit Content Repository Issue Type: Bug Components: clustering Affects Versions: 2.4.3 Environment: CentOS 6.4 running WebSphere Application Server 7.0.0.19. Jackrabbit cluster configuration with 2 WAS servers. Repository on DB2 9.7. Reporter: Cody Burleson Fix For: 2.4.3 Attachments: JackrabbitCluster-ResponseTime.png, Node1repository.xml, Node2repository.xml
In our performance analysis, we are seeing a strange effect, which we does not make sense to us. It may or may not be a defect, but we need to understand why the effect occurs. In a 2 node cluster, we can run a certain load (reading and writing) directly on Node1 and an equivalent load (reading and writing on Node2). We measure the response time on both nodes, and it's less than 2 seconds. If we stop the load to one of the servers, the response time on the other server triples (with no additional load). See attached image "JackrabbitCluster-ResponseTime.png". The left side of the report shows when only one node (Node1) has load and Node2 has no load. In this case, the response times on Node1 are at about 6 seconds. Then, on the right side of the report, we add an equivalent load to Node2 and then the response times on Node1 drop to 2 seconds. So, the load on Node1 was always consistent, yet ADDING load to Node2 actually improves response time on Node1. Logically, it doesn't make much sense, eh? Someone, please, at least help us understand why this may be happening. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira