[ https://issues.apache.org/jira/browse/GEODE-1704?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Kevin Duling updated GEODE-1704: -------------------------------- Labels: ci (was: ) > FixedPartitioningDUnitTest.testPut_ValidateDataOnMember_PrimarySecondary_Datastore > ---------------------------------------------------------------------------------- > > Key: GEODE-1704 > URL: https://issues.apache.org/jira/browse/GEODE-1704 > Project: Geode > Issue Type: Bug > Reporter: Kevin Duling > Labels: ci > > In Geode_develop_DistributedTests/3414: > {noformat} > Error Message > com.gemstone.gemfire.test.dunit.RMIException: While invoking > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningDUnitTest$$Lambda$132/528994226.run > in VM 3 running on Host japan.gemstone.com with 4 VMs > Stacktrace > com.gemstone.gemfire.test.dunit.RMIException: While invoking > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningDUnitTest$$Lambda$132/528994226.run > in VM 3 running on Host japan.gemstone.com with 4 VMs > at com.gemstone.gemfire.test.dunit.VM.invoke(VM.java:389) > at com.gemstone.gemfire.test.dunit.VM.invoke(VM.java:355) > at com.gemstone.gemfire.test.dunit.VM.invoke(VM.java:293) > at > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningDUnitTest.testPut_ValidateDataOnMember_PrimarySecondary_Datastore(FixedPartitioningDUnitTest.java:1029) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:497) > at > org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50) > at > org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) > at > org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47) > at > org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) > at > org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) > at > org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) > at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:55) > at org.junit.rules.RunRules.evaluate(RunRules.java:20) > at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:325) > at > org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:78) > at > org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:57) > at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290) > at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71) > at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288) > at org.junit.runners.ParentRunner.access$000(ParentRunner.java:58) > at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268) > at > org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) > at org.junit.runners.ParentRunner.run(ParentRunner.java:363) > at > org.gradle.api.internal.tasks.testing.junit.JUnitTestClassExecuter.runTestClass(JUnitTestClassExecuter.java:112) > at > org.gradle.api.internal.tasks.testing.junit.JUnitTestClassExecuter.execute(JUnitTestClassExecuter.java:56) > at > org.gradle.api.internal.tasks.testing.junit.JUnitTestClassProcessor.processTestClass(JUnitTestClassProcessor.java:66) > at > org.gradle.api.internal.tasks.testing.SuiteTestClassProcessor.processTestClass(SuiteTestClassProcessor.java:51) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:497) > at > org.gradle.messaging.dispatch.ReflectionDispatch.dispatch(ReflectionDispatch.java:35) > at > org.gradle.messaging.dispatch.ReflectionDispatch.dispatch(ReflectionDispatch.java:24) > at > org.gradle.messaging.dispatch.ContextClassLoaderDispatch.dispatch(ContextClassLoaderDispatch.java:32) > at > org.gradle.messaging.dispatch.ProxyDispatchAdapter$DispatchingInvocationHandler.invoke(ProxyDispatchAdapter.java:93) > at com.sun.proxy.$Proxy2.processTestClass(Unknown Source) > at > org.gradle.api.internal.tasks.testing.worker.TestWorker.processTestClass(TestWorker.java:109) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:497) > at > org.gradle.messaging.dispatch.ReflectionDispatch.dispatch(ReflectionDispatch.java:35) > at > org.gradle.messaging.dispatch.ReflectionDispatch.dispatch(ReflectionDispatch.java:24) > at > org.gradle.messaging.remote.internal.hub.MessageHub$Handler.run(MessageHub.java:360) > at > org.gradle.internal.concurrent.ExecutorPolicy$CatchAndRecordFailures.onExecute(ExecutorPolicy.java:54) > at > org.gradle.internal.concurrent.StoppableExecutorImpl$1.run(StoppableExecutorImpl.java:40) > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) > at java.lang.Thread.run(Thread.java:745) > Caused by: java.lang.AssertionError: Failed while creating the cache > at com.gemstone.gemfire.test.dunit.Assert.fail(Assert.java:64) > at > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningTestBase.createCache(FixedPartitioningTestBase.java:151) > at > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningTestBase.createCacheOnMember(FixedPartitioningTestBase.java:131) > at > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningDUnitTest.lambda$testPut_ValidateDataOnMember_PrimarySecondary_Datastore$84651174$4(FixedPartitioningDUnitTest.java:1029) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:497) > at hydra.MethExecutor.executeObject(MethExecutor.java:268) > at > com.gemstone.gemfire.test.dunit.standalone.RemoteDUnitVM.executeMethodOnObject(RemoteDUnitVM.java:82) > at sun.reflect.GeneratedMethodAccessor5.invoke(Unknown Source) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:497) > at sun.rmi.server.UnicastServerRef.dispatch(UnicastServerRef.java:323) > at sun.rmi.transport.Transport$1.run(Transport.java:200) > at sun.rmi.transport.Transport$1.run(Transport.java:197) > at java.security.AccessController.doPrivileged(Native Method) > at sun.rmi.transport.Transport.serviceCall(Transport.java:196) > at > sun.rmi.transport.tcp.TCPTransport.handleMessages(TCPTransport.java:568) > at > sun.rmi.transport.tcp.TCPTransport$ConnectionHandler.run0(TCPTransport.java:826) > at > sun.rmi.transport.tcp.TCPTransport$ConnectionHandler.lambda$run$95(TCPTransport.java:683) > at java.security.AccessController.doPrivileged(Native Method) > at > sun.rmi.transport.tcp.TCPTransport$ConnectionHandler.run(TCPTransport.java:682) > ... 3 more > Caused by: com.gemstone.gemfire.SystemConnectException: One or more peers > generated exceptions during connection attempt > at > com.gemstone.gemfire.distributed.internal.DistributionManager.sendStartupMessage(DistributionManager.java:2675) > at > com.gemstone.gemfire.distributed.internal.DistributionManager.create(DistributionManager.java:557) > at > com.gemstone.gemfire.distributed.internal.InternalDistributedSystem.initialize(InternalDistributedSystem.java:616) > at > com.gemstone.gemfire.distributed.internal.InternalDistributedSystem.newInstance(InternalDistributedSystem.java:241) > at > com.gemstone.gemfire.distributed.DistributedSystem.connect(DistributedSystem.java:238) > at > com.gemstone.gemfire.test.dunit.internal.JUnit4DistributedTestCase.getSystem(JUnit4DistributedTestCase.java:178) > at > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningTestBase.createCache(FixedPartitioningTestBase.java:146) > ... 24 more > Caused by: > com.gemstone.gemfire.distributed.DistributedSystemDisconnectedException: > GemFire on japan(16139)<ec><v118>:1028 started at Tue Jul 26 17:32:11 PDT > 2016: Message distribution has terminated, caused by > com.gemstone.gemfire.ForcedDisconnectException: Member isn't responding to > heartbeat requests > at > com.gemstone.gemfire.distributed.internal.DistributionManager$Stopper.generateCancelledException(DistributionManager.java:699) > at > com.gemstone.gemfire.CancelCriterion.checkCancelInProgress(CancelCriterion.java:91) > at > com.gemstone.gemfire.distributed.internal.DistributionManager.addMemberEvent(DistributionManager.java:2456) > at > com.gemstone.gemfire.distributed.internal.DistributionManager.handleManagerDeparture(DistributionManager.java:3268) > at > com.gemstone.gemfire.distributed.internal.StartupOperation.sendStartupMessage(StartupOperation.java:105) > at > com.gemstone.gemfire.distributed.internal.DistributionManager.sendStartupMessage(DistributionManager.java:2671) > ... 30 more > Caused by: com.gemstone.gemfire.ForcedDisconnectException: Member isn't > responding to heartbeat requests > at > com.gemstone.gemfire.distributed.internal.membership.gms.mgr.GMSMembershipManager.forceDisconnect(GMSMembershipManager.java:2605) > at > com.gemstone.gemfire.distributed.internal.membership.gms.membership.GMSJoinLeave.forceDisconnect(GMSJoinLeave.java:918) > at > com.gemstone.gemfire.distributed.internal.membership.gms.membership.GMSJoinLeave.processRemoveRequest(GMSJoinLeave.java:593) > at > com.gemstone.gemfire.distributed.internal.membership.gms.membership.GMSJoinLeave.processMessage(GMSJoinLeave.java:1555) > at > com.gemstone.gemfire.distributed.internal.membership.gms.messenger.JGroupsMessenger$JGroupsReceiver.receive(JGroupsMessenger.java:1065) > at org.jgroups.JChannel.invokeCallback(JChannel.java:816) > at org.jgroups.JChannel.up(JChannel.java:741) > at org.jgroups.stack.ProtocolStack.up(ProtocolStack.java:1030) > at org.jgroups.protocols.FRAG2.up(FRAG2.java:165) > at org.jgroups.protocols.FlowControl.up(FlowControl.java:390) > at org.jgroups.protocols.UNICAST3.deliverMessage(UNICAST3.java:1070) > at org.jgroups.protocols.UNICAST3.handleDataReceived(UNICAST3.java:785) > at org.jgroups.protocols.UNICAST3.up(UNICAST3.java:426) > at > com.gemstone.gemfire.distributed.internal.membership.gms.messenger.StatRecorder.up(StatRecorder.java:75) > at > com.gemstone.gemfire.distributed.internal.membership.gms.messenger.AddressManager.up(AddressManager.java:75) > at org.jgroups.protocols.TP.passMessageUp(TP.java:1601) > at org.jgroups.protocols.TP$SingleMessageHandler.run(TP.java:1817) > at org.jgroups.util.DirectExecutor.execute(DirectExecutor.java:10) > at org.jgroups.protocols.TP.handleSingleMessage(TP.java:1729) > at org.jgroups.protocols.TP.receive(TP.java:1654) > at > com.gemstone.gemfire.distributed.internal.membership.gms.messenger.Transport.receive(Transport.java:162) > at org.jgroups.protocols.UDP$PacketReceiver.run(UDP.java:701) > ... 1 more > Standard Output > Previously run tests: [PartitionedRegionLoaderWriterDUnitTest, > FixedPartitioningDUnitTest] > [vm_0][info 2016/07/26 17:30:41.110 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Received method: com.gemstone.gemfire.test.dunit.NamedRunnable.run > with 0 args on object: runnable(setupVM) > [vm_1][info 2016/07/26 17:30:41.111 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Received method: com.gemstone.gemfire.test.dunit.NamedRunnable.run > with 0 args on object: runnable(setupVM) > [vm_1][info 2016/07/26 17:30:41.111 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Got result: null > [vm_1] from com.gemstone.gemfire.test.dunit.NamedRunnable.run with 0 args on > object: runnable(setupVM) (took 0 ms) > [vm_0][info 2016/07/26 17:30:41.110 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Got result: null > [vm_0] from com.gemstone.gemfire.test.dunit.NamedRunnable.run with 0 args on > object: runnable(setupVM) (took 0 ms) > [vm_2][info 2016/07/26 17:30:41.112 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Received method: com.gemstone.gemfire.test.dunit.NamedRunnable.run > with 0 args on object: runnable(setupVM) > [vm_2][info 2016/07/26 17:30:41.112 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Got result: null > [vm_2] from com.gemstone.gemfire.test.dunit.NamedRunnable.run with 0 args on > object: runnable(setupVM) (took 0 ms) > [vm_3][info 2016/07/26 17:30:41.113 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Received method: com.gemstone.gemfire.test.dunit.NamedRunnable.run > with 0 args on object: runnable(setupVM) > [vm_3][info 2016/07/26 17:30:41.113 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Got result: null > [vm_3] from com.gemstone.gemfire.test.dunit.NamedRunnable.run with 0 args on > object: runnable(setupVM) (took 0 ms) > [setup] START TEST FixedPartitioningDUnitTest.testDelete_WithoutPut > [vm_0][info 2016/07/26 17:30:41.132 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Received method: > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningDUnitTest$$Lambda$12/1305889715.run > with 0 args on object: > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningDUnitTest$$Lambda$12/1305889715@76d2217e > [vm_0]distributed system properties: {locators=localhost[33275], > enable-cluster-configuration=false, use-cluster-configuration=false, > mcast-port=0, disable-auto-reconnect=true, log-level=info} > [vm_0][info 2016/07/26 17:30:41.138 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Startup Configuration: > [vm_0] ### GemFire Properties defined with api ### > [vm_0]disable-auto-reconnect=true > [vm_0]enable-cluster-configuration=false > [vm_0]locators=localhost[33275] > [vm_0]log-level=info > [vm_0]mcast-port=0 > [vm_0]use-cluster-configuration=false > [vm_0]### GemFire Properties using default values ### > [vm_0]ack-severe-alert-threshold=0 > [vm_0]ack-wait-threshold=15 > [vm_0]archive-disk-space-limit=0 > [vm_0]archive-file-size-limit=0 > [vm_0]async-distribution-timeout=0 > [vm_0]async-max-queue-size=8 > [vm_0]async-queue-timeout=60000 > [vm_0]bind-address= > [vm_0]cache-xml-file=cache.xml > [vm_0]cluster-configuration-dir= > [vm_0]cluster-ssl-ciphers=any > [vm_0]cluster-ssl-enabled=false > [vm_0]cluster-ssl-keystore= > [vm_0]cluster-ssl-keystore-password= > [vm_0]cluster-ssl-keystore-type= > [vm_0]cluster-ssl-protocols=any > [vm_0]cluster-ssl-require-authentication=true > [vm_0]cluster-ssl-truststore= > [vm_0]cluster-ssl-truststore-password= > [vm_0]conflate-events=server > [vm_0]conserve-sockets=true > [vm_0]delta-propagation=true > [vm_0]deploy-working-dir=. > [vm_0]disable-tcp=false > [vm_0]distributed-system-id=-1 > [vm_0]distributed-transactions=false > [vm_0]durable-client-id= > [vm_0]durable-client-timeout=300 > [vm_0]enable-network-partition-detection=false > [vm_0]enable-time-statistics=false > [vm_0]enforce-unique-host=false > [vm_0]gateway-ssl-ciphers=any > [vm_0]gateway-ssl-enabled=false > [vm_0]gateway-ssl-keystore= > [vm_0]gateway-ssl-keystore-password= > [vm_0]gateway-ssl-keystore-type= > [vm_0]gateway-ssl-protocols=any > [vm_0]gateway-ssl-require-authentication=true > [vm_0]gateway-ssl-truststore= > [vm_0]gateway-ssl-truststore-password= > [vm_0]groups= > [vm_0]http-service-bind-address= > [vm_0]http-service-port=7070 > [vm_0]http-service-ssl-ciphers=any > [vm_0]http-service-ssl-enabled=false > [vm_0]http-service-ssl-keystore= > [vm_0]http-service-ssl-keystore-password= > [vm_0]http-service-ssl-keystore-type= > [vm_0]http-service-ssl-protocols=any > [vm_0]http-service-ssl-require-authentication=false > [vm_0]http-service-ssl-truststore= > [vm_0]http-service-ssl-truststore-password= > [vm_0]jmx-manager=false > [vm_0]jmx-manager-access-file= > [vm_0]jmx-manager-bind-address= > [vm_0]jmx-manager-hostname-for-clients= > [vm_0]jmx-manager-http-port=7070 > [vm_0]jmx-manager-password-file= > [vm_0]jmx-manager-port=1099 > [vm_0]jmx-manager-ssl=false > [vm_0]jmx-manager-ssl-ciphers=any > [vm_0]jmx-manager-ssl-enabled=false > [vm_0]jmx-manager-ssl-keystore= > [vm_0]jmx-manager-ssl-keystore-password= > [vm_0]jmx-manager-ssl-keystore-type= > [vm_0]jmx-manager-ssl-protocols=any > [vm_0]jmx-manager-ssl-require-authentication=true > [vm_0]jmx-manager-ssl-truststore= > [vm_0]jmx-manager-ssl-truststore-password= > [vm_0]jmx-manager-start=false > [vm_0]jmx-manager-update-rate=2000 > [vm_0]load-cluster-configuration-from-dir=false > [vm_0]locator-wait-time=0 > [vm_0]lock-memory=false > [vm_0]log-disk-space-limit=0 > [vm_0]log-file= > [vm_0]log-file-size-limit=0 > [vm_0]max-num-reconnect-tries=3 > [vm_0]max-wait-time-reconnect=60000 > [vm_0]mcast-address=239.192.81.1 > [vm_0]mcast-flow-control=1048576, 0.25, 5000 > [vm_0]mcast-recv-buffer-size=1048576 > [vm_0]mcast-send-buffer-size=65535 > [vm_0]mcast-ttl=32 > [vm_0]member-timeout=5000 > [vm_0]membership-port-range=1024-65535 > [vm_0]memcached-bind-address= > [vm_0]memcached-port=0 > [vm_0]memcached-protocol=ASCII > [vm_0]name= > [vm_0]off-heap-memory-size= > [vm_0]redis-bind-address= > [vm_0]redis-password= > [vm_0]redis-port=0 > [vm_0]redundancy-zone= > [vm_0]remote-locators= > [vm_0]remove-unresponsive-client=false > [vm_0]roles= > [vm_0]security-client-accessor= > [vm_0]security-client-accessor-pp= > [vm_0]security-client-auth-init= > [vm_0]security-client-authenticator= > [vm_0]security-client-dhalgo= > [vm_0]security-log-file= > [vm_0]security-log-level=config > [vm_0]security-manager= > [vm_0]security-peer-auth-init= > [vm_0]security-peer-authenticator= > [vm_0]security-peer-verifymember-timeout=1000 > [vm_0]security-post-processor= > [vm_0]server-bind-address= > [vm_0]server-ssl-ciphers=any > [vm_0]server-ssl-enabled=false > [vm_0]server-ssl-keystore= > [vm_0]server-ssl-keystore-password= > [vm_0]server-ssl-keystore-type= > [vm_0]server-ssl-protocols=any > [vm_0]server-ssl-require-authentication=true > [vm_0]server-ssl-truststore= > [vm_0]server-ssl-truststore-password= > [vm_0]socket-buffer-size=32768 > [vm_0]socket-lease-time=60000 > [vm_0]ssl-ciphers=any > [vm_0]ssl-enabled=false > [vm_0]ssl-protocols=any > [vm_0]ssl-require-authentication=true > [vm_0]start-dev-rest-api=false > [vm_0]start-locator= > [vm_0]statistic-archive-file= > [vm_0]statistic-sample-rate=1000 > [vm_0]statistic-sampling-enabled=true > [vm_0]tcp-port=0 > [vm_0]udp-fragment-size=60000 > [vm_0]udp-recv-buffer-size=1048576 > [vm_0]udp-send-buffer-size=65535 > [vm_0]user-command-packages= > [vm_0][info 2016/07/26 17:30:41.140 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Starting membership services > [vm_0][info 2016/07/26 17:30:41.149 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] JGroups channel created (took 9ms) > [vm_0][info 2016/07/26 17:30:41.152 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] GemFire P2P Listener started on null > [vm_0][info 2016/07/26 17:30:41.153 PDT <Geode Failure Detection Server > thread 0> tid=0x78] Started failure detection server thread on > japan.gemstone.com/10.118.32.37:43786. > [vm_0][info 2016/07/26 17:30:41.154 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Attempting to join the distributed system through coordinator > japan(16059:locator)<ec><v0>:1024 using address japan(16120)<ec>:1025 > [locator][info 2016/07/26 17:30:41.155 PDT <unicast receiver,japan-50141> > tid=0x21] received join request from japan(16120)<ec>:1025 > [locator][info 2016/07/26 17:30:41.155 PDT <Geode Membership View Creator> > tid=0x27] View Creator is processing 5 requests for the next membership view > [locator][info 2016/07/26 17:30:41.155 PDT <Geode Membership View Creator> > tid=0x27] preparing new view View[japan(16059:locator)<ec><v0>:1024|42] > members: [japan(16059:locator)<ec><v0>:1024, > japan(16120)<ec><v42>:1025{lead}] shutdown: [japan(16130)<ec><v39>:1025, > japan(16124)<ec><v41>:1026] > [locator]failure detection ports: 17959 43786 > [locator][info 2016/07/26 17:30:41.156 PDT <Geode Membership View Creator> > tid=0x27] finished waiting for responses to view preparation > [locator][info 2016/07/26 17:30:41.156 PDT <Geode Membership View Creator> > tid=0x27] received new view: View[japan(16059:locator)<ec><v0>:1024|42] > members: [japan(16059:locator)<ec><v0>:1024, > japan(16120)<ec><v42>:1025{lead}] shutdown: [japan(16130)<ec><v39>:1025, > japan(16124)<ec><v41>:1026] > [locator]old view is: View[japan(16059:locator)<ec><v0>:1024|41] members: > [japan(16059:locator)<ec><v0>:1024, japan(16130)<ec><v39>:1025{lead}, > japan(16124)<ec><v41>:1026] shutdown: [japan(16124)<ec><v40>:1026] > [vm_0][info 2016/07/26 17:30:41.156 PDT <unicast receiver,japan-21143> > tid=0x74] received new view: View[japan(16059:locator)<ec><v0>:1024|42] > members: [japan(16059:locator)<ec><v0>:1024, > japan(16120)<ec><v42>:1025{lead}] shutdown: [japan(16130)<ec><v39>:1025, > japan(16124)<ec><v41>:1026] > [vm_0]old view is: null > [locator][info 2016/07/26 17:30:41.157 PDT <Geode Membership View Creator> > tid=0x27] Peer locator received new membership view: > View[japan(16059:locator)<ec><v0>:1024|42] members: > [japan(16059:locator)<ec><v0>:1024, japan(16120)<ec><v42>:1025{lead}] > shutdown: [japan(16130)<ec><v39>:1025, japan(16124)<ec><v41>:1026] > [locator][info 2016/07/26 17:30:41.157 PDT <Geode Membership View Creator> > tid=0x27] sending new view View[japan(16059:locator)<ec><v0>:1024|42] > members: [japan(16059:locator)<ec><v0>:1024, > japan(16120)<ec><v42>:1025{lead}] shutdown: [japan(16130)<ec><v39>:1025, > japan(16124)<ec><v41>:1026] > [locator]failure detection ports: 17959 43786 > [vm_0][info 2016/07/26 17:30:41.157 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Finished joining (took 4ms). > [vm_0][info 2016/07/26 17:30:41.157 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Starting DistributionManager japan(16120)<ec><v42>:1025. (took 18 > ms) > [locator][info 2016/07/26 17:30:41.157 PDT <View Message Processor> tid=0x33] > Membership: Processing addition < japan(16120)<ec><v42>:1025 > > [locator][info 2016/07/26 17:30:41.157 PDT <View Message Processor> tid=0x33] > Admitting member <japan(16120)<ec><v42>:1025>. Now there are 2 non-admin > member(s). > [vm_0][info 2016/07/26 17:30:41.160 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Initial (distribution manager) view = > View[japan(16059:locator)<ec><v0>:1024|42] members: > [japan(16059:locator)<ec><v0>:1024, japan(16120)<ec><v42>:1025{lead}] > shutdown: [japan(16130)<ec><v39>:1025, japan(16124)<ec><v41>:1026] > [vm_0][info 2016/07/26 17:30:41.160 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Admitting member <japan(16059:locator)<ec><v0>:1024>. Now there are > 1 non-admin member(s). > [vm_0][info 2016/07/26 17:30:41.160 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Admitting member <japan(16120)<ec><v42>:1025>. Now there are 2 > non-admin member(s). > [locator][info 2016/07/26 17:30:41.163 PDT <Pooled High Priority Message > Processor 1> tid=0x37] Member japan(16120)<ec><v42>:1025 is equivalent or in > the same redundancy zone. > [vm_0][info 2016/07/26 17:30:41.166 PDT <P2P message reader for > japan(16059:locator)<ec><v0>:1024 shared unordered uid=37 port=49582> > tid=0x7e] Member japan(16059:locator)<ec><v0>:1024 is equivalent or in the > same redundancy zone. > [vm_0][info 2016/07/26 17:30:41.170 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Shutting down DistributionManager japan(16120)<ec><v42>:1025. > [locator][info 2016/07/26 17:30:41.172 PDT <Pooled High Priority Message > Processor 1> tid=0x37] received leave request from japan(16120)<ec><v42>:1025 > for japan(16120)<ec><v42>:1025 > [locator][info 2016/07/26 17:30:41.173 PDT <Pooled High Priority Message > Processor 1> tid=0x37] Member at japan(16120)<ec><v42>:1025 gracefully left > the distributed cache: shutdown message received > [vm_0][info 2016/07/26 17:30:41.273 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Now closing distribution for japan(16120)<ec><v42>:1025 > [vm_0][info 2016/07/26 17:30:41.274 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Stopping membership services > [vm_0][info 2016/07/26 17:30:41.275 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] GMSHealthMonitor server socket is closed in stopServices(). > [vm_0][info 2016/07/26 17:30:41.275 PDT <Geode Failure Detection Server > thread 0> tid=0x78] GMSHealthMonitor server thread exiting > [locator][info 2016/07/26 17:30:41.275 PDT <unicast receiver,japan-50141> > tid=0x21] received leave request from japan(16120)<ec><v42>:1025 for > japan(16120)<ec><v42>:1025 > [vm_0][info 2016/07/26 17:30:41.275 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] GMSHealthMonitor serverSocketExecutor is terminated > [vm_0][info 2016/07/26 17:30:41.278 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] DistributionManager stopped in 108ms. > [vm_0][info 2016/07/26 17:30:41.278 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Marking DistributionManager japan(16120)<ec><v42>:1025 as closed. > [vm_0]distributed system properties: {locators=localhost[33275], > enable-cluster-configuration=false, use-cluster-configuration=false, > mcast-port=0, disable-auto-reconnect=true, log-level=info} > [vm_0][info 2016/07/26 17:30:41.282 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Startup Configuration: > [vm_0] ### GemFire Properties defined with api ### > [vm_0]disable-auto-reconnect=true > [vm_0]enable-cluster-configuration=false > [vm_0]locators=localhost[33275] > [vm_0]log-level=info > [vm_0]mcast-port=0 > [vm_0]use-cluster-configuration=false > [vm_0]### GemFire Properties using default values ### > [vm_0]ack-severe-alert-threshold=0 > [vm_0]ack-wait-threshold=15 > [vm_0]archive-disk-space-limit=0 > [vm_0]archive-file-size-limit=0 > [vm_0]async-distribution-timeout=0 > [vm_0]async-max-queue-size=8 > [vm_0]async-queue-timeout=60000 > [vm_0]bind-address= > [vm_0]cache-xml-file=cache.xml > [vm_0]cluster-configuration-dir= > [vm_0]cluster-ssl-ciphers=any > [vm_0]cluster-ssl-enabled=false > [vm_0]cluster-ssl-keystore= > [vm_0]cluster-ssl-keystore-password= > [vm_0]cluster-ssl-keystore-type= > [vm_0]cluster-ssl-protocols=any > [vm_0]cluster-ssl-require-authentication=true > [vm_0]cluster-ssl-truststore= > [vm_0]cluster-ssl-truststore-password= > [vm_0]conflate-events=server > [vm_0]conserve-sockets=true > [vm_0]delta-propagation=true > [vm_0]deploy-working-dir=. > [vm_0]disable-tcp=false > [vm_0]distributed-system-id=-1 > [vm_0]distributed-transactions=false > [vm_0]durable-client-id= > [vm_0]durable-client-timeout=300 > [vm_0]enable-network-partition-detection=false > [vm_0]enable-time-statistics=false > [vm_0]enforce-unique-host=false > [vm_0]gateway-ssl-ciphers=any > [vm_0]gateway-ssl-enabled=false > [vm_0]gateway-ssl-keystore= > [vm_0]gateway-ssl-keystore-password= > [vm_0]gateway-ssl-keystore-type= > [vm_0]gateway-ssl-protocols=any > [vm_0]gateway-ssl-require-authentication=true > [vm_0]gateway-ssl-truststore= > [vm_0]gateway-ssl-truststore-password= > [vm_0]groups= > [vm_0]http-service-bind-address= > [vm_0]http-service-port=7070 > [vm_0]http-service-ssl-ciphers=any > [vm_0]http-service-ssl-enabled=false > [vm_0]http-service-ssl-keystore= > [vm_0]http-service-ssl-keystore-password= > [vm_0]http-service-ssl-keystore-type= > [vm_0]http-service-ssl-protocols=any > [vm_0]http-service-ssl-require-authentication=false > [vm_0]http-service-ssl-truststore= > [vm_0]http-service-ssl-truststore-password= > [vm_0]jmx-manager=false > [vm_0]jmx-manager-access-file= > [vm_0]jmx-manager-bind-address= > [vm_0]jmx-manager-hostname-for-clients= > [vm_0]jmx-manager-http-port=7070 > [vm_0]jmx-manager-password-file= > [vm_0]jmx-manager-port=1099 > [vm_0]jmx-manager-ssl=false > [vm_0]jmx-manager-ssl-ciphers=any > [vm_0]jmx-manager-ssl-enabled=false > [vm_0]jmx-manager-ssl-keystore= > [vm_0]jmx-manager-ssl-keystore-password= > [vm_0]jmx-manager-ssl-keystore-type= > [vm_0]jmx-manager-ssl-protocols=any > [vm_0]jmx-manager-ssl-require-authentication=true > [vm_0]jmx-manager-ssl-truststore= > [vm_0]jmx-manager-ssl-truststore-password= > [vm_0]jmx-manager-start=false > [vm_0]jmx-manager-update-rate=2000 > [vm_0]load-cluster-configuration-from-dir=false > [vm_0]locator-wait-time=0 > [vm_0]lock-memory=false > [vm_0]log-disk-space-limit=0 > [vm_0]log-file= > [vm_0]log-file-size-limit=0 > [vm_0]max-num-reconnect-tries=3 > [vm_0]max-wait-time-reconnect=60000 > [vm_0]mcast-address=239.192.81.1 > [vm_0]mcast-flow-control=1048576, 0.25, 5000 > [vm_0]mcast-recv-buffer-size=1048576 > [vm_0]mcast-send-buffer-size=65535 > [vm_0]mcast-ttl=32 > [vm_0]member-timeout=5000 > [vm_0]membership-port-range=1024-65535 > [vm_0]memcached-bind-address= > [vm_0]memcached-port=0 > [vm_0]memcached-protocol=ASCII > [vm_0]name= > [vm_0]off-heap-memory-size= > [vm_0]redis-bind-address= > [vm_0]redis-password= > [vm_0]redis-port=0 > [vm_0]redundancy-zone= > [vm_0]remote-locators= > [vm_0]remove-unresponsive-client=false > [vm_0]roles= > [vm_0]security-client-accessor= > [vm_0]security-client-accessor-pp= > [vm_0]security-client-auth-init= > [vm_0]security-client-authenticator= > [vm_0]security-client-dhalgo= > [vm_0]security-log-file= > [vm_0]security-log-level=config > [vm_0]security-manager= > [vm_0]security-peer-auth-init= > [vm_0]security-peer-authenticator= > [vm_0]security-peer-verifymember-timeout=1000 > [vm_0]security-post-processor= > [vm_0]server-bind-address= > [vm_0]server-ssl-ciphers=any > [vm_0]server-ssl-enabled=false > [vm_0]server-ssl-keystore= > [vm_0]server-ssl-keystore-password= > [vm_0]server-ssl-keystore-type= > [vm_0]server-ssl-protocols=any > [vm_0]server-ssl-require-authentication=true > [vm_0]server-ssl-truststore= > [vm_0]server-ssl-truststore-password= > [vm_0]socket-buffer-size=32768 > [vm_0]socket-lease-time=60000 > [vm_0]ssl-ciphers=any > [vm_0]ssl-enabled=false > [vm_0]ssl-protocols=any > [vm_0]ssl-require-authentication=true > [vm_0]start-dev-rest-api=false > [vm_0]start-locator= > [vm_0]statistic-archive-file= > [vm_0]statistic-sample-rate=1000 > [vm_0]statistic-sampling-enabled=true > [vm_0]tcp-port=0 > [vm_0]udp-fragment-size=60000 > [vm_0]udp-recv-buffer-size=1048576 > [vm_0]udp-send-buffer-size=65535 > [vm_0]user-command-packages= > [vm_0][info 2016/07/26 17:30:41.284 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Starting membership services > [vm_0][info 2016/07/26 17:30:41.293 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] JGroups channel created (took 8ms) > [vm_0][info 2016/07/26 17:30:41.294 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] GemFire P2P Listener started on null > [vm_0][info 2016/07/26 17:30:41.295 PDT <Geode Failure Detection Server > thread 0> tid=0x96] Started failure detection server thread on > /10.118.32.37:5113. > [vm_0][info 2016/07/26 17:30:41.296 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Attempting to join the distributed system through coordinator > japan(16059:locator)<ec><v0>:1024 using address japan(16120)<ec>:1025 > [locator][info 2016/07/26 17:30:41.296 PDT <unicast receiver,japan-50141> > tid=0x21] received join request from japan(16120)<ec>:1025 > [locator][info 2016/07/26 17:30:41.497 PDT <Geode Membership View Creator> > tid=0x27] View Creator is processing 3 requests for the next membership view > [locator][info 2016/07/26 17:30:41.497 PDT <Geode Membership View Creator> > tid=0x27] preparing new view View[japan(16059:locator)<ec><v0>:1024|43] > members: [japan(16059:locator)<ec><v0>:1024, > japan(16120)<ec><v43>:1025{lead}] shutdown: [japan(16120)<ec><v42>:1025] > [locator]failure detection ports: 17959 5113 > [vm_0][info 2016/07/26 17:30:41.498 PDT <unicast receiver,japan-59771> > tid=0x92] received new view: View[japan(16059:locator)<ec><v0>:1024|43] > members: [japan(16059:locator)<ec><v0>:1024, > japan(16120)<ec><v43>:1025{lead}] shutdown: [japan(16120)<ec><v42>:1025] > [vm_0]old view is: null > [locator][info 2016/07/26 17:30:41.498 PDT <Geode Membership View Creator> > tid=0x27] finished waiting for responses to view preparation > [locator][info 2016/07/26 17:30:41.498 PDT <Geode Membership View Creator> > tid=0x27] received new view: View[japan(16059:locator)<ec><v0>:1024|43] > members: [japan(16059:locator)<ec><v0>:1024, > japan(16120)<ec><v43>:1025{lead}] shutdown: [japan(16120)<ec><v42>:1025] > [locator]old view is: View[japan(16059:locator)<ec><v0>:1024|42] members: > [japan(16059:locator)<ec><v0>:1024, japan(16120)<ec><v42>:1025{lead}] > shutdown: [japan(16130)<ec><v39>:1025, japan(16124)<ec><v41>:1026] > [vm_0][info 2016/07/26 17:30:41.499 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Finished joining (took 204ms). > [locator][info 2016/07/26 17:30:41.498 PDT <Geode Membership View Creator> > tid=0x27] Peer locator received new membership view: > View[japan(16059:locator)<ec><v0>:1024|43] members: > [japan(16059:locator)<ec><v0>:1024, japan(16120)<ec><v43>:1025{lead}] > shutdown: [japan(16120)<ec><v42>:1025] > [locator][info 2016/07/26 17:30:41.499 PDT <Geode Membership View Creator> > tid=0x27] sending new view View[japan(16059:locator)<ec><v0>:1024|43] > members: [japan(16059:locator)<ec><v0>:1024, > japan(16120)<ec><v43>:1025{lead}] shutdown: [japan(16120)<ec><v42>:1025] > [locator]failure detection ports: 17959 5113 > [locator][info 2016/07/26 17:30:41.499 PDT <View Message Processor> tid=0x33] > Membership: Processing addition < japan(16120)<ec><v43>:1025 > > [vm_0][info 2016/07/26 17:30:41.499 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Starting DistributionManager japan(16120)<ec><v43>:1025. (took 215 > ms) > [locator][info 2016/07/26 17:30:41.499 PDT <View Message Processor> tid=0x33] > Admitting member <japan(16120)<ec><v43>:1025>. Now there are 2 non-admin > member(s). > [vm_0][info 2016/07/26 17:30:41.499 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Initial (distribution manager) view = > View[japan(16059:locator)<ec><v0>:1024|43] members: > [japan(16059:locator)<ec><v0>:1024, japan(16120)<ec><v43>:1025{lead}] > shutdown: [japan(16120)<ec><v42>:1025] > [vm_0][info 2016/07/26 17:30:41.499 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Admitting member <japan(16059:locator)<ec><v0>:1024>. Now there are > 1 non-admin member(s). > [vm_0][info 2016/07/26 17:30:41.500 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Admitting member <japan(16120)<ec><v43>:1025>. Now there are 2 > non-admin member(s). > [locator][info 2016/07/26 17:30:41.502 PDT <Pooled High Priority Message > Processor 1> tid=0x37] Member japan(16120)<ec><v43>:1025 is equivalent or in > the same redundancy zone. > [vm_0][info 2016/07/26 17:30:41.503 PDT <P2P message reader for > japan(16059:locator)<ec><v0>:1024 shared unordered uid=38 port=43393> > tid=0x9b] Member japan(16059:locator)<ec><v0>:1024 is equivalent or in the > same redundancy zone. > [vm_0][info 2016/07/26 17:30:41.512 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Initializing region _monitoringRegion_10.118.32.37<v43>1025 > [vm_0][info 2016/07/26 17:30:41.513 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Initialization of region _monitoringRegion_10.118.32.37<v43>1025 > completed > [vm_0][info 2016/07/26 17:30:41.517 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] The cache has been created with "use-cluster-configuration=false". > It will not receive any cluster configuration > [vm_0][info 2016/07/26 17:30:41.518 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Initializing region PdxTypes > [vm_0][info 2016/07/26 17:30:41.519 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Initialization of region PdxTypes completed > [vm_0][info 2016/07/26 17:30:41.520 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Got result: null > [vm_0] from > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningDUnitTest$$Lambda$12/1305889715.run > with 0 args on object: > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningDUnitTest$$Lambda$12/1305889715@76d2217e > (took 387 ms) > [vm_0][info 2016/07/26 17:30:41.526 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Received method: > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningDUnitTest$$Lambda$13/1734475610.run > with 0 args on object: > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningDUnitTest$$Lambda$13/1734475610@55adb50f > [vm_0][info 2016/07/26 17:30:41.527 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Initializing region __PR > [vm_0][info 2016/07/26 17:30:41.528 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Initialization of region __PR completed > [vm_0][info 2016/07/26 17:30:41.537 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Partitioned Region /Quarter is born with prId=8 ident:#Quarter > [vm_0][info 2016/07/26 17:30:41.543 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Got result: null > [vm_0] from > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningDUnitTest$$Lambda$13/1734475610.run > with 0 args on object: > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningDUnitTest$$Lambda$13/1734475610@55adb50f > (took 16 ms) > [vm_1][info 2016/07/26 17:30:41.561 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Received method: > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningDUnitTest$$Lambda$19/133104116.run > with 0 args on object: > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningDUnitTest$$Lambda$19/133104116@37e534d3 > [vm_1]distributed system properties: {locators=localhost[33275], > enable-cluster-configuration=false, use-cluster-configuration=false, > mcast-port=0, disable-auto-reconnect=true, log-level=info} > [vm_1][info 2016/07/26 17:30:41.563 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Startup Configuration: > [vm_1] ### GemFire Properties defined with api ### > [vm_1]disable-auto-reconnect=true > [vm_1]enable-cluster-configuration=false > [vm_1]locators=localhost[33275] > [vm_1]log-level=info > [vm_1]mcast-port=0 > [vm_1]use-cluster-configuration=false > [vm_1]### GemFire Properties using default values ### > [vm_1]ack-severe-alert-threshold=0 > [vm_1]ack-wait-threshold=15 > [vm_1]archive-disk-space-limit=0 > [vm_1]archive-file-size-limit=0 > [vm_1]async-distribution-timeout=0 > [vm_1]async-max-queue-size=8 > [vm_1]async-queue-timeout=60000 > [vm_1]bind-address= > [vm_1]cache-xml-file=cache.xml > [vm_1]cluster-configuration-dir= > [vm_1]cluster-ssl-ciphers=any > [vm_1]cluster-ssl-enabled=false > [vm_1]cluster-ssl-keystore= > [vm_1]cluster-ssl-keystore-password= > [vm_1]cluster-ssl-keystore-type= > [vm_1]cluster-ssl-protocols=any > [vm_1]cluster-ssl-require-authentication=true > [vm_1]cluster-ssl-truststore= > [vm_1]cluster-ssl-truststore-password= > [vm_1]conflate-events=server > [vm_1]conserve-sockets=true > [vm_1]delta-propagation=true > [vm_1]deploy-working-dir=. > [vm_1]disable-tcp=false > [vm_1]distributed-system-id=-1 > [vm_1]distributed-transactions=false > [vm_1]durable-client-id= > [vm_1]durable-client-timeout=300 > [vm_1]enable-network-partition-detection=false > [vm_1]enable-time-statistics=false > [vm_1]enforce-unique-host=false > [vm_1]gateway-ssl-ciphers=any > [vm_1]gateway-ssl-enabled=false > [vm_1]gateway-ssl-keystore= > [vm_1]gateway-ssl-keystore-password= > [vm_1]gateway-ssl-keystore-type= > [vm_1]gateway-ssl-protocols=any > [vm_1]gateway-ssl-require-authentication=true > [vm_1]gateway-ssl-truststore= > [vm_1]gateway-ssl-truststore-password= > [vm_1]groups= > [vm_1]http-service-bind-address= > [vm_1]http-service-port=7070 > [vm_1]http-service-ssl-ciphers=any > [vm_1]http-service-ssl-enabled=false > [vm_1]http-service-ssl-keystore= > [vm_1]http-service-ssl-keystore-password= > [vm_1]http-service-ssl-keystore-type= > [vm_1]http-service-ssl-protocols=any > [vm_1]http-service-ssl-require-authentication=false > [vm_1]http-service-ssl-truststore= > [vm_1]http-service-ssl-truststore-password= > [vm_1]jmx-manager=false > [vm_1]jmx-manager-access-file= > [vm_1]jmx-manager-bind-address= > [vm_1]jmx-manager-hostname-for-clients= > [vm_1]jmx-manager-http-port=7070 > [vm_1]jmx-manager-password-file= > [vm_1]jmx-manager-port=1099 > [vm_1]jmx-manager-ssl=false > [vm_1]jmx-manager-ssl-ciphers=any > [vm_1]jmx-manager-ssl-enabled=false > [vm_1]jmx-manager-ssl-keystore= > [vm_1]jmx-manager-ssl-keystore-password= > [vm_1]jmx-manager-ssl-keystore-type= > [vm_1]jmx-manager-ssl-protocols=any > [vm_1]jmx-manager-ssl-require-authentication=true > [vm_1]jmx-manager-ssl-truststore= > [vm_1]jmx-manager-ssl-truststore-password= > [vm_1]jmx-manager-start=false > [vm_1]jmx-manager-update-rate=2000 > [vm_1]load-cluster-configuration-from-dir=false > [vm_1]locator-wait-time=0 > [vm_1]lock-memory=false > [vm_1]log-disk-space-limit=0 > [vm_1]log-file= > [vm_1]log-file-size-limit=0 > [vm_1]max-num-reconnect-tries=3 > [vm_1]max-wait-time-reconnect=60000 > [vm_1]mcast-address=239.192.81.1 > [vm_1]mcast-flow-control=1048576, 0.25, 5000 > [vm_1]mcast-recv-buffer-size=1048576 > [vm_1]mcast-send-buffer-size=65535 > [vm_1]mcast-ttl=32 > [vm_1]member-timeout=5000 > [vm_1]membership-port-range=1024-65535 > [vm_1]memcached-bind-address= > [vm_1]memcached-port=0 > [vm_1]memcached-protocol=ASCII > [vm_1]name= > [vm_1]off-heap-memory-size= > [vm_1]redis-bind-address= > [vm_1]redis-password= > [vm_1]redis-port=0 > [vm_1]redundancy-zone= > [vm_1]remote-locators= > [vm_1]remove-unresponsive-client=false > [vm_1]roles= > [vm_1]security-client-accessor= > [vm_1]security-client-accessor-pp= > [vm_1]security-client-auth-init= > [vm_1]security-client-authenticator= > [vm_1]security-client-dhalgo= > [vm_1]security-log-file= > [vm_1]security-log-level=config > [vm_1]security-manager= > [vm_1]security-peer-auth-init= > [vm_1]security-peer-authenticator= > [vm_1]security-peer-verifymember-timeout=1000 > [vm_1]security-post-processor= > [vm_1]server-bind-address= > [vm_1]server-ssl-ciphers=any > [vm_1]server-ssl-enabled=false > [vm_1]server-ssl-keystore= > [vm_1]server-ssl-keystore-password= > [vm_1]server-ssl-keystore-type= > [vm_1]server-ssl-protocols=any > [vm_1]server-ssl-require-authentication=true > [vm_1]server-ssl-truststore= > [vm_1]server-ssl-truststore-password= > [vm_1]socket-buffer-size=32768 > [vm_1]socket-lease-time=60000 > [vm_1]ssl-ciphers=any > [vm_1]ssl-enabled=false > [vm_1]ssl-protocols=any > [vm_1]ssl-require-authentication=true > [vm_1]start-dev-rest-api=false > [vm_1]start-locator= > [vm_1]statistic-archive-file= > [vm_1]statistic-sample-rate=1000 > [vm_1]statistic-sampling-enabled=true > [vm_1]tcp-port=0 > [vm_1]udp-fragment-size=60000 > [vm_1]udp-recv-buffer-size=1048576 > [vm_1]udp-send-buffer-size=65535 > [vm_1]user-command-packages= > [vm_1][info 2016/07/26 17:30:41.565 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Starting membership services > [vm_1][info 2016/07/26 17:30:41.574 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] JGroups channel created (took 9ms) > [vm_1][info 2016/07/26 17:30:41.575 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] GemFire P2P Listener started on null > [vm_1][info 2016/07/26 17:30:41.575 PDT <Geode Failure Detection Server > thread 0> tid=0x2a2] Started failure detection server thread on > japan.gemstone.com/10.118.32.37:59695. > [vm_1][info 2016/07/26 17:30:41.576 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Attempting to join the distributed system through coordinator > japan(16059:locator)<ec><v0>:1024 using address japan(16124)<ec>:1026 > [locator][info 2016/07/26 17:30:41.579 PDT <unicast receiver,japan-50141> > tid=0x21] received join request from japan(16124)<ec>:1026 > [locator][info 2016/07/26 17:30:41.879 PDT <Geode Membership View Creator> > tid=0x27] View Creator is processing 1 requests for the next membership view > [locator][info 2016/07/26 17:30:41.880 PDT <Geode Membership View Creator> > tid=0x27] preparing new view View[japan(16059:locator)<ec><v0>:1024|44] > members: [japan(16059:locator)<ec><v0>:1024, > japan(16120)<ec><v43>:1025{lead}, japan(16124)<ec><v44>:1026] > [locator]failure detection ports: 17959 5113 59695 > [vm_1][info 2016/07/26 17:30:41.880 PDT <unicast receiver,japan-13613> > tid=0x29e] received new view: View[japan(16059:locator)<ec><v0>:1024|44] > members: [japan(16059:locator)<ec><v0>:1024, > japan(16120)<ec><v43>:1025{lead}, japan(16124)<ec><v44>:1026] > [vm_1]old view is: null > [locator][info 2016/07/26 17:30:41.881 PDT <Geode Membership View Creator> > tid=0x27] finished waiting for responses to view preparation > [locator][info 2016/07/26 17:30:41.881 PDT <Geode Membership View Creator> > tid=0x27] received new view: View[japan(16059:locator)<ec><v0>:1024|44] > members: [japan(16059:locator)<ec><v0>:1024, > japan(16120)<ec><v43>:1025{lead}, japan(16124)<ec><v44>:1026] > [locator]old view is: View[japan(16059:locator)<ec><v0>:1024|43] members: > [japan(16059:locator)<ec><v0>:1024, japan(16120)<ec><v43>:1025{lead}] > shutdown: [japan(16120)<ec><v42>:1025] > [locator][info 2016/07/26 17:30:41.881 PDT <Geode Membership View Creator> > tid=0x27] Peer locator received new membership view: > View[japan(16059:locator)<ec><v0>:1024|44] members: > [japan(16059:locator)<ec><v0>:1024, japan(16120)<ec><v43>:1025{lead}, > japan(16124)<ec><v44>:1026] > [vm_1][info 2016/07/26 17:30:41.881 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Finished joining (took 306ms). > [vm_1][info 2016/07/26 17:30:41.881 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Starting DistributionManager japan(16124)<ec><v44>:1026. (took 317 > ms) > [locator][info 2016/07/26 17:30:41.881 PDT <Geode Membership View Creator> > tid=0x27] sending new view View[japan(16059:locator)<ec><v0>:1024|44] > members: [japan(16059:locator)<ec><v0>:1024, > japan(16120)<ec><v43>:1025{lead}, japan(16124)<ec><v44>:1026] > [locator]failure detection ports: 17959 5113 59695 > [vm_1][info 2016/07/26 17:30:41.881 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Initial (distribution manager) view = > View[japan(16059:locator)<ec><v0>:1024|44] members: > [japan(16059:locator)<ec><v0>:1024, japan(16120)<ec><v43>:1025{lead}, > japan(16124)<ec><v44>:1026] > [vm_1][info 2016/07/26 17:30:41.881 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Admitting member <japan(16059:locator)<ec><v0>:1024>. Now there are > 1 non-admin member(s). > [vm_1][info 2016/07/26 17:30:41.881 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Admitting member <japan(16120)<ec><v43>:1025>. Now there are 2 > non-admin member(s). > [vm_1][info 2016/07/26 17:30:41.881 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Admitting member <japan(16124)<ec><v44>:1026>. Now there are 3 > non-admin member(s). > [locator][info 2016/07/26 17:30:41.881 PDT <View Message Processor> tid=0x33] > Membership: Processing addition < japan(16124)<ec><v44>:1026 > > [locator][info 2016/07/26 17:30:41.881 PDT <View Message Processor> tid=0x33] > Admitting member <japan(16124)<ec><v44>:1026>. Now there are 3 non-admin > member(s). > [vm_0][info 2016/07/26 17:30:41.882 PDT <unicast receiver,japan-59771> > tid=0x92] received new view: View[japan(16059:locator)<ec><v0>:1024|44] > members: [japan(16059:locator)<ec><v0>:1024, > japan(16120)<ec><v43>:1025{lead}, japan(16124)<ec><v44>:1026] > [vm_0]old view is: View[japan(16059:locator)<ec><v0>:1024|43] members: > [japan(16059:locator)<ec><v0>:1024, japan(16120)<ec><v43>:1025{lead}] > shutdown: [japan(16120)<ec><v42>:1025] > [vm_0][info 2016/07/26 17:30:41.885 PDT <P2P message reader@645cb237> > tid=0xab] Admitting member <japan(16124)<ec><v44>:1026>. Now there are 3 > non-admin member(s). > [locator][info 2016/07/26 17:30:41.887 PDT <Pooled High Priority Message > Processor 1> tid=0x37] Member japan(16124)<ec><v44>:1026 is equivalent or in > the same redundancy zone. > [vm_0][info 2016/07/26 17:30:41.887 PDT <Pooled High Priority Message > Processor 1> tid=0xa6] Member japan(16124)<ec><v44>:1026 is equivalent or in > the same redundancy zone. > [vm_1][info 2016/07/26 17:30:41.889 PDT <P2P message reader for > japan(16059:locator)<ec><v0>:1024 shared unordered uid=39 port=53241> > tid=0x2a9] Member japan(16059:locator)<ec><v0>:1024 is equivalent or in the > same redundancy zone. > [vm_1][info 2016/07/26 17:30:41.890 PDT <P2P message reader for > japan(16120)<ec><v43>:1025 shared unordered uid=8 port=53242> tid=0x2aa] > Member japan(16120)<ec><v43>:1025 is equivalent or in the same redundancy > zone. > [vm_1][info 2016/07/26 17:30:41.897 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Shutting down DistributionManager japan(16124)<ec><v44>:1026. > [vm_0][info 2016/07/26 17:30:41.899 PDT <Pooled High Priority Message > Processor 1> tid=0xa6] received leave request from japan(16124)<ec><v44>:1026 > for japan(16124)<ec><v44>:1026 > [vm_0][info 2016/07/26 17:30:41.899 PDT <Pooled High Priority Message > Processor 1> tid=0xa6] Member at japan(16124)<ec><v44>:1026 gracefully left > the distributed cache: shutdown message received > [locator][info 2016/07/26 17:30:41.899 PDT <Pooled High Priority Message > Processor 1> tid=0x37] received leave request from japan(16124)<ec><v44>:1026 > for japan(16124)<ec><v44>:1026 > [locator][info 2016/07/26 17:30:41.899 PDT <Pooled High Priority Message > Processor 1> tid=0x37] Member at japan(16124)<ec><v44>:1026 gracefully left > the distributed cache: shutdown message received > [vm_1][info 2016/07/26 17:30:41.999 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Now closing distribution for japan(16124)<ec><v44>:1026 > [vm_1][info 2016/07/26 17:30:41.999 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Stopping membership services > [vm_1][info 2016/07/26 17:30:42.000 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] GMSHealthMonitor server socket is closed in stopServices(). > [locator][info 2016/07/26 17:30:42.000 PDT <unicast receiver,japan-50141> > tid=0x21] received leave request from japan(16124)<ec><v44>:1026 for > japan(16124)<ec><v44>:1026 > [vm_1][info 2016/07/26 17:30:42.000 PDT <Geode Failure Detection Server > thread 0> tid=0x2a2] GMSHealthMonitor server thread exiting > [vm_1][info 2016/07/26 17:30:42.000 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] GMSHealthMonitor serverSocketExecutor is terminated > [vm_0][info 2016/07/26 17:30:42.000 PDT <unicast receiver,japan-59771> > tid=0x92] received leave request from japan(16124)<ec><v44>:1026 for > japan(16124)<ec><v44>:1026 > [vm_1][info 2016/07/26 17:30:42.002 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] DistributionManager stopped in 105ms. > [vm_1][info 2016/07/26 17:30:42.002 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Marking DistributionManager japan(16124)<ec><v44>:1026 as closed. > [vm_1]distributed system properties: {locators=localhost[33275], > enable-cluster-configuration=false, use-cluster-configuration=false, > mcast-port=0, disable-auto-reconnect=true, log-level=info} > [vm_1][info 2016/07/26 17:30:42.006 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Startup Configuration: > [vm_1] ### GemFire Properties defined with api ### > [vm_1]disable-auto-reconnect=true > [vm_1]enable-cluster-configuration=false > [vm_1]locators=localhost[33275] > [vm_1]log-level=info > [vm_1]mcast-port=0 > [vm_1]use-cluster-configuration=false > [vm_1]### GemFire Properties using default values ### > [vm_1]ack-severe-alert-threshold=0 > [vm_1]ack-wait-threshold=15 > [vm_1]archive-disk-space-limit=0 > [vm_1]archive-file-size-limit=0 > [vm_1]async-distribution-timeout=0 > [vm_1]async-max-queue-size=8 > [vm_1]async-queue-timeout=60000 > [vm_1]bind-address= > [vm_1]cache-xml-file=cache.xml > [vm_1]cluster-configuration-dir= > [vm_1]cluster-ssl-ciphers=any > [vm_1]cluster-ssl-enabled=false > [vm_1]cluster-ssl-keystore= > [vm_1]cluster-ssl-keystore-password= > [vm_1]cluster-ssl-keystore-type= > [vm_1]cluster-ssl-protocols=any > [vm_1]cluster-ssl-require-authentication=true > [vm_1]cluster-ssl-truststore= > [vm_1]cluster-ssl-truststore-password= > [vm_1]conflate-events=server > [vm_1]conserve-sockets=true > [vm_1]delta-propagation=true > [vm_1]deploy-working-dir=. > [vm_1]disable-tcp=false > [vm_1]distributed-system-id=-1 > [vm_1]distributed-transactions=false > [vm_1]durable-client-id= > [vm_1]durable-client-timeout=300 > [vm_1]enable-network-partition-detection=false > [vm_1]enable-time-statistics=false > [vm_1]enforce-unique-host=false > [vm_1]gateway-ssl-ciphers=any > [vm_1]gateway-ssl-enabled=false > [vm_1]gateway-ssl-keystore= > [vm_1]gateway-ssl-keystore-password= > [vm_1]gateway-ssl-keystore-type= > [vm_1]gateway-ssl-protocols=any > [vm_1]gateway-ssl-require-authentication=true > [vm_1]gateway-ssl-truststore= > [vm_1]gateway-ssl-truststore-password= > [vm_1]groups= > [vm_1]http-service-bind-address= > [vm_1]http-service-port=7070 > [vm_1]http-service-ssl-ciphers=any > [vm_1]http-service-ssl-enabled=false > [vm_1]http-service-ssl-keystore= > [vm_1]http-service-ssl-keystore-password= > [vm_1]http-service-ssl-keystore-type= > [vm_1]http-service-ssl-protocols=any > [vm_1]http-service-ssl-require-authentication=false > [vm_1]http-service-ssl-truststore= > [vm_1]http-service-ssl-truststore-password= > [vm_1]jmx-manager=false > [vm_1]jmx-manager-access-file= > [vm_1]jmx-manager-bind-address= > [vm_1]jmx-manager-hostname-for-clients= > [vm_1]jmx-manager-http-port=7070 > [vm_1]jmx-manager-password-file= > [vm_1]jmx-manager-port=1099 > [vm_1]jmx-manager-ssl=false > [vm_1]jmx-manager-ssl-ciphers=any > [vm_1]jmx-manager-ssl-enabled=false > [vm_1]jmx-manager-ssl-keystore= > [vm_1]jmx-manager-ssl-keystore-password= > [vm_1]jmx-manager-ssl-keystore-type= > [vm_1]jmx-manager-ssl-protocols=any > [vm_1]jmx-manager-ssl-require-authentication=true > [vm_1]jmx-manager-ssl-truststore= > [vm_1]jmx-manager-ssl-truststore-password= > [vm_1]jmx-manager-start=false > [vm_1]jmx-manager-update-rate=2000 > [vm_1]load-cluster-configuration-from-dir=false > [vm_1]locator-wait-time=0 > [vm_1]lock-memory=false > [vm_1]log-disk-space-limit=0 > [vm_1]log-file= > [vm_1]log-file-size-limit=0 > [vm_1]max-num-reconnect-tries=3 > [vm_1]max-wait-time-reconnect=60000 > [vm_1]mcast-address=239.192.81.1 > [vm_1]mcast-flow-control=1048576, 0.25, 5000 > [vm_1]mcast-recv-buffer-size=1048576 > [vm_1]mcast-send-buffer-size=65535 > [vm_1]mcast-ttl=32 > [vm_1]member-timeout=5000 > [vm_1]membership-port-range=1024-65535 > [vm_1]memcached-bind-address= > [vm_1]memcached-port=0 > [vm_1]memcached-protocol=ASCII > [vm_1]name= > [vm_1]off-heap-memory-size= > [vm_1]redis-bind-address= > [vm_1]redis-password= > [vm_1]redis-port=0 > [vm_1]redundancy-zone= > [vm_1]remote-locators= > [vm_1]remove-unresponsive-client=false > [vm_1]roles= > [vm_1]security-client-accessor= > [vm_1]security-client-accessor-pp= > [vm_1]security-client-auth-init= > [vm_1]security-client-authenticator= > [vm_1]security-client-dhalgo= > [vm_1]security-log-file= > [vm_1]security-log-level=config > [vm_1]security-manager= > [vm_1]security-peer-auth-init= > [vm_1]security-peer-authenticator= > [vm_1]security-peer-verifymember-timeout=1000 > [vm_1]security-post-processor= > [vm_1]server-bind-address= > [vm_1]server-ssl-ciphers=any > [vm_1]server-ssl-enabled=false > [vm_1]server-ssl-keystore= > [vm_1]server-ssl-keystore-password= > [vm_1]server-ssl-keystore-type= > [vm_1]server-ssl-protocols=any > [vm_1]server-ssl-require-authentication=true > [vm_1]server-ssl-truststore= > [vm_1]server-ssl-truststore-password= > [vm_1]socket-buffer-size=32768 > [vm_1]socket-lease-time=60000 > [vm_1]ssl-ciphers=any > [vm_1]ssl-enabled=false > [vm_1]ssl-protocols=any > [vm_1]ssl-require-authentication=true > [vm_1]start-dev-rest-api=false > [vm_1]start-locator= > [vm_1]statistic-archive-file= > [vm_1]statistic-sample-rate=1000 > [vm_1]statistic-sampling-enabled=true > [vm_1]tcp-port=0 > [vm_1]udp-fragment-size=60000 > [vm_1]udp-recv-buffer-size=1048576 > [vm_1]udp-send-buffer-size=65535 > [vm_1]user-command-packages= > [vm_1][info 2016/07/26 17:30:42.007 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Starting membership services > [vm_1][info 2016/07/26 17:30:42.016 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] JGroups channel created (took 9ms) > [vm_1][info 2016/07/26 17:30:42.017 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] GemFire P2P Listener started on null > [vm_1][info 2016/07/26 17:30:42.017 PDT <Geode Failure Detection Server > thread 0> tid=0x2c3] Started failure detection server thread on > /10.118.32.37:15110. > [vm_1][info 2016/07/26 17:30:42.019 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Attempting to join the distributed system through coordinator > japan(16059:locator)<ec><v0>:1024 using address japan(16124)<ec>:1026 > [locator][info 2016/07/26 17:30:42.021 PDT <unicast receiver,japan-50141> > tid=0x21] received join request from japan(16124)<ec>:1026 > [locator][info 2016/07/26 17:30:42.221 PDT <Geode Membership View Creator> > tid=0x27] View Creator is processing 3 requests for the next membership view > [locator][info 2016/07/26 17:30:42.221 PDT <Geode Membership View Creator> > tid=0x27] preparing new view View[japan(16059:locator)<ec><v0>:1024|45] > members: [japan(16059:locator)<ec><v0>:1024, > japan(16120)<ec><v43>:1025{lead}, japan(16124)<ec><v45>:1026] shutdown: > [japan(16124)<ec><v44>:1026] > [locator]failure detection ports: 17959 5113 15110 > [vm_1][info 2016/07/26 17:30:42.222 PDT <unicast receiver,japan-19913> > tid=0x2bf] received new view: View[japan(16059:locator)<ec><v0>:1024|45] > members: [japan(16059:locator)<ec><v0>:1024, > japan(16120)<ec><v43>:1025{lead}, japan(16124)<ec><v45>:1026] shutdown: > [japan(16124)<ec><v44>:1026] > [vm_1]old view is: null > [locator][info 2016/07/26 17:30:42.222 PDT <Geode Membership View Creator> > tid=0x27] finished waiting for responses to view preparation > [vm_1][info 2016/07/26 17:30:42.223 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Finished joining (took 206ms). > [vm_1][info 2016/07/26 17:30:42.223 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Starting DistributionManager japan(16124)<ec><v45>:1026. (took 216 > ms) > [locator][info 2016/07/26 17:30:42.223 PDT <Geode Membership View Creator> > tid=0x27] received new view: View[japan(16059:locator)<ec><v0>:1024|45] > members: [japan(16059:locator)<ec><v0>:1024, > japan(16120)<ec><v43>:1025{lead}, japan(16124)<ec><v45>:1026] shutdown: > [japan(16124)<ec><v44>:1026] > [locator]old view is: View[japan(16059:locator)<ec><v0>:1024|44] members: > [japan(16059:locator)<ec><v0>:1024, japan(16120)<ec><v43>:1025{lead}, > japan(16124)<ec><v44>:1026] > [locator][info 2016/07/26 17:30:42.223 PDT <Geode Membership View Creator> > tid=0x27] Peer locator received new membership view: > View[japan(16059:locator)<ec><v0>:1024|45] members: > [japan(16059:locator)<ec><v0>:1024, japan(16120)<ec><v43>:1025{lead}, > japan(16124)<ec><v45>:1026] shutdown: [japan(16124)<ec><v44>:1026] > [locator][info 2016/07/26 17:30:42.223 PDT <Geode Membership View Creator> > tid=0x27] sending new view View[japan(16059:locator)<ec><v0>:1024|45] > members: [japan(16059:locator)<ec><v0>:1024, > japan(16120)<ec><v43>:1025{lead}, japan(16124)<ec><v45>:1026] shutdown: > [japan(16124)<ec><v44>:1026] > [locator]failure detection ports: 17959 5113 15110 > [vm_1][info 2016/07/26 17:30:42.223 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Initial (distribution manager) view = > View[japan(16059:locator)<ec><v0>:1024|45] members: > [japan(16059:locator)<ec><v0>:1024, japan(16120)<ec><v43>:1025{lead}, > japan(16124)<ec><v45>:1026] shutdown: [japan(16124)<ec><v44>:1026] > [vm_1][info 2016/07/26 17:30:42.223 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Admitting member <japan(16059:locator)<ec><v0>:1024>. Now there are > 1 non-admin member(s). > [vm_1][info 2016/07/26 17:30:42.223 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Admitting member <japan(16120)<ec><v43>:1025>. Now there are 2 > non-admin member(s). > [vm_1][info 2016/07/26 17:30:42.223 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Admitting member <japan(16124)<ec><v45>:1026>. Now there are 3 > non-admin member(s). > [locator][info 2016/07/26 17:30:42.223 PDT <View Message Processor> tid=0x33] > Membership: Processing addition < japan(16124)<ec><v45>:1026 > > [locator][info 2016/07/26 17:30:42.223 PDT <View Message Processor> tid=0x33] > Admitting member <japan(16124)<ec><v45>:1026>. Now there are 3 non-admin > member(s). > [vm_0][info 2016/07/26 17:30:42.225 PDT <unicast receiver,japan-59771> > tid=0x92] received new view: View[japan(16059:locator)<ec><v0>:1024|45] > members: [japan(16059:locator)<ec><v0>:1024, > japan(16120)<ec><v43>:1025{lead}, japan(16124)<ec><v45>:1026] shutdown: > [japan(16124)<ec><v44>:1026] > [vm_0]old view is: View[japan(16059:locator)<ec><v0>:1024|44] members: > [japan(16059:locator)<ec><v0>:1024, japan(16120)<ec><v43>:1025{lead}, > japan(16124)<ec><v44>:1026] > [vm_0][info 2016/07/26 17:30:42.226 PDT <P2P message reader@4b0e3139> > tid=0xb0] Admitting member <japan(16124)<ec><v45>:1026>. Now there are 3 > non-admin member(s). > [vm_0][info 2016/07/26 17:30:42.228 PDT <Pooled High Priority Message > Processor 1> tid=0xa6] Member japan(16124)<ec><v45>:1026 is equivalent or in > the same redundancy zone. > [locator][info 2016/07/26 17:30:42.230 PDT <Pooled High Priority Message > Processor 1> tid=0x37] Member japan(16124)<ec><v45>:1026 is equivalent or in > the same redundancy zone. > [vm_1][info 2016/07/26 17:30:42.232 PDT <P2P message reader for > japan(16120)<ec><v43>:1025 shared unordered uid=9 port=49962> tid=0x2ca] > Member japan(16120)<ec><v43>:1025 is equivalent or in the same redundancy > zone. > [vm_1][info 2016/07/26 17:30:42.233 PDT <P2P message reader for > japan(16059:locator)<ec><v0>:1024 shared unordered uid=40 port=49963> > tid=0x2cb] Member japan(16059:locator)<ec><v0>:1024 is equivalent or in the > same redundancy zone. > [vm_1][info 2016/07/26 17:30:42.241 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Initializing region _monitoringRegion_10.118.32.37<v45>1026 > [vm_1][info 2016/07/26 17:30:42.242 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Initialization of region _monitoringRegion_10.118.32.37<v45>1026 > completed > [vm_1][info 2016/07/26 17:30:42.246 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] The cache has been created with "use-cluster-configuration=false". > It will not receive any cluster configuration > [vm_1][info 2016/07/26 17:30:42.247 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Initializing region PdxTypes > [vm_1][info 2016/07/26 17:30:42.250 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Region PdxTypes requesting initial image from > japan(16120)<ec><v43>:1025 > [vm_1][info 2016/07/26 17:30:42.254 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] PdxTypes is done getting image from japan(16120)<ec><v43>:1025. > isDeltaGII is false > [vm_1][info 2016/07/26 17:30:42.254 PDT <RMI TCP Connection(1)-10.118.32.37> > tid=0x12] Initialization of region PdxTypes completed > [vm_1][info 2016/07/26 17:30:42.255 PDT <RMI > ...[truncated 3646857 chars]... > alMaxMemory=2147483647;totalNumBuckets=12;partitionResolver=com.gemstone.gemfire.internal.cache.partitioned.fixed.QuarterPartitionResolver@492b0293;colocatedWith=null;recoveryDelay=-1;startupRecoveryDelay=0;FixedPartitionAttributes=[FixedPartitionAttributes@[partitionName=Q3;isPrimary=true;numBuckets=3], > FixedPartitionAttributes@[partitionName=Q1;isPrimary=false;numBuckets=3], > FixedPartitionAttributes@[partitionName=Q2;isPrimary=false;numBuckets=3]];partitionListeners=null]; > on VM japan(16130)<ec><v250>:1028] in this VM :- 3 > [vm_2][info 2016/07/26 17:38:16.102 PDT <RMI TCP Connection(4)-10.118.32.37> > tid=0x12] Lit of Primaries in this VM :- [3, 4, 5] > [vm_2][info 2016/07/26 17:38:16.102 PDT <RMI TCP Connection(4)-10.118.32.37> > tid=0x12] Got result: null > [vm_2] from > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningDUnitTest$$Lambda$89/1448967926.run > with 0 args on object: > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningDUnitTest$$Lambda$89/1448967926@4a83ee5d > (took 0 ms) > [vm_3][info 2016/07/26 17:38:16.104 PDT <RMI TCP Connection(6)-10.118.32.37> > tid=0x12] Received method: > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningDUnitTest$$Lambda$76/1867127127.run > with 0 args on object: > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningDUnitTest$$Lambda$76/1867127127@4814ca10 > [vm_3][info 2016/07/26 17:38:16.104 PDT <RMI TCP Connection(6)-10.118.32.37> > tid=0x12] Size of the Partitioned Region @f3056ce [path='/Quarter'; > dataPolicy=PARTITION; prId=30; isDestroyed=false; isClosed=false; > retryTimeout=3600000; serialNumber=614; partition > attributes=PartitionAttributes@1883691941[redundantCopies=3;localMaxMemory=40;totalMaxMemory=2147483647;totalNumBuckets=12;partitionResolver=com.gemstone.gemfire.internal.cache.partitioned.fixed.QuarterPartitionResolver@74794175;colocatedWith=null;recoveryDelay=-1;startupRecoveryDelay=0;FixedPartitionAttributes=[FixedPartitionAttributes@[partitionName=Q4;isPrimary=true;numBuckets=3], > FixedPartitionAttributes@[partitionName=Q1;isPrimary=false;numBuckets=3], > FixedPartitionAttributes@[partitionName=Q2;isPrimary=false;numBuckets=3]];partitionListeners=null]; > on VM japan(16139)<ec><v252>:1029] in this VM :- 9List of buckets : [0, 1, > 2, 6, 7, 8, 9, 10, 11] > [vm_3][info 2016/07/26 17:38:16.104 PDT <RMI TCP Connection(6)-10.118.32.37> > tid=0x12] Size of primary buckets the Partitioned Region @f3056ce > [path='/Quarter'; dataPolicy=PARTITION; prId=30; isDestroyed=false; > isClosed=false; retryTimeout=3600000; serialNumber=614; partition > attributes=PartitionAttributes@1883691941[redundantCopies=3;localMaxMemory=40;totalMaxMemory=2147483647;totalNumBuckets=12;partitionResolver=com.gemstone.gemfire.internal.cache.partitioned.fixed.QuarterPartitionResolver@74794175;colocatedWith=null;recoveryDelay=-1;startupRecoveryDelay=0;FixedPartitionAttributes=[FixedPartitionAttributes@[partitionName=Q4;isPrimary=true;numBuckets=3], > FixedPartitionAttributes@[partitionName=Q1;isPrimary=false;numBuckets=3], > FixedPartitionAttributes@[partitionName=Q2;isPrimary=false;numBuckets=3]];partitionListeners=null]; > on VM japan(16139)<ec><v252>:1029] in this VM :- 3 > [vm_3][info 2016/07/26 17:38:16.104 PDT <RMI TCP Connection(6)-10.118.32.37> > tid=0x12] Lit of Primaries in this VM :- [6, 7, 8] > [vm_3][info 2016/07/26 17:38:16.104 PDT <RMI TCP Connection(6)-10.118.32.37> > tid=0x12] Got result: null > [vm_3] from > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningDUnitTest$$Lambda$76/1867127127.run > with 0 args on object: > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningDUnitTest$$Lambda$76/1867127127@4814ca10 > (took 0 ms) > [vm_3][info 2016/07/26 17:38:16.106 PDT <RMI TCP Connection(6)-10.118.32.37> > tid=0x12] Received method: > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningDUnitTest$$Lambda$77/1386422795.run > with 0 args on object: > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningDUnitTest$$Lambda$77/1386422795@6d9c09a9 > [vm_3][info 2016/07/26 17:38:16.107 PDT <RMI TCP Connection(6)-10.118.32.37> > tid=0x12] GemFireCache[id = 1260355656; isClosing = true; isShutDownAll = > false; created = Tue Jul 26 17:38:15 PDT 2016; server = false; copyOnRead = > false; lockLease = 120; lockTimeout = 60]: Now closing. > [vm_3][info 2016/07/26 17:38:16.160 PDT <RMI TCP Connection(6)-10.118.32.37> > tid=0x12] Shutting down DistributionManager japan(16139)<ec><v252>:1029. > [locator][info 2016/07/26 17:38:16.165 PDT <Pooled High Priority Message > Processor 1> tid=0x37] received leave request from > japan(16139)<ec><v252>:1029 for japan(16139)<ec><v252>:1029 > [locator][info 2016/07/26 17:38:16.165 PDT <Pooled High Priority Message > Processor 1> tid=0x37] Member at japan(16139)<ec><v252>:1029 gracefully left > the distributed cache: shutdown message received > [info 2016/07/26 17:38:16.165 PDT <Pooled High Priority Message Processor 1> > tid=0x163] received leave request from japan(16139)<ec><v252>:1029 for > japan(16139)<ec><v252>:1029 > [vm_2][info 2016/07/26 17:38:16.165 PDT <Pooled High Priority Message > Processor 2> tid=0xc1b] received leave request from > japan(16139)<ec><v252>:1029 for japan(16139)<ec><v252>:1029 > [vm_2][info 2016/07/26 17:38:16.165 PDT <Pooled High Priority Message > Processor 2> tid=0xc1b] Member at japan(16139)<ec><v252>:1029 gracefully left > the distributed cache: shutdown message received > [info 2016/07/26 17:38:16.165 PDT <Pooled High Priority Message Processor 1> > tid=0x163] Member at japan(16139)<ec><v252>:1029 gracefully left the > distributed cache: shutdown message received > [vm_1][info 2016/07/26 17:38:16.166 PDT <Pooled High Priority Message > Processor 3> tid=0xd17] received leave request from > japan(16139)<ec><v252>:1029 for japan(16139)<ec><v252>:1029 > [vm_1][info 2016/07/26 17:38:16.166 PDT <Pooled High Priority Message > Processor 3> tid=0xd17] Member at japan(16139)<ec><v252>:1029 gracefully left > the distributed cache: shutdown message received > [vm_0][info 2016/07/26 17:38:16.166 PDT <Pooled High Priority Message > Processor 1> tid=0xb30] received leave request from > japan(16139)<ec><v252>:1029 for japan(16139)<ec><v252>:1029 > [vm_0][info 2016/07/26 17:38:16.167 PDT <Pooled High Priority Message > Processor 1> tid=0xb30] Member at japan(16139)<ec><v252>:1029 gracefully left > the distributed cache: shutdown message received > [vm_3][info 2016/07/26 17:38:16.267 PDT <RMI TCP Connection(6)-10.118.32.37> > tid=0x12] Now closing distribution for japan(16139)<ec><v252>:1029 > [vm_3][info 2016/07/26 17:38:16.267 PDT <RMI TCP Connection(6)-10.118.32.37> > tid=0x12] Stopping membership services > [info 2016/07/26 17:38:16.267 PDT <unicast receiver,japan-2663> tid=0x14e] > received leave request from japan(16139)<ec><v252>:1029 for > japan(16139)<ec><v252>:1029 > [vm_0][info 2016/07/26 17:38:16.267 PDT <unicast receiver,japan-16947> > tid=0xb19] received leave request from japan(16139)<ec><v252>:1029 for > japan(16139)<ec><v252>:1029 > [vm_3][info 2016/07/26 17:38:16.267 PDT <RMI TCP Connection(6)-10.118.32.37> > tid=0x12] GMSHealthMonitor server socket is closed in stopServices(). > [vm_3][info 2016/07/26 17:38:16.268 PDT <Geode Failure Detection Server > thread 0> tid=0xa0c] GMSHealthMonitor server thread exiting > [locator][info 2016/07/26 17:38:16.267 PDT <unicast receiver,japan-50141> > tid=0x21] received leave request from japan(16139)<ec><v252>:1029 for > japan(16139)<ec><v252>:1029 > [vm_2][info 2016/07/26 17:38:16.268 PDT <unicast receiver,japan-39240> > tid=0xbfd] received leave request from japan(16139)<ec><v252>:1029 for > japan(16139)<ec><v252>:1029 > [vm_3][info 2016/07/26 17:38:16.268 PDT <RMI TCP Connection(6)-10.118.32.37> > tid=0x12] GMSHealthMonitor serverSocketExecutor is terminated > [vm_1][info 2016/07/26 17:38:16.268 PDT <unicast receiver,japan-620> > tid=0xcfb] received leave request from japan(16139)<ec><v252>:1029 for > japan(16139)<ec><v252>:1029 > [vm_3][info 2016/07/26 17:38:16.272 PDT <RMI TCP Connection(6)-10.118.32.37> > tid=0x12] DistributionManager stopped in 112ms. > [vm_3][info 2016/07/26 17:38:16.273 PDT <RMI TCP Connection(6)-10.118.32.37> > tid=0x12] Marking DistributionManager japan(16139)<ec><v252>:1029 as closed. > [vm_3][info 2016/07/26 17:38:16.273 PDT <RMI TCP Connection(6)-10.118.32.37> > tid=0x12] Got result: null > [vm_3] from > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningDUnitTest$$Lambda$77/1386422795.run > with 0 args on object: > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningDUnitTest$$Lambda$77/1386422795@6d9c09a9 > (took 167 ms) > [info 2016/07/26 17:38:16.274 PDT <Test worker> tid=0xa] Pausing for 1000 > ms... > [locator][info 2016/07/26 17:38:16.468 PDT <Geode Membership View Creator> > tid=0x27] View Creator is processing 2 requests for the next membership view > [locator][info 2016/07/26 17:38:16.468 PDT <Geode Membership View Creator> > tid=0x27] preparing new view View[japan(16059:locator)<ec><v0>:1024|253] > members: [japan(16059:locator)<ec><v0>:1024, > japan(16028)<ec><v244>:1025{lead}, japan(16120)<ec><v246>:1026, > japan(16124)<ec><v248>:1027, japan(16130)<ec><v250>:1028] shutdown: > [japan(16139)<ec><v252>:1029] > [locator]failure detection ports: 17959 32315 20343 14847 47154 > [locator][info 2016/07/26 17:38:16.469 PDT <Geode Membership View Creator> > tid=0x27] finished waiting for responses to view preparation > [locator][info 2016/07/26 17:38:16.469 PDT <Geode Membership View Creator> > tid=0x27] received new view: View[japan(16059:locator)<ec><v0>:1024|253] > members: [japan(16059:locator)<ec><v0>:1024, > japan(16028)<ec><v244>:1025{lead}, japan(16120)<ec><v246>:1026, > japan(16124)<ec><v248>:1027, japan(16130)<ec><v250>:1028] shutdown: > [japan(16139)<ec><v252>:1029] > [locator]old view is: View[japan(16059:locator)<ec><v0>:1024|252] members: > [japan(16059:locator)<ec><v0>:1024, japan(16028)<ec><v244>:1025{lead}, > japan(16120)<ec><v246>:1026, japan(16124)<ec><v248>:1027, > japan(16130)<ec><v250>:1028, japan(16139)<ec><v252>:1029] shutdown: > [japan(16139)<ec><v251>:1029] > [locator][info 2016/07/26 17:38:16.469 PDT <Geode Membership View Creator> > tid=0x27] Peer locator received new membership view: > View[japan(16059:locator)<ec><v0>:1024|253] members: > [japan(16059:locator)<ec><v0>:1024, japan(16028)<ec><v244>:1025{lead}, > japan(16120)<ec><v246>:1026, japan(16124)<ec><v248>:1027, > japan(16130)<ec><v250>:1028] shutdown: [japan(16139)<ec><v252>:1029] > [locator][info 2016/07/26 17:38:16.470 PDT <Geode Membership View Creator> > tid=0x27] sending new view View[japan(16059:locator)<ec><v0>:1024|253] > members: [japan(16059:locator)<ec><v0>:1024, > japan(16028)<ec><v244>:1025{lead}, japan(16120)<ec><v246>:1026, > japan(16124)<ec><v248>:1027, japan(16130)<ec><v250>:1028] shutdown: > [japan(16139)<ec><v252>:1029] > [locator]failure detection ports: 17959 32315 20343 14847 47154 > [vm_2][info 2016/07/26 17:38:16.470 PDT <unicast receiver,japan-39240> > tid=0xbfd] received new view: View[japan(16059:locator)<ec><v0>:1024|253] > members: [japan(16059:locator)<ec><v0>:1024, > japan(16028)<ec><v244>:1025{lead}, japan(16120)<ec><v246>:1026, > japan(16124)<ec><v248>:1027, japan(16130)<ec><v250>:1028] shutdown: > [japan(16139)<ec><v252>:1029] > [vm_2]old view is: View[japan(16059:locator)<ec><v0>:1024|252] members: > [japan(16059:locator)<ec><v0>:1024, japan(16028)<ec><v244>:1025{lead}, > japan(16120)<ec><v246>:1026, japan(16124)<ec><v248>:1027, > japan(16130)<ec><v250>:1028, japan(16139)<ec><v252>:1029] shutdown: > [japan(16139)<ec><v251>:1029] > [vm_0][info 2016/07/26 17:38:16.470 PDT <unicast receiver,japan-16947> > tid=0xb19] received new view: View[japan(16059:locator)<ec><v0>:1024|253] > members: [japan(16059:locator)<ec><v0>:1024, > japan(16028)<ec><v244>:1025{lead}, japan(16120)<ec><v246>:1026, > japan(16124)<ec><v248>:1027, japan(16130)<ec><v250>:1028] shutdown: > [japan(16139)<ec><v252>:1029] > [vm_0]old view is: View[japan(16059:locator)<ec><v0>:1024|252] members: > [japan(16059:locator)<ec><v0>:1024, japan(16028)<ec><v244>:1025{lead}, > japan(16120)<ec><v246>:1026, japan(16124)<ec><v248>:1027, > japan(16130)<ec><v250>:1028, japan(16139)<ec><v252>:1029] shutdown: > [japan(16139)<ec><v251>:1029] > [vm_1][info 2016/07/26 17:38:16.471 PDT <unicast receiver,japan-620> > tid=0xcfb] received new view: View[japan(16059:locator)<ec><v0>:1024|253] > members: [japan(16059:locator)<ec><v0>:1024, > japan(16028)<ec><v244>:1025{lead}, japan(16120)<ec><v246>:1026, > japan(16124)<ec><v248>:1027, japan(16130)<ec><v250>:1028] shutdown: > [japan(16139)<ec><v252>:1029] > [vm_1]old view is: View[japan(16059:locator)<ec><v0>:1024|252] members: > [japan(16059:locator)<ec><v0>:1024, japan(16028)<ec><v244>:1025{lead}, > japan(16120)<ec><v246>:1026, japan(16124)<ec><v248>:1027, > japan(16130)<ec><v250>:1028, japan(16139)<ec><v252>:1029] shutdown: > [japan(16139)<ec><v251>:1029] > [info 2016/07/26 17:38:16.471 PDT <unicast receiver,japan-2663> tid=0x14e] > received new view: View[japan(16059:locator)<ec><v0>:1024|253] members: > [japan(16059:locator)<ec><v0>:1024, japan(16028)<ec><v244>:1025{lead}, > japan(16120)<ec><v246>:1026, japan(16124)<ec><v248>:1027, > japan(16130)<ec><v250>:1028] shutdown: [japan(16139)<ec><v252>:1029] > old view is: View[japan(16059:locator)<ec><v0>:1024|252] members: > [japan(16059:locator)<ec><v0>:1024, japan(16028)<ec><v244>:1025{lead}, > japan(16120)<ec><v246>:1026, japan(16124)<ec><v248>:1027, > japan(16130)<ec><v250>:1028, japan(16139)<ec><v252>:1029] shutdown: > [japan(16139)<ec><v251>:1029] > [vm_0][info 2016/07/26 17:38:17.276 PDT <RMI TCP Connection(4)-10.118.32.37> > tid=0x44e] Received method: > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningDUnitTest$$Lambda$110/888029628.run > with 0 args on object: > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningDUnitTest$$Lambda$110/888029628@395b645 > [vm_0][info 2016/07/26 17:38:17.283 PDT <RMI TCP Connection(4)-10.118.32.37> > tid=0x44e] Got result: null > [vm_0] from > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningDUnitTest$$Lambda$110/888029628.run > with 0 args on object: > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningDUnitTest$$Lambda$110/888029628@395b645 > (took 6 ms) > [vm_1][info 2016/07/26 17:38:17.285 PDT <RMI TCP Connection(5)-10.118.32.37> > tid=0x680] Received method: > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningDUnitTest$$Lambda$95/2077110724.run > with 0 args on object: > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningDUnitTest$$Lambda$95/2077110724@22491221 > [vm_1][info 2016/07/26 17:38:17.293 PDT <RMI TCP Connection(5)-10.118.32.37> > tid=0x680] Got result: null > [vm_1] from > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningDUnitTest$$Lambda$95/2077110724.run > with 0 args on object: > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningDUnitTest$$Lambda$95/2077110724@22491221 > (took 8 ms) > [vm_2][info 2016/07/26 17:38:17.295 PDT <RMI TCP Connection(4)-10.118.32.37> > tid=0x12] Received method: > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningDUnitTest$$Lambda$90/1573184613.run > with 0 args on object: > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningDUnitTest$$Lambda$90/1573184613@60539be5 > [vm_2][info 2016/07/26 17:38:17.311 PDT <RMI TCP Connection(4)-10.118.32.37> > tid=0x12] Got result: null > [vm_2] from > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningDUnitTest$$Lambda$90/1573184613.run > with 0 args on object: > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningDUnitTest$$Lambda$90/1573184613@60539be5 > (took 15 ms) > [vm_0][info 2016/07/26 17:38:17.313 PDT <RMI TCP Connection(4)-10.118.32.37> > tid=0x44e] Received method: > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningDUnitTest$$Lambda$111/867459734.run > with 0 args on object: > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningDUnitTest$$Lambda$111/867459734@759136fd > [vm_0][info 2016/07/26 17:38:17.313 PDT <RMI TCP Connection(4)-10.118.32.37> > tid=0x44e] Size of the Partitioned Region @330ffcc1 [path='/Quarter'; > dataPolicy=PARTITION; prId=30; isDestroyed=false; isClosed=false; > retryTimeout=3600000; serialNumber=582; partition > attributes=PartitionAttributes@81032140[redundantCopies=3;localMaxMemory=40;totalMaxMemory=2147483647;totalNumBuckets=12;partitionResolver=com.gemstone.gemfire.internal.cache.partitioned.fixed.QuarterPartitionResolver@a139d26;colocatedWith=null;recoveryDelay=-1;startupRecoveryDelay=0;FixedPartitionAttributes=[FixedPartitionAttributes@[partitionName=Q1;isPrimary=true;numBuckets=3], > FixedPartitionAttributes@[partitionName=Q3;isPrimary=false;numBuckets=3], > FixedPartitionAttributes@[partitionName=Q4;isPrimary=false;numBuckets=3]];partitionListeners=null]; > on VM japan(16120)<ec><v246>:1026] in this VM :- 9List of buckets : [0, 1, > 2, 3, 4, 5, 6, 7, 8] > [vm_0][info 2016/07/26 17:38:17.313 PDT <RMI TCP Connection(4)-10.118.32.37> > tid=0x44e] Size of primary buckets the Partitioned Region @330ffcc1 > [path='/Quarter'; dataPolicy=PARTITION; prId=30; isDestroyed=false; > isClosed=false; retryTimeout=3600000; serialNumber=582; partition > attributes=PartitionAttributes@81032140[redundantCopies=3;localMaxMemory=40;totalMaxMemory=2147483647;totalNumBuckets=12;partitionResolver=com.gemstone.gemfire.internal.cache.partitioned.fixed.QuarterPartitionResolver@a139d26;colocatedWith=null;recoveryDelay=-1;startupRecoveryDelay=0;FixedPartitionAttributes=[FixedPartitionAttributes@[partitionName=Q1;isPrimary=true;numBuckets=3], > FixedPartitionAttributes@[partitionName=Q3;isPrimary=false;numBuckets=3], > FixedPartitionAttributes@[partitionName=Q4;isPrimary=false;numBuckets=3]];partitionListeners=null]; > on VM japan(16120)<ec><v246>:1026] in this VM :- 6 > [vm_0][info 2016/07/26 17:38:17.313 PDT <RMI TCP Connection(4)-10.118.32.37> > tid=0x44e] Lit of Primaries in this VM :- [0, 1, 2, 6, 7, 8] > [vm_0][info 2016/07/26 17:38:17.313 PDT <RMI TCP Connection(4)-10.118.32.37> > tid=0x44e] Got result: null > [vm_0] from > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningDUnitTest$$Lambda$111/867459734.run > with 0 args on object: > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningDUnitTest$$Lambda$111/867459734@759136fd > (took 0 ms) > [vm_1][info 2016/07/26 17:38:17.315 PDT <RMI TCP Connection(5)-10.118.32.37> > tid=0x680] Received method: > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningDUnitTest$$Lambda$96/334480815.run > with 0 args on object: > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningDUnitTest$$Lambda$96/334480815@19a58607 > [vm_1][info 2016/07/26 17:38:17.316 PDT <RMI TCP Connection(5)-10.118.32.37> > tid=0x680] Size of the Partitioned Region @166dc9d4 [path='/Quarter'; > dataPolicy=PARTITION; prId=30; isDestroyed=false; isClosed=false; > retryTimeout=3600000; serialNumber=1317; partition > attributes=PartitionAttributes@989561565[redundantCopies=3;localMaxMemory=40;totalMaxMemory=2147483647;totalNumBuckets=12;partitionResolver=com.gemstone.gemfire.internal.cache.partitioned.fixed.QuarterPartitionResolver@3ec4115a;colocatedWith=null;recoveryDelay=-1;startupRecoveryDelay=0;FixedPartitionAttributes=[FixedPartitionAttributes@[partitionName=Q2;isPrimary=true;numBuckets=3], > FixedPartitionAttributes@[partitionName=Q3;isPrimary=false;numBuckets=3], > FixedPartitionAttributes@[partitionName=Q4;isPrimary=false;numBuckets=3]];partitionListeners=null]; > on VM japan(16124)<ec><v248>:1027] in this VM :- 9List of buckets : [3, 4, > 5, 6, 7, 8, 9, 10, 11] > [vm_1][info 2016/07/26 17:38:17.316 PDT <RMI TCP Connection(5)-10.118.32.37> > tid=0x680] Size of primary buckets the Partitioned Region @166dc9d4 > [path='/Quarter'; dataPolicy=PARTITION; prId=30; isDestroyed=false; > isClosed=false; retryTimeout=3600000; serialNumber=1317; partition > attributes=PartitionAttributes@989561565[redundantCopies=3;localMaxMemory=40;totalMaxMemory=2147483647;totalNumBuckets=12;partitionResolver=com.gemstone.gemfire.internal.cache.partitioned.fixed.QuarterPartitionResolver@3ec4115a;colocatedWith=null;recoveryDelay=-1;startupRecoveryDelay=0;FixedPartitionAttributes=[FixedPartitionAttributes@[partitionName=Q2;isPrimary=true;numBuckets=3], > FixedPartitionAttributes@[partitionName=Q3;isPrimary=false;numBuckets=3], > FixedPartitionAttributes@[partitionName=Q4;isPrimary=false;numBuckets=3]];partitionListeners=null]; > on VM japan(16124)<ec><v248>:1027] in this VM :- 3 > [vm_1][info 2016/07/26 17:38:17.316 PDT <RMI TCP Connection(5)-10.118.32.37> > tid=0x680] Lit of Primaries in this VM :- [9, 10, 11] > [vm_1][info 2016/07/26 17:38:17.316 PDT <RMI TCP Connection(5)-10.118.32.37> > tid=0x680] Got result: null > [vm_1] from > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningDUnitTest$$Lambda$96/334480815.run > with 0 args on object: > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningDUnitTest$$Lambda$96/334480815@19a58607 > (took 0 ms) > [vm_2][info 2016/07/26 17:38:17.320 PDT <RMI TCP Connection(4)-10.118.32.37> > tid=0x12] Received method: > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningDUnitTest$$Lambda$91/2125348876.run > with 0 args on object: > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningDUnitTest$$Lambda$91/2125348876@4b37470 > [vm_2][info 2016/07/26 17:38:17.321 PDT <RMI TCP Connection(4)-10.118.32.37> > tid=0x12] Size of the Partitioned Region @ff09555 [path='/Quarter'; > dataPolicy=PARTITION; prId=30; isDestroyed=false; isClosed=false; > retryTimeout=3600000; serialNumber=1161; partition > attributes=PartitionAttributes@1293197289[redundantCopies=3;localMaxMemory=40;totalMaxMemory=2147483647;totalNumBuckets=12;partitionResolver=com.gemstone.gemfire.internal.cache.partitioned.fixed.QuarterPartitionResolver@492b0293;colocatedWith=null;recoveryDelay=-1;startupRecoveryDelay=0;FixedPartitionAttributes=[FixedPartitionAttributes@[partitionName=Q3;isPrimary=true;numBuckets=3], > FixedPartitionAttributes@[partitionName=Q1;isPrimary=false;numBuckets=3], > FixedPartitionAttributes@[partitionName=Q2;isPrimary=false;numBuckets=3]];partitionListeners=null]; > on VM japan(16130)<ec><v250>:1028] in this VM :- 9List of buckets : [0, 1, > 2, 3, 4, 5, 9, 10, 11] > [vm_2][info 2016/07/26 17:38:17.321 PDT <RMI TCP Connection(4)-10.118.32.37> > tid=0x12] Size of primary buckets the Partitioned Region @ff09555 > [path='/Quarter'; dataPolicy=PARTITION; prId=30; isDestroyed=false; > isClosed=false; retryTimeout=3600000; serialNumber=1161; partition > attributes=PartitionAttributes@1293197289[redundantCopies=3;localMaxMemory=40;totalMaxMemory=2147483647;totalNumBuckets=12;partitionResolver=com.gemstone.gemfire.internal.cache.partitioned.fixed.QuarterPartitionResolver@492b0293;colocatedWith=null;recoveryDelay=-1;startupRecoveryDelay=0;FixedPartitionAttributes=[FixedPartitionAttributes@[partitionName=Q3;isPrimary=true;numBuckets=3], > FixedPartitionAttributes@[partitionName=Q1;isPrimary=false;numBuckets=3], > FixedPartitionAttributes@[partitionName=Q2;isPrimary=false;numBuckets=3]];partitionListeners=null]; > on VM japan(16130)<ec><v250>:1028] in this VM :- 3 > [vm_2][info 2016/07/26 17:38:17.321 PDT <RMI TCP Connection(4)-10.118.32.37> > tid=0x12] Lit of Primaries in this VM :- [3, 4, 5] > [vm_2][info 2016/07/26 17:38:17.321 PDT <RMI TCP Connection(4)-10.118.32.37> > tid=0x12] Got result: null > [vm_2] from > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningDUnitTest$$Lambda$91/2125348876.run > with 0 args on object: > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningDUnitTest$$Lambda$91/2125348876@4b37470 > (took 0 ms) > [info 2016/07/26 17:38:17.323 PDT <Test worker> tid=0xa] GemFireCache[id = > 479697129; isClosing = true; isShutDownAll = false; created = Tue Jul 26 > 17:38:12 PDT 2016; server = false; copyOnRead = false; lockLease = 120; > lockTimeout = 60]: Now closing. > [info 2016/07/26 17:38:17.390 PDT <Test worker> tid=0xa] Shutting down > DistributionManager japan(16028)<ec><v244>:1025. > [vm_1][info 2016/07/26 17:38:17.402 PDT <Pooled High Priority Message > Processor 3> tid=0xd17] received leave request from > japan(16028)<ec><v244>:1025 for japan(16028)<ec><v244>:1025 > [vm_1][info 2016/07/26 17:38:17.402 PDT <Pooled High Priority Message > Processor 3> tid=0xd17] Member at japan(16028)<ec><v244>:1025 gracefully left > the distributed cache: shutdown message received > [locator][info 2016/07/26 17:38:17.402 PDT <Pooled High Priority Message > Processor 1> tid=0x37] received leave request from > japan(16028)<ec><v244>:1025 for japan(16028)<ec><v244>:1025 > [locator][info 2016/07/26 17:38:17.402 PDT <Pooled High Priority Message > Processor 1> tid=0x37] Member at japan(16028)<ec><v244>:1025 gracefully left > the distributed cache: shutdown message received > [vm_0][info 2016/07/26 17:38:17.402 PDT <Pooled High Priority Message > Processor 1> tid=0xb30] received leave request from > japan(16028)<ec><v244>:1025 for japan(16028)<ec><v244>:1025 > [vm_0][info 2016/07/26 17:38:17.402 PDT <Pooled High Priority Message > Processor 1> tid=0xb30] Member at japan(16028)<ec><v244>:1025 gracefully left > the distributed cache: shutdown message received > [vm_2][info 2016/07/26 17:38:17.402 PDT <Pooled High Priority Message > Processor 2> tid=0xc1b] received leave request from > japan(16028)<ec><v244>:1025 for japan(16028)<ec><v244>:1025 > [vm_2][info 2016/07/26 17:38:17.402 PDT <Pooled High Priority Message > Processor 2> tid=0xc1b] Member at japan(16028)<ec><v244>:1025 gracefully left > the distributed cache: shutdown message received > [info 2016/07/26 17:38:17.503 PDT <Test worker> tid=0xa] Now closing > distribution for japan(16028)<ec><v244>:1025 > [info 2016/07/26 17:38:17.503 PDT <Test worker> tid=0xa] Stopping membership > services > [info 2016/07/26 17:38:17.504 PDT <Test worker> tid=0xa] GMSHealthMonitor > server socket is closed in stopServices(). > [vm_1][info 2016/07/26 17:38:17.505 PDT <unicast receiver,japan-620> > tid=0xcfb] received leave request from japan(16028)<ec><v244>:1025 for > japan(16028)<ec><v244>:1025 > [vm_2][info 2016/07/26 17:38:17.505 PDT <unicast receiver,japan-39240> > tid=0xbfd] received leave request from japan(16028)<ec><v244>:1025 for > japan(16028)<ec><v244>:1025 > [info 2016/07/26 17:38:17.505 PDT <Geode Failure Detection Server thread 0> > tid=0x152] GMSHealthMonitor server thread exiting > [locator][info 2016/07/26 17:38:17.505 PDT <unicast receiver,japan-50141> > tid=0x21] received leave request from japan(16028)<ec><v244>:1025 for > japan(16028)<ec><v244>:1025 > [info 2016/07/26 17:38:17.505 PDT <Test worker> tid=0xa] GMSHealthMonitor > serverSocketExecutor is terminated > [vm_0][info 2016/07/26 17:38:17.505 PDT <unicast receiver,japan-16947> > tid=0xb19] received leave request from japan(16028)<ec><v244>:1025 for > japan(16028)<ec><v244>:1025 > [info 2016/07/26 17:38:17.516 PDT <Test worker> tid=0xa] DistributionManager > stopped in 126ms. > [info 2016/07/26 17:38:17.516 PDT <Test worker> tid=0xa] Marking > DistributionManager japan(16028)<ec><v244>:1025 as closed. > [vm_0][info 2016/07/26 17:38:17.517 PDT <RMI TCP Connection(4)-10.118.32.37> > tid=0x44e] Received method: > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningTestBase$$Lambda$17/2140581019.run > with 0 args on object: > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningTestBase$$Lambda$17/2140581019@279116aa > [vm_0][info 2016/07/26 17:38:17.519 PDT <RMI TCP Connection(4)-10.118.32.37> > tid=0x44e] GemFireCache[id = 500122850; isClosing = true; isShutDownAll = > false; created = Tue Jul 26 17:38:13 PDT 2016; server = false; copyOnRead = > false; lockLease = 120; lockTimeout = 60]: Now closing. > [vm_0][info 2016/07/26 17:38:17.536 PDT <RMI TCP Connection(4)-10.118.32.37> > tid=0x44e] Shutting down DistributionManager japan(16120)<ec><v246>:1026. > [vm_1][info 2016/07/26 17:38:17.537 PDT <Pooled High Priority Message > Processor 3> tid=0xd17] received leave request from > japan(16120)<ec><v246>:1026 for japan(16120)<ec><v246>:1026 > [locator][info 2016/07/26 17:38:17.537 PDT <Pooled High Priority Message > Processor 1> tid=0x37] received leave request from > japan(16120)<ec><v246>:1026 for japan(16120)<ec><v246>:1026 > [locator][info 2016/07/26 17:38:17.537 PDT <Pooled High Priority Message > Processor 1> tid=0x37] Member at japan(16120)<ec><v246>:1026 gracefully left > the distributed cache: shutdown message received > [vm_1][info 2016/07/26 17:38:17.537 PDT <Pooled High Priority Message > Processor 3> tid=0xd17] Member at japan(16120)<ec><v246>:1026 gracefully left > the distributed cache: shutdown message received > [vm_2][info 2016/07/26 17:38:17.537 PDT <Pooled High Priority Message > Processor 4> tid=0xc1e] received leave request from > japan(16120)<ec><v246>:1026 for japan(16120)<ec><v246>:1026 > [vm_2][info 2016/07/26 17:38:17.537 PDT <Pooled High Priority Message > Processor 4> tid=0xc1e] Member at japan(16120)<ec><v246>:1026 gracefully left > the distributed cache: shutdown message received > [vm_0][warn 2016/07/26 17:38:17.638 PDT <Shutdown Message Thread for > japan(16120)<ec><v246>:1026> tid=0xb64] Attempting TCP/IP reconnect to > japan(16028)<ec><v244>:1025 > [vm_0][info 2016/07/26 17:38:17.638 PDT <Shutdown Message Thread for > japan(16120)<ec><v246>:1026> tid=0xb64] Connection: shared=true ordered=false > failed to connect to peer japan(16028)<ec><v244>:1025 because: > java.net.ConnectException: Connection refused > [locator][info 2016/07/26 17:38:17.738 PDT <Geode Membership View Creator> > tid=0x27] View Creator is processing 3 requests for the next membership view > [locator][info 2016/07/26 17:38:17.738 PDT <Geode Membership View Creator> > tid=0x27] preparing new view View[japan(16059:locator)<ec><v0>:1024|254] > members: [japan(16059:locator)<ec><v0>:1024, > japan(16124)<ec><v248>:1027{lead}, japan(16130)<ec><v250>:1028] shutdown: > [japan(16028)<ec><v244>:1025, japan(16120)<ec><v246>:1026] > [locator]failure detection ports: 17959 14847 47154 > [locator][info 2016/07/26 17:38:17.739 PDT <Geode Membership View Creator> > tid=0x27] finished waiting for responses to view preparation > [locator][info 2016/07/26 17:38:17.739 PDT <Geode Membership View Creator> > tid=0x27] received new view: View[japan(16059:locator)<ec><v0>:1024|254] > members: [japan(16059:locator)<ec><v0>:1024, > japan(16124)<ec><v248>:1027{lead}, japan(16130)<ec><v250>:1028] shutdown: > [japan(16028)<ec><v244>:1025, japan(16120)<ec><v246>:1026] > [locator]old view is: View[japan(16059:locator)<ec><v0>:1024|253] members: > [japan(16059:locator)<ec><v0>:1024, japan(16028)<ec><v244>:1025{lead}, > japan(16120)<ec><v246>:1026, japan(16124)<ec><v248>:1027, > japan(16130)<ec><v250>:1028] shutdown: [japan(16139)<ec><v252>:1029] > [locator][info 2016/07/26 17:38:17.739 PDT <Geode Membership View Creator> > tid=0x27] Peer locator received new membership view: > View[japan(16059:locator)<ec><v0>:1024|254] members: > [japan(16059:locator)<ec><v0>:1024, japan(16124)<ec><v248>:1027{lead}, > japan(16130)<ec><v250>:1028] shutdown: [japan(16028)<ec><v244>:1025, > japan(16120)<ec><v246>:1026] > [locator][info 2016/07/26 17:38:17.739 PDT <Geode Membership View Creator> > tid=0x27] sending new view View[japan(16059:locator)<ec><v0>:1024|254] > members: [japan(16059:locator)<ec><v0>:1024, > japan(16124)<ec><v248>:1027{lead}, japan(16130)<ec><v250>:1028] shutdown: > [japan(16028)<ec><v244>:1025, japan(16120)<ec><v246>:1026] > [locator]failure detection ports: 17959 14847 47154 > [vm_1][info 2016/07/26 17:38:17.739 PDT <unicast receiver,japan-620> > tid=0xcfb] received new view: View[japan(16059:locator)<ec><v0>:1024|254] > members: [japan(16059:locator)<ec><v0>:1024, > japan(16124)<ec><v248>:1027{lead}, japan(16130)<ec><v250>:1028] shutdown: > [japan(16028)<ec><v244>:1025, japan(16120)<ec><v246>:1026] > [vm_1]old view is: View[japan(16059:locator)<ec><v0>:1024|253] members: > [japan(16059:locator)<ec><v0>:1024, japan(16028)<ec><v244>:1025{lead}, > japan(16120)<ec><v246>:1026, japan(16124)<ec><v248>:1027, > japan(16130)<ec><v250>:1028] shutdown: [japan(16139)<ec><v252>:1029] > [vm_2][info 2016/07/26 17:38:17.739 PDT <unicast receiver,japan-39240> > tid=0xbfd] received new view: View[japan(16059:locator)<ec><v0>:1024|254] > members: [japan(16059:locator)<ec><v0>:1024, > japan(16124)<ec><v248>:1027{lead}, japan(16130)<ec><v250>:1028] shutdown: > [japan(16028)<ec><v244>:1025, japan(16120)<ec><v246>:1026] > [vm_2]old view is: View[japan(16059:locator)<ec><v0>:1024|253] members: > [japan(16059:locator)<ec><v0>:1024, japan(16028)<ec><v244>:1025{lead}, > japan(16120)<ec><v246>:1026, japan(16124)<ec><v248>:1027, > japan(16130)<ec><v250>:1028] shutdown: [japan(16139)<ec><v252>:1029] > [vm_0][warn 2016/07/26 17:38:19.639 PDT <Shutdown Message Thread for > japan(16120)<ec><v246>:1026> tid=0xb64] Connection: Attempting reconnect to > peer japan(16028)<ec><v244>:1025 > [vm_0][info 2016/07/26 17:38:20.743 PDT <P2P message reader for > japan(16059:locator)<ec><v0>:1024 shared unordered uid=213 port=42601> > tid=0xb24] Performing final check for suspect member > japan(16059:locator)<ec><v0>:1024 reason=member unexpectedly shut down > shared, unordered connection > [vm_0][info 2016/07/26 17:38:20.743 PDT <P2P message reader for > japan(16124)<ec><v248>:1027 shared unordered uid=291 port=42612> tid=0xb3d] > Performing final check for suspect member japan(16124)<ec><v248>:1027 > reason=member unexpectedly shut down shared, unordered connection > [vm_0][info 2016/07/26 17:38:20.744 PDT <P2P message reader for > japan(16130)<ec><v250>:1028 shared unordered uid=292 port=42631> tid=0xb47] > Performing final check for suspect member japan(16130)<ec><v250>:1028 > reason=member unexpectedly shut down shared, unordered connection > [vm_0][info 2016/07/26 17:38:20.744 PDT <P2P message reader for > japan(16059:locator)<ec><v0>:1024 shared unordered uid=213 port=42601> > tid=0xb24] Final check passed for suspect member > japan(16059:locator)<ec><v0>:1024 > [vm_0][info 2016/07/26 17:38:20.745 PDT <P2P message reader for > japan(16124)<ec><v248>:1027 shared unordered uid=291 port=42612> tid=0xb3d] > Final check passed for suspect member japan(16124)<ec><v248>:1027 > [vm_0][info 2016/07/26 17:38:20.745 PDT <P2P message reader for > japan(16130)<ec><v250>:1028 shared unordered uid=292 port=42631> tid=0xb47] > Final check passed for suspect member japan(16130)<ec><v250>:1028 > [vm_0][warn 2016/07/26 17:38:37.536 PDT <RMI TCP Connection(4)-10.118.32.37> > tid=0x44e] Failed sending shutdown message to peers (timeout) > [vm_0][info 2016/07/26 17:38:37.537 PDT <Shutdown Message Thread for > japan(16120)<ec><v246>:1026> tid=0xb64] Ending reconnect attempt to > japan(16028)<ec><v244>:1025 because shutdown has started. > [vm_0][info 2016/07/26 17:38:37.637 PDT <RMI TCP Connection(4)-10.118.32.37> > tid=0x44e] Now closing distribution for japan(16120)<ec><v246>:1026 > [vm_0][info 2016/07/26 17:38:37.637 PDT <RMI TCP Connection(4)-10.118.32.37> > tid=0x44e] Stopping membership services > [locator][info 2016/07/26 17:38:37.638 PDT <unicast receiver,japan-50141> > tid=0x21] received leave request from japan(16120)<ec><v246>:1026 for > japan(16120)<ec><v246>:1026 > [vm_0][info 2016/07/26 17:38:37.638 PDT <RMI TCP Connection(4)-10.118.32.37> > tid=0x44e] GMSHealthMonitor server socket is closed in stopServices(). > [vm_0][info 2016/07/26 17:38:37.638 PDT <Geode Failure Detection Server > thread 0> tid=0xb1d] GMSHealthMonitor server thread exiting > [vm_2][info 2016/07/26 17:38:37.638 PDT <unicast receiver,japan-39240> > tid=0xbfd] received leave request from japan(16120)<ec><v246>:1026 for > japan(16120)<ec><v246>:1026 > [vm_1][info 2016/07/26 17:38:37.638 PDT <unicast receiver,japan-620> > tid=0xcfb] received leave request from japan(16120)<ec><v246>:1026 for > japan(16120)<ec><v246>:1026 > [vm_0][info 2016/07/26 17:38:37.638 PDT <RMI TCP Connection(4)-10.118.32.37> > tid=0x44e] GMSHealthMonitor serverSocketExecutor is terminated > [vm_0][info 2016/07/26 17:38:37.640 PDT <RMI TCP Connection(4)-10.118.32.37> > tid=0x44e] DistributionManager stopped in 20,104ms. > [vm_0][info 2016/07/26 17:38:37.640 PDT <RMI TCP Connection(4)-10.118.32.37> > tid=0x44e] Marking DistributionManager japan(16120)<ec><v246>:1026 as closed. > [vm_0][info 2016/07/26 17:38:37.640 PDT <RMI TCP Connection(4)-10.118.32.37> > tid=0x44e] Got result: null > [vm_0] from > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningTestBase$$Lambda$17/2140581019.run > with 0 args on object: > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningTestBase$$Lambda$17/2140581019@279116aa > (took 20122 ms) > [vm_1][info 2016/07/26 17:38:37.641 PDT <RMI TCP Connection(5)-10.118.32.37> > tid=0x680] Received method: > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningTestBase$$Lambda$21/440574307.run > with 0 args on object: > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningTestBase$$Lambda$21/440574307@6dc87f91 > [vm_1][info 2016/07/26 17:38:37.642 PDT <RMI TCP Connection(5)-10.118.32.37> > tid=0x680] GemFireCache[id = 91855103; isClosing = true; isShutDownAll = > false; created = Tue Jul 26 17:38:14 PDT 2016; server = false; copyOnRead = > false; lockLease = 120; lockTimeout = 60]: Now closing. > [vm_1][info 2016/07/26 17:38:37.647 PDT <RMI TCP Connection(5)-10.118.32.37> > tid=0x680] Shutting down DistributionManager japan(16124)<ec><v248>:1027. > [vm_2][info 2016/07/26 17:38:37.647 PDT <Pooled High Priority Message > Processor 4> tid=0xc1e] received leave request from > japan(16124)<ec><v248>:1027 for japan(16124)<ec><v248>:1027 > [vm_2][info 2016/07/26 17:38:37.647 PDT <Pooled High Priority Message > Processor 4> tid=0xc1e] Member at japan(16124)<ec><v248>:1027 gracefully left > the distributed cache: shutdown message received > [locator][info 2016/07/26 17:38:37.647 PDT <Pooled High Priority Message > Processor 1> tid=0x37] received leave request from > japan(16124)<ec><v248>:1027 for japan(16124)<ec><v248>:1027 > [locator][info 2016/07/26 17:38:37.647 PDT <Pooled High Priority Message > Processor 1> tid=0x37] Member at japan(16124)<ec><v248>:1027 gracefully left > the distributed cache: shutdown message received > [vm_1][info 2016/07/26 17:38:37.748 PDT <RMI TCP Connection(5)-10.118.32.37> > tid=0x680] Now closing distribution for japan(16124)<ec><v248>:1027 > [vm_1][info 2016/07/26 17:38:37.748 PDT <RMI TCP Connection(5)-10.118.32.37> > tid=0x680] Stopping membership services > [locator][info 2016/07/26 17:38:37.748 PDT <unicast receiver,japan-50141> > tid=0x21] received leave request from japan(16124)<ec><v248>:1027 for > japan(16124)<ec><v248>:1027 > [vm_2][info 2016/07/26 17:38:37.748 PDT <unicast receiver,japan-39240> > tid=0xbfd] received leave request from japan(16124)<ec><v248>:1027 for > japan(16124)<ec><v248>:1027 > [vm_1][info 2016/07/26 17:38:37.748 PDT <RMI TCP Connection(5)-10.118.32.37> > tid=0x680] GMSHealthMonitor server socket is closed in stopServices(). > [vm_1][info 2016/07/26 17:38:37.749 PDT <Geode Failure Detection Server > thread 0> tid=0xcff] GMSHealthMonitor server thread exiting > [vm_1][info 2016/07/26 17:38:37.749 PDT <RMI TCP Connection(5)-10.118.32.37> > tid=0x680] GMSHealthMonitor serverSocketExecutor is terminated > [vm_1][info 2016/07/26 17:38:37.750 PDT <RMI TCP Connection(5)-10.118.32.37> > tid=0x680] DistributionManager stopped in 103ms. > [vm_1][info 2016/07/26 17:38:37.751 PDT <RMI TCP Connection(5)-10.118.32.37> > tid=0x680] Marking DistributionManager japan(16124)<ec><v248>:1027 as closed. > [vm_1][info 2016/07/26 17:38:37.751 PDT <RMI TCP Connection(5)-10.118.32.37> > tid=0x680] Got result: null > [vm_1] from > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningTestBase$$Lambda$21/440574307.run > with 0 args on object: > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningTestBase$$Lambda$21/440574307@6dc87f91 > (took 109 ms) > [vm_2][info 2016/07/26 17:38:37.751 PDT <RMI TCP Connection(4)-10.118.32.37> > tid=0x12] Received method: > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningTestBase$$Lambda$20/1166003088.run > with 0 args on object: > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningTestBase$$Lambda$20/1166003088@51c60f47 > [vm_2][info 2016/07/26 17:38:37.753 PDT <RMI TCP Connection(4)-10.118.32.37> > tid=0x12] GemFireCache[id = 520797903; isClosing = true; isShutDownAll = > false; created = Tue Jul 26 17:38:14 PDT 2016; server = false; copyOnRead = > false; lockLease = 120; lockTimeout = 60]: Now closing. > [vm_2][info 2016/07/26 17:38:37.756 PDT <RMI TCP Connection(4)-10.118.32.37> > tid=0x12] Shutting down DistributionManager japan(16130)<ec><v250>:1028. > [locator][info 2016/07/26 17:38:37.756 PDT <Pooled High Priority Message > Processor 1> tid=0x37] received leave request from > japan(16130)<ec><v250>:1028 for japan(16130)<ec><v250>:1028 > [locator][info 2016/07/26 17:38:37.756 PDT <Pooled High Priority Message > Processor 1> tid=0x37] Member at japan(16130)<ec><v250>:1028 gracefully left > the distributed cache: shutdown message received > [vm_2][info 2016/07/26 17:38:37.857 PDT <RMI TCP Connection(4)-10.118.32.37> > tid=0x12] Now closing distribution for japan(16130)<ec><v250>:1028 > [vm_2][info 2016/07/26 17:38:37.857 PDT <RMI TCP Connection(4)-10.118.32.37> > tid=0x12] Stopping membership services > [locator][info 2016/07/26 17:38:37.858 PDT <unicast receiver,japan-50141> > tid=0x21] received leave request from japan(16130)<ec><v250>:1028 for > japan(16130)<ec><v250>:1028 > [vm_2][info 2016/07/26 17:38:37.858 PDT <RMI TCP Connection(4)-10.118.32.37> > tid=0x12] GMSHealthMonitor server socket is closed in stopServices(). > [vm_2][info 2016/07/26 17:38:37.858 PDT <Geode Failure Detection Server > thread 0> tid=0xc01] GMSHealthMonitor server thread exiting > [vm_2][info 2016/07/26 17:38:37.858 PDT <RMI TCP Connection(4)-10.118.32.37> > tid=0x12] GMSHealthMonitor serverSocketExecutor is terminated > [vm_2][info 2016/07/26 17:38:37.861 PDT <RMI TCP Connection(4)-10.118.32.37> > tid=0x12] DistributionManager stopped in 105ms. > [vm_2][info 2016/07/26 17:38:37.861 PDT <RMI TCP Connection(4)-10.118.32.37> > tid=0x12] Marking DistributionManager japan(16130)<ec><v250>:1028 as closed. > [vm_2][info 2016/07/26 17:38:37.861 PDT <RMI TCP Connection(4)-10.118.32.37> > tid=0x12] Got result: null > [vm_2] from > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningTestBase$$Lambda$20/1166003088.run > with 0 args on object: > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningTestBase$$Lambda$20/1166003088@51c60f47 > (took 109 ms) > [vm_3][info 2016/07/26 17:38:37.862 PDT <RMI TCP Connection(6)-10.118.32.37> > tid=0x12] Received method: > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningTestBase$$Lambda$18/1058655980.run > with 0 args on object: > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningTestBase$$Lambda$18/1058655980@520ab810 > [vm_3][info 2016/07/26 17:38:37.862 PDT <RMI TCP Connection(6)-10.118.32.37> > tid=0x12] Got result: null > [vm_3] from > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningTestBase$$Lambda$18/1058655980.run > with 0 args on object: > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningTestBase$$Lambda$18/1058655980@520ab810 > (took 0 ms) > [vm_0][info 2016/07/26 17:38:37.862 PDT <RMI TCP Connection(4)-10.118.32.37> > tid=0x44e] Received method: > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningTestBase$$Lambda$18/1180821158.run > with 0 args on object: > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningTestBase$$Lambda$18/1180821158@69e6f47 > [vm_0][info 2016/07/26 17:38:37.863 PDT <RMI TCP Connection(4)-10.118.32.37> > tid=0x44e] Got result: null > [vm_0] from > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningTestBase$$Lambda$18/1180821158.run > with 0 args on object: > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningTestBase$$Lambda$18/1180821158@69e6f47 > (took 0 ms) > [vm_1][info 2016/07/26 17:38:37.863 PDT <RMI TCP Connection(5)-10.118.32.37> > tid=0x680] Received method: > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningTestBase$$Lambda$22/1198789181.run > with 0 args on object: > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningTestBase$$Lambda$22/1198789181@75d7a804 > [vm_1][info 2016/07/26 17:38:37.863 PDT <RMI TCP Connection(5)-10.118.32.37> > tid=0x680] Got result: null > [vm_1] from > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningTestBase$$Lambda$22/1198789181.run > with 0 args on object: > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningTestBase$$Lambda$22/1198789181@75d7a804 > (took 0 ms) > [vm_2][info 2016/07/26 17:38:37.864 PDT <RMI TCP Connection(4)-10.118.32.37> > tid=0x12] Received method: > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningTestBase$$Lambda$21/1279519409.run > with 0 args on object: > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningTestBase$$Lambda$21/1279519409@24e03dfc > [vm_2][info 2016/07/26 17:38:37.864 PDT <RMI TCP Connection(4)-10.118.32.37> > tid=0x12] Got result: null > [vm_2] from > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningTestBase$$Lambda$21/1279519409.run > with 0 args on object: > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningTestBase$$Lambda$21/1279519409@24e03dfc > (took 0 ms) > [vm_3][info 2016/07/26 17:38:37.864 PDT <RMI TCP Connection(6)-10.118.32.37> > tid=0x12] Received method: > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningTestBase$$Lambda$19/1296152409.run > with 0 args on object: > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningTestBase$$Lambda$19/1296152409@22407782 > [vm_3][info 2016/07/26 17:38:37.865 PDT <RMI TCP Connection(6)-10.118.32.37> > tid=0x12] Got result: null > [vm_3] from > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningTestBase$$Lambda$19/1296152409.run > with 0 args on object: > com.gemstone.gemfire.internal.cache.partitioned.fixed.FixedPartitioningTestBase$$Lambda$19/1296152409@22407782 > (took 0 ms) > [vm_0][info 2016/07/26 17:38:37.865 PDT <RMI TCP Connection(4)-10.118.32.37> > tid=0x44e] Received method: com.gemstone.gemfire.test.dunit.NamedRunnable.run > with 0 args on object: runnable(tearDownCreationStackGenerator) > [vm_0][info 2016/07/26 17:38:37.865 PDT <RMI TCP Connection(4)-10.118.32.37> > tid=0x44e] Got result: null > [vm_0] from com.gemstone.gemfire.test.dunit.NamedRunnable.run with 0 args on > object: runnable(tearDownCreationStackGenerator) (took 0 ms) > [vm_1][info 2016/07/26 17:38:37.866 PDT <RMI TCP Connection(5)-10.118.32.37> > tid=0x680] Received method: com.gemstone.gemfire.test.dunit.NamedRunnable.run > with 0 args on object: runnable(tearDownCreationStackGenerator) > [vm_1][info 2016/07/26 17:38:37.866 PDT <RMI TCP Connection(5)-10.118.32.37> > tid=0x680] Got result: null > [vm_1] from com.gemstone.gemfire.test.dunit.NamedRunnable.run with 0 args on > object: runnable(tearDownCreationStackGenerator) (took 0 ms) > [vm_2][info 2016/07/26 17:38:37.866 PDT <RMI TCP Connection(4)-10.118.32.37> > tid=0x12] Received method: com.gemstone.gemfire.test.dunit.NamedRunnable.run > with 0 args on object: runnable(tearDownCreationStackGenerator) > [vm_2][info 2016/07/26 17:38:37.866 PDT <RMI TCP Connection(4)-10.118.32.37> > tid=0x12] Got result: null > [vm_2] from com.gemstone.gemfire.test.dunit.NamedRunnable.run with 0 args on > object: runnable(tearDownCreationStackGenerator) (took 0 ms) > [vm_3][info 2016/07/26 17:38:37.867 PDT <RMI TCP Connection(6)-10.118.32.37> > tid=0x12] Received method: com.gemstone.gemfire.test.dunit.NamedRunnable.run > with 0 args on object: runnable(tearDownCreationStackGenerator) > [vm_3][info 2016/07/26 17:38:37.867 PDT <RMI TCP Connection(6)-10.118.32.37> > tid=0x12] Got result: null > [vm_3] from com.gemstone.gemfire.test.dunit.NamedRunnable.run with 0 args on > object: runnable(tearDownCreationStackGenerator) (took 0 ms) > [vm_0][info 2016/07/26 17:38:37.867 PDT <RMI TCP Connection(4)-10.118.32.37> > tid=0x44e] Received method: com.gemstone.gemfire.test.dunit.NamedRunnable.run > with 0 args on object: runnable(tearDownVM) > [vm_0][info 2016/07/26 17:38:37.867 PDT <RMI TCP Connection(4)-10.118.32.37> > tid=0x44e] Got result: null > [vm_0] from com.gemstone.gemfire.test.dunit.NamedRunnable.run with 0 args on > object: runnable(tearDownVM) (took 0 ms) > [vm_1][info 2016/07/26 17:38:37.868 PDT <RMI TCP Connection(5)-10.118.32.37> > tid=0x680] Received method: com.gemstone.gemfire.test.dunit.NamedRunnable.run > with 0 args on object: runnable(tearDownVM) > [vm_1][info 2016/07/26 17:38:37.868 PDT <RMI TCP Connection(5)-10.118.32.37> > tid=0x680] Got result: null > [vm_1] from com.gemstone.gemfire.test.dunit.NamedRunnable.run with 0 args on > object: runnable(tearDownVM) (took 0 ms) > [vm_2][info 2016/07/26 17:38:37.868 PDT <RMI TCP Connection(4)-10.118.32.37> > tid=0x12] Received method: com.gemstone.gemfire.test.dunit.NamedRunnable.run > with 0 args on object: runnable(tearDownVM) > [vm_2][info 2016/07/26 17:38:37.868 PDT <RMI TCP Connection(4)-10.118.32.37> > tid=0x12] Got result: null > [vm_2] from com.gemstone.gemfire.test.dunit.NamedRunnable.run with 0 args on > object: runnable(tearDownVM) (took 0 ms) > [vm_3][info 2016/07/26 17:38:37.869 PDT <RMI TCP Connection(6)-10.118.32.37> > tid=0x12] Received method: com.gemstone.gemfire.test.dunit.NamedRunnable.run > with 0 args on object: runnable(tearDownVM) > [vm_3][info 2016/07/26 17:38:37.869 PDT <RMI TCP Connection(6)-10.118.32.37> > tid=0x12] Got result: null > [vm_3] from com.gemstone.gemfire.test.dunit.NamedRunnable.run with 0 args on > object: runnable(tearDownVM) (took 0 ms) > [locator][info 2016/07/26 17:38:37.870 PDT <RMI TCP > Connection(9)-10.118.32.37> tid=0x248] Received method: > com.gemstone.gemfire.test.dunit.internal.JUnit4DistributedTestCase$$Lambda$6/210471107.run > with 0 args on object: > com.gemstone.gemfire.test.dunit.internal.JUnit4DistributedTestCase$$Lambda$6/210471107@7c07e12 > [locator][info 2016/07/26 17:38:37.870 PDT <RMI TCP > Connection(9)-10.118.32.37> tid=0x248] Got result: null > [locator] from > com.gemstone.gemfire.test.dunit.internal.JUnit4DistributedTestCase$$Lambda$6/210471107.run > with 0 args on object: > com.gemstone.gemfire.test.dunit.internal.JUnit4DistributedTestCase$$Lambda$6/210471107@7c07e12 > (took 0 ms) > Standard Error > Suspicious strings were written to the log during this run. > Fix the strings or use IgnoredException.addIgnoredException to ignore. > ----------------------------------------------------------------------- > Found suspect string in log4j at line 2007 > [fatal 2016/07/26 17:32:44.637 PDT <Geode Membership View Creator> tid=0x27] > Possible loss of quorum due to the loss of 2 cache processes: > [japan(16120)<ec><v112>:1025, japan(16124)<ec><v114>:1026] > ----------------------------------------------------------------------- > Found suspect string in log4j at line 2025 > [fatal 2016/07/26 17:32:44.639 PDT <unicast receiver,japan-56628> tid=0x4e1] > Possible loss of quorum due to the loss of 2 cache processes: > [japan(16120)<ec><v112>:1025, japan(16124)<ec><v114>:1026] > ----------------------------------------------------------------------- > Found suspect string in log4j at line 2076 > [fatal 2016/07/26 17:36:22.280 PDT <unicast receiver,japan-45900> tid=0x5e5] > Membership service failure: Failed to acknowledge a new membership view and > then failed tcp/ip connection attempt > com.gemstone.gemfire.ForcedDisconnectException: Failed to acknowledge a new > membership view and then failed tcp/ip connection attempt > at > com.gemstone.gemfire.distributed.internal.membership.gms.mgr.GMSMembershipManager.forceDisconnect(GMSMembershipManager.java:2605) > at > com.gemstone.gemfire.distributed.internal.membership.gms.membership.GMSJoinLeave.forceDisconnect(GMSJoinLeave.java:918) > at > com.gemstone.gemfire.distributed.internal.membership.gms.membership.GMSJoinLeave.processRemoveRequest(GMSJoinLeave.java:593) > at > com.gemstone.gemfire.distributed.internal.membership.gms.membership.GMSJoinLeave.processMessage(GMSJoinLeave.java:1555) > at > com.gemstone.gemfire.distributed.internal.membership.gms.messenger.JGroupsMessenger$JGroupsReceiver.receive(JGroupsMessenger.java:1065) > at org.jgroups.JChannel.invokeCallback(JChannel.java:816) > at org.jgroups.JChannel.up(JChannel.java:741) > at org.jgroups.stack.ProtocolStack.up(ProtocolStack.java:1030) > at org.jgroups.protocols.FRAG2.up(FRAG2.java:165) > at org.jgroups.protocols.FlowControl.up(FlowControl.java:390) > at org.jgroups.protocols.UNICAST3.deliverMessage(UNICAST3.java:1070) > at org.jgroups.protocols.UNICAST3.handleDataReceived(UNICAST3.java:785) > at org.jgroups.protocols.UNICAST3.up(UNICAST3.java:426) > at > com.gemstone.gemfire.distributed.internal.membership.gms.messenger.StatRecorder.up(StatRecorder.java:75) > at > com.gemstone.gemfire.distributed.internal.membership.gms.messenger.AddressManager.up(AddressManager.java:75) > at org.jgroups.protocols.TP.passMessageUp(TP.java:1601) > at org.jgroups.protocols.TP$SingleMessageHandler.run(TP.java:1817) > at org.jgroups.util.DirectExecutor.execute(DirectExecutor.java:10) > at org.jgroups.protocols.TP.handleSingleMessage(TP.java:1729) > at org.jgroups.protocols.TP.receive(TP.java:1654) > at > com.gemstone.gemfire.distributed.internal.membership.gms.messenger.Transport.receive(Transport.java:162) > at org.jgroups.protocols.UDP$PacketReceiver.run(UDP.java:701) > at java.lang.Thread.run(Thread.java:745) > ----------------------------------------------------------------------- > Found suspect string in log4j at line 2108 > [fatal 2016/07/26 17:36:22.293 PDT <unicast receiver,japan-21319> tid=0x411] > Membership service failure: Failed to acknowledge a new membership view and > then failed tcp/ip connection attempt > com.gemstone.gemfire.ForcedDisconnectException: Failed to acknowledge a new > membership view and then failed tcp/ip connection attempt > at > com.gemstone.gemfire.distributed.internal.membership.gms.mgr.GMSMembershipManager.forceDisconnect(GMSMembershipManager.java:2605) > at > com.gemstone.gemfire.distributed.internal.membership.gms.membership.GMSJoinLeave.forceDisconnect(GMSJoinLeave.java:918) > at > com.gemstone.gemfire.distributed.internal.membership.gms.membership.GMSJoinLeave.processRemoveRequest(GMSJoinLeave.java:593) > at > com.gemstone.gemfire.distributed.internal.membership.gms.membership.GMSJoinLeave.processMessage(GMSJoinLeave.java:1555) > at > com.gemstone.gemfire.distributed.internal.membership.gms.messenger.JGroupsMessenger$JGroupsReceiver.receive(JGroupsMessenger.java:1065) > at org.jgroups.JChannel.invokeCallback(JChannel.java:816) > at org.jgroups.JChannel.up(JChannel.java:741) > at org.jgroups.stack.ProtocolStack.up(ProtocolStack.java:1030) > at org.jgroups.protocols.FRAG2.up(FRAG2.java:165) > at org.jgroups.protocols.FlowControl.up(FlowControl.java:390) > at org.jgroups.protocols.UNICAST3.deliverMessage(UNICAST3.java:1070) > at org.jgroups.protocols.UNICAST3.handleDataReceived(UNICAST3.java:785) > at org.jgroups.protocols.UNICAST3.up(UNICAST3.java:426) > at > com.gemstone.gemfire.distributed.internal.membership.gms.messenger.StatRecorder.up(StatRecorder.java:75) > at > com.gemstone.gemfire.distributed.internal.membership.gms.messenger.AddressManager.up(AddressManager.java:75) > at org.jgroups.protocols.TP.passMessageUp(TP.java:1601) > at org.jgroups.protocols.TP$SingleMessageHandler.run(TP.java:1817) > at org.jgroups.util.DirectExecutor.execute(DirectExecutor.java:10) > at org.jgroups.protocols.TP.handleSingleMessage(TP.java:1729) > at org.jgroups.protocols.TP.receive(TP.java:1654) > at > com.gemstone.gemfire.distributed.internal.membership.gms.messenger.Transport.receive(Transport.java:162) > at org.jgroups.protocols.UDP$PacketReceiver.run(UDP.java:701) > at java.lang.Thread.run(Thread.java:745) > ----------------------------------------------------------------------- > Found suspect string in log4j at line 2142 > [fatal 2016/07/26 17:36:22.301 PDT <unicast receiver,japan-56628> tid=0x4e1] > Membership service failure: Member isn't responding to heartbeat requests > com.gemstone.gemfire.ForcedDisconnectException: Member isn't responding to > heartbeat requests > at > com.gemstone.gemfire.distributed.internal.membership.gms.mgr.GMSMembershipManager.forceDisconnect(GMSMembershipManager.java:2605) > at > com.gemstone.gemfire.distributed.internal.membership.gms.membership.GMSJoinLeave.forceDisconnect(GMSJoinLeave.java:918) > at > com.gemstone.gemfire.distributed.internal.membership.gms.membership.GMSJoinLeave.processRemoveRequest(GMSJoinLeave.java:593) > at > com.gemstone.gemfire.distributed.internal.membership.gms.membership.GMSJoinLeave.processMessage(GMSJoinLeave.java:1555) > at > com.gemstone.gemfire.distributed.internal.membership.gms.messenger.JGroupsMessenger$JGroupsReceiver.receive(JGroupsMessenger.java:1065) > at org.jgroups.JChannel.invokeCallback(JChannel.java:816) > at org.jgroups.JChannel.up(JChannel.java:741) > at org.jgroups.stack.ProtocolStack.up(ProtocolStack.java:1030) > at org.jgroups.protocols.FRAG2.up(FRAG2.java:165) > at org.jgroups.protocols.FlowControl.up(FlowControl.java:390) > at org.jgroups.protocols.UNICAST3.deliverMessage(UNICAST3.java:1070) > at org.jgroups.protocols.UNICAST3.handleDataReceived(UNICAST3.java:785) > at org.jgroups.protocols.UNICAST3.up(UNICAST3.java:426) > at > com.gemstone.gemfire.distributed.internal.membership.gms.messenger.StatRecorder.up(StatRecorder.java:75) > at > com.gemstone.gemfire.distributed.internal.membership.gms.messenger.AddressManager.up(AddressManager.java:75) > at org.jgroups.protocols.TP.passMessageUp(TP.java:1601) > at org.jgroups.protocols.TP$SingleMessageHandler.run(TP.java:1817) > at org.jgroups.util.DirectExecutor.execute(DirectExecutor.java:10) > at org.jgroups.protocols.TP.handleSingleMessage(TP.java:1729) > at org.jgroups.protocols.TP.receive(TP.java:1654) > at > com.gemstone.gemfire.distributed.internal.membership.gms.messenger.Transport.receive(Transport.java:162) > at org.jgroups.protocols.UDP$PacketReceiver.run(UDP.java:701) > at java.lang.Thread.run(Thread.java:745) > ----------------------------------------------------------------------- > Found suspect string in log4j at line 2182 > [fatal 2016/07/26 17:36:22.331 PDT <unicast receiver,japan-57884> tid=0x64c] > Membership service failure: Member isn't responding to heartbeat requests > com.gemstone.gemfire.ForcedDisconnectException: Member isn't responding to > heartbeat requests > at > com.gemstone.gemfire.distributed.internal.membership.gms.mgr.GMSMembershipManager.forceDisconnect(GMSMembershipManager.java:2605) > at > com.gemstone.gemfire.distributed.internal.membership.gms.membership.GMSJoinLeave.forceDisconnect(GMSJoinLeave.java:918) > at > com.gemstone.gemfire.distributed.internal.membership.gms.membership.GMSJoinLeave.processRemoveRequest(GMSJoinLeave.java:593) > at > com.gemstone.gemfire.distributed.internal.membership.gms.membership.GMSJoinLeave.processMessage(GMSJoinLeave.java:1555) > at > com.gemstone.gemfire.distributed.internal.membership.gms.messenger.JGroupsMessenger$JGroupsReceiver.receive(JGroupsMessenger.java:1065) > at org.jgroups.JChannel.invokeCallback(JChannel.java:816) > at org.jgroups.JChannel.up(JChannel.java:741) > at org.jgroups.stack.ProtocolStack.up(ProtocolStack.java:1030) > at org.jgroups.protocols.FRAG2.up(FRAG2.java:165) > at org.jgroups.protocols.FlowControl.up(FlowControl.java:390) > at org.jgroups.protocols.UNICAST3.deliverMessage(UNICAST3.java:1070) > at org.jgroups.protocols.UNICAST3.handleDataReceived(UNICAST3.java:785) > at org.jgroups.protocols.UNICAST3.up(UNICAST3.java:426) > at > com.gemstone.gemfire.distributed.internal.membership.gms.messenger.StatRecorder.up(StatRecorder.java:75) > at > com.gemstone.gemfire.distributed.internal.membership.gms.messenger.AddressManager.up(AddressManager.java:75) > at org.jgroups.protocols.TP.passMessageUp(TP.java:1601) > at org.jgroups.protocols.TP$SingleMessageHandler.run(TP.java:1817) > at org.jgroups.util.DirectExecutor.execute(DirectExecutor.java:10) > at org.jgroups.protocols.TP.handleSingleMessage(TP.java:1729) > at org.jgroups.protocols.TP.receive(TP.java:1654) > at > com.gemstone.gemfire.distributed.internal.membership.gms.messenger.Transport.receive(Transport.java:162) > at org.jgroups.protocols.UDP$PacketReceiver.run(UDP.java:701) > at java.lang.Thread.run(Thread.java:745) > {noformat} -- This message was sent by Atlassian JIRA (v6.3.4#6332)