Christopher Tubbs created ACCUMULO-2764: -------------------------------------------
Summary: Stopping MAC before it's processes have fully started causes an indefinite hang Key: ACCUMULO-2764 URL: https://issues.apache.org/jira/browse/ACCUMULO-2764 Project: Accumulo Issue Type: Bug Components: mini Affects Versions: 1.6.0 Environment: OpenJDK 1.6.0, CentOS 6.5, 2CPU, 6GB RAM (virtual hardware) Reporter: Christopher Tubbs Fix For: 1.6.1, 1.7.0 I saw this testing 1.6.0-RC5. Calling process.destroy() and then process.waitFor(), as MiniAccumuloCluster does in it's stop method, before the process is fully started, appears to create an indefinite hang. I saw this most recently in MiniAccumuloClusterGCTest.testAccurateProcessListReturned, which gets a ProcessReference and then immediately shuts down MAC, though it was also the root cause of ACCUMULO-2756. In this instance, the test got stuck in the MAC teardown. {code:java} "main" prio=10 tid=0x00007f3cf4008800 nid=0x2b19 in Object.wait() [0x00007f3cf8f9c000] java.lang.Thread.State: WAITING (on object monitor) at java.lang.Object.wait(Native Method) - waiting on <0x00000000e29dd2e8> (a java.lang.UNIXProcess) at java.lang.Object.wait(Object.java:502) at java.lang.UNIXProcess.waitFor(UNIXProcess.java:181) - locked <0x00000000e29dd2e8> (a java.lang.UNIXProcess) at org.apache.accumulo.minicluster.impl.MiniAccumuloClusterImpl.stop(MiniAccumuloClusterImpl.java:607) at org.apache.accumulo.minicluster.impl.MiniAccumuloClusterGCTest.tearDownMiniCluster(MiniAccumuloClusterGCTest.java:74) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:622) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.runners.ParentRunner.run(ParentRunner.java:309) at org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:264) at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:153) at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:124) at org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoader(ForkedBooter.java:200) at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:153) at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:103) {code} It appears that destroy() doesn't actually succeed in destroying a process which is still starting, so the waitFor() waits indefinitely. I haven't debugged further. It may be a JVM bug, or a limitation in the java Process API, or some UNIX signal handling quirk with process instantiation that destroy() cannot know. One fix could be to make start() wait until the metadata table can be scanned before it returns, to ensure all processes are actually running and ready. Another fix would be to have the teardown code try another destroy if waitFor() doesn't return after a reasonable amount of time. -- This message was sent by Atlassian JIRA (v6.2#6252)