[ https://issues.apache.org/jira/browse/PHOENIX-4159?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16155653#comment-16155653 ]
Josh Mahonin commented on PHOENIX-4159: --------------------------------------- Patch attached. The specific issue had to do with not properly deleting the 'TemporaryFolder' in the underlying BaseHBaseManagedTimeIT when used across two tests, not were we properly invoking the underlying 'cleanUpAfterTest' method. Scala isn't wonderful about extending abstract Java classes, and Scalatest doesn't seem to support the @ClassRule annotation properly, so we're mocking the behaviour ourselves. It's not perfect, but should work well enough and is back to the originally intended behaviour. In fixing the above issue, it raised another one, in that JUnit 4.12 doesn't support parallel test execution using TemporaryFolders. It's fixed in JUnit 4.13 (not yet released), so I've disabled the parallel testing for now. The phoenix-spark integration tests take 4m48s on my machine. For review [~samarthjain] > phoenix-spark tests are failing > ------------------------------- > > Key: PHOENIX-4159 > URL: https://issues.apache.org/jira/browse/PHOENIX-4159 > Project: Phoenix > Issue Type: Bug > Reporter: Samarth Jain > Attachments: PHOENIX-4159.patch > > > In few of the runs where we were able to get successful test runs for > phoenix-core, we ran into failures for the phoenix-spark module. > Sample run - https://builds.apache.org/job/Phoenix-master/1762/console > [~jmahonin] - would you mind taking a look. Copy pasting here a possibly > relevant stacktrace in case the link is no longer working: > {code} > Formatting using clusterid: testClusterID > 1 [ScalaTest-4] ERROR org.apache.hadoop.hdfs.MiniDFSCluster - IOE > creating namenodes. Permissions dump: > path > '/home/jenkins/jenkins-slave/workspace/Phoenix-master/phoenix-spark/target/test-data/fa615cb3-a0d9-4c9e-90eb-acd0c7d46d9b/dfscluster_1ce5f5c4-f355-4111-a763-4ab777941386/dfs/data': > > > absolute:/home/jenkins/jenkins-slave/workspace/Phoenix-master/phoenix-spark/target/test-data/fa615cb3-a0d9-4c9e-90eb-acd0c7d46d9b/dfscluster_1ce5f5c4-f355-4111-a763-4ab777941386/dfs/data > permissions: ---- > path > '/home/jenkins/jenkins-slave/workspace/Phoenix-master/phoenix-spark/target/test-data/fa615cb3-a0d9-4c9e-90eb-acd0c7d46d9b/dfscluster_1ce5f5c4-f355-4111-a763-4ab777941386/dfs': > > > absolute:/home/jenkins/jenkins-slave/workspace/Phoenix-master/phoenix-spark/target/test-data/fa615cb3-a0d9-4c9e-90eb-acd0c7d46d9b/dfscluster_1ce5f5c4-f355-4111-a763-4ab777941386/dfs > permissions: drwx > path > '/home/jenkins/jenkins-slave/workspace/Phoenix-master/phoenix-spark/target/test-data/fa615cb3-a0d9-4c9e-90eb-acd0c7d46d9b/dfscluster_1ce5f5c4-f355-4111-a763-4ab777941386': > > > absolute:/home/jenkins/jenkins-slave/workspace/Phoenix-master/phoenix-spark/target/test-data/fa615cb3-a0d9-4c9e-90eb-acd0c7d46d9b/dfscluster_1ce5f5c4-f355-4111-a763-4ab777941386 > permissions: drwx > path > '/home/jenkins/jenkins-slave/workspace/Phoenix-master/phoenix-spark/target/test-data/fa615cb3-a0d9-4c9e-90eb-acd0c7d46d9b': > > > absolute:/home/jenkins/jenkins-slave/workspace/Phoenix-master/phoenix-spark/target/test-data/fa615cb3-a0d9-4c9e-90eb-acd0c7d46d9b > permissions: drwx > path > '/home/jenkins/jenkins-slave/workspace/Phoenix-master/phoenix-spark/target/test-data': > > > absolute:/home/jenkins/jenkins-slave/workspace/Phoenix-master/phoenix-spark/target/test-data > permissions: drwx > path > '/home/jenkins/jenkins-slave/workspace/Phoenix-master/phoenix-spark/target': > > absolute:/home/jenkins/jenkins-slave/workspace/Phoenix-master/phoenix-spark/target > permissions: drwx > path '/home/jenkins/jenkins-slave/workspace/Phoenix-master/phoenix-spark': > > absolute:/home/jenkins/jenkins-slave/workspace/Phoenix-master/phoenix-spark > permissions: drwx > path '/home/jenkins/jenkins-slave/workspace/Phoenix-master': > absolute:/home/jenkins/jenkins-slave/workspace/Phoenix-master > permissions: drwx > path '/home/jenkins/jenkins-slave/workspace': > absolute:/home/jenkins/jenkins-slave/workspace > permissions: drwx > path '/home/jenkins/jenkins-slave': > absolute:/home/jenkins/jenkins-slave > permissions: drwx > path '/home/jenkins': > absolute:/home/jenkins > permissions: drwx > path '/home': > absolute:/home > permissions: dr-x > path '/': > absolute:/ > permissions: dr-x > java.io.IOException: Cannot create directory > /home/jenkins/jenkins-slave/workspace/Phoenix-master/phoenix-spark/target/test-data/fa615cb3-a0d9-4c9e-90eb-acd0c7d46d9b/dfscluster_1ce5f5c4-f355-4111-a763-4ab777941386/dfs/name1/current > at > org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.clearDirectory(Storage.java:337) > at > org.apache.hadoop.hdfs.server.namenode.NNStorage.format(NNStorage.java:548) > at > org.apache.hadoop.hdfs.server.namenode.NNStorage.format(NNStorage.java:569) > at > org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:161) > at > org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:991) > at > org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:342) > at > org.apache.hadoop.hdfs.DFSTestUtil.formatNameNode(DFSTestUtil.java:176) > at > org.apache.hadoop.hdfs.MiniDFSCluster.createNameNodesAndSetConf(MiniDFSCluster.java:973) > at > org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:811) > at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:742) > at > org.apache.hadoop.hbase.HBaseTestingUtility.startMiniDFSCluster(HBaseTestingUtility.java:625) > at > org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:1022) > at > org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:903) > at > org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:885) > at > org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:872) > at org.apache.phoenix.query.BaseTest.initMiniCluster(BaseTest.java:522) > at org.apache.phoenix.query.BaseTest.setUpTestCluster(BaseTest.java:442) > at > org.apache.phoenix.query.BaseTest.checkClusterInitialized(BaseTest.java:424) > at org.apache.phoenix.query.BaseTest.setUpTestDriver(BaseTest.java:495) > at org.apache.phoenix.query.BaseTest.setUpTestDriver(BaseTest.java:490) > at > org.apache.phoenix.end2end.BaseHBaseManagedTimeIT.doSetup(BaseHBaseManagedTimeIT.java:57) > at > org.apache.phoenix.spark.PhoenixSparkITHelper$.doSetup(AbstractPhoenixSparkIT.scala:33) > at > org.apache.phoenix.spark.AbstractPhoenixSparkIT.beforeAll(AbstractPhoenixSparkIT.scala:88) > at > org.scalatest.BeforeAndAfterAll$class.beforeAll(BeforeAndAfterAll.scala:187) > at > org.apache.phoenix.spark.AbstractPhoenixSparkIT.beforeAll(AbstractPhoenixSparkIT.scala:44) > at > org.scalatest.BeforeAndAfterAll$class.run(BeforeAndAfterAll.scala:253) > at > org.apache.phoenix.spark.AbstractPhoenixSparkIT.run(AbstractPhoenixSparkIT.scala:44) > at org.scalatest.tools.SuiteRunner.run(SuiteRunner.scala:55) > at > java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471) > at java.util.concurrent.FutureTask.run(FutureTask.java:262) > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) > at java.lang.Thread.run(Thread.java:745) > Exception encountered when invoking run on a nested suite - > java.io.IOException: Cannot create directory > /home/jenkins/jenkins-slave/workspace/Phoenix-master/phoenix-spark/target/test-data/fa615cb3-a0d9-4c9e-90eb-acd0c7d46d9b/dfscluster_1ce5f5c4-f355-4111-a763-4ab777941386/dfs/name1/current > *** ABORTED *** > java.lang.RuntimeException: java.io.IOException: Cannot create directory > /home/jenkins/jenkins-slave/workspace/Phoenix-master/phoenix-spark/target/test-data/fa615cb3-a0d9-4c9e-90eb-acd0c7d46d9b/dfscluster_1ce5f5c4-f355-4111-a763-4ab777941386/dfs/name1/current > at org.apache.phoenix.query.BaseTest.initMiniCluster(BaseTest.java:525) > at org.apache.phoenix.query.BaseTest.setUpTestCluster(BaseTest.java:442) > at > org.apache.phoenix.query.BaseTest.checkClusterInitialized(BaseTest.java:424) > at org.apache.phoenix.query.BaseTest.setUpTestDriver(BaseTest.java:495) > at org.apache.phoenix.query.BaseTest.setUpTestDriver(BaseTest.java:490) > at > org.apache.phoenix.end2end.BaseHBaseManagedTimeIT.doSetup(BaseHBaseManagedTimeIT.java:57) > at > org.apache.phoenix.spark.PhoenixSparkITHelper$.doSetup(AbstractPhoenixSparkIT.scala:33) > at > org.apache.phoenix.spark.AbstractPhoenixSparkIT.beforeAll(AbstractPhoenixSparkIT.scala:88) > at > org.scalatest.BeforeAndAfterAll$class.beforeAll(BeforeAndAfterAll.scala:187) > at > org.apache.phoenix.spark.AbstractPhoenixSparkIT.beforeAll(AbstractPhoenixSparkIT.scala:44) > ... > Cause: java.io.IOException: Cannot create directory > /home/jenkins/jenkins-slave/workspace/Phoenix-master/phoenix-spark/target/test-data/fa615cb3-a0d9-4c9e-90eb-acd0c7d46d9b/dfscluster_1ce5f5c4-f355-4111-a763-4ab777941386/dfs/name1/current > at > org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.clearDirectory(Storage.java:337) > at > org.apache.hadoop.hdfs.server.namenode.NNStorage.format(NNStorage.java:548) > at > org.apache.hadoop.hdfs.server.namenode.NNStorage.format(NNStorage.java:569) > at org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:161) > at org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:991) > at org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:342) > at org.apache.hadoop.hdfs.DFSTestUtil.formatNameNode(DFSTestUtil.java:176) > at > org.apache.hadoop.hdfs.MiniDFSCluster.createNameNodesAndSetConf(MiniDFSCluster.java:973) > at > org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:811) > at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:742) > ... > {code} -- This message was sent by Atlassian JIRA (v6.4.14#64029)