[ 
https://issues.apache.org/jira/browse/PHOENIX-7065?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17789260#comment-17789260
 ] 

ASF GitHub Bot commented on PHOENIX-7065:
-----------------------------------------

stoty commented on PR #118:
URL: 
https://github.com/apache/phoenix-connectors/pull/118#issuecomment-1824899171

   :broken_heart: **-1 overall**
   
   
   
   
   
   
   | Vote | Subsystem | Runtime | Comment |
   |:----:|----------:|--------:|:--------|
   | +0 :ok: |  reexec  |   3m 24s |  Docker mode activated.  |
   ||| _ Prechecks _ |
   | +1 :green_heart: |  dupname  |   0m  0s |  No case conflicting files 
found.  |
   | +1 :green_heart: |  hbaseanti  |   0m  0s |  Patch does not have any 
anti-patterns.  |
   | +1 :green_heart: |  @author  |   0m  0s |  The patch does not contain any 
@author tags.  |
   | -1 :x: |  test4tests  |   0m  0s |  The patch doesn't appear to include 
any new or modified tests. Please justify why no new tests are needed for this 
patch. Also please list what manual steps were performed to verify this patch.  
|
   ||| _ master Compile Tests _ |
   | +0 :ok: |  mvndep  |  10m 28s |  Maven dependency ordering for branch  |
   | +1 :green_heart: |  mvninstall  |   8m 30s |  master passed  |
   | +1 :green_heart: |  compile  |   1m 11s |  master passed  |
   | +1 :green_heart: |  checkstyle  |   0m  9s |  master passed  |
   | +1 :green_heart: |  javadoc  |   0m 52s |  master passed  |
   | -1 :x: |  scaladoc  |   0m 26s |  root in master failed.  |
   | +0 :ok: |  spotbugs  |   0m 54s |  phoenix5-spark in master has 40 extant 
spotbugs warnings.  |
   | +0 :ok: |  spotbugs  |   0m 36s |  phoenix5-spark3 in master has 22 extant 
spotbugs warnings.  |
   | +0 :ok: |  spotbugs  |   1m 17s |  root in master has 99 extant spotbugs 
warnings.  |
   ||| _ Patch Compile Tests _ |
   | +0 :ok: |  mvndep  |   0m 14s |  Maven dependency ordering for patch  |
   | +1 :green_heart: |  mvninstall  |   7m 42s |  the patch passed  |
   | +1 :green_heart: |  compile  |   1m 10s |  the patch passed  |
   | +1 :green_heart: |  javac  |   1m 10s |  the patch passed  |
   | +1 :green_heart: |  scalac  |   1m 10s |  the patch passed  |
   | +1 :green_heart: |  checkstyle  |   0m 10s |  the patch passed  |
   | +1 :green_heart: |  markdownlint  |   0m  2s |  There were no new 
markdownlint issues.  |
   | +1 :green_heart: |  whitespace  |   0m  0s |  The patch has no whitespace 
issues.  |
   | +1 :green_heart: |  xml  |   0m  4s |  The patch has no ill-formed XML 
file.  |
   | -1 :x: |  javadoc  |   0m 21s |  root generated 20 new + 1401 unchanged - 
20 fixed = 1421 total (was 1421)  |
   | -1 :x: |  scaladoc  |   0m 53s |  root generated 9 new + 0 unchanged - 0 
fixed = 9 total (was 0)  |
   | +1 :green_heart: |  spotbugs  |   3m 12s |  the patch passed  |
   ||| _ Other Tests _ |
   | -1 :x: |  unit  |   9m 49s |  root in the patch failed.  |
   | -1 :x: |  asflicense  |   0m 23s |  The patch generated 3 ASF License 
warnings.  |
   |  |   |  55m 24s |   |
   
   
   | Subsystem | Report/Notes |
   |----------:|:-------------|
   | Docker | ClientAPI=1.43 ServerAPI=1.43 base: 
https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-Connectors-PreCommit-GitHub-PR/job/PR-118/1/artifact/yetus-general-check/output/Dockerfile
 |
   | GITHUB PR | https://github.com/apache/phoenix-connectors/pull/118 |
   | Optional Tests | dupname asflicense javac javadoc unit xml compile scalac 
scaladoc markdownlint spotbugs hbaseanti checkstyle |
   | uname | Linux 5d89d13c9dfc 5.15.0-88-generic #98-Ubuntu SMP Mon Oct 2 
15:18:56 UTC 2023 x86_64 x86_64 x86_64 GNU/Linux |
   | Build tool | maven |
   | Personality | dev/phoenix-connectors-personality.sh |
   | git revision | master / 26c1aba |
   | Default Java | Private Build-1.8.0_242-8u242-b08-0ubuntu3~16.04-b08 |
   | scaladoc | 
https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-Connectors-PreCommit-GitHub-PR/job/PR-118/1/artifact/yetus-general-check/output/branch-scaladoc-root.txt
 |
   | javadoc | 
https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-Connectors-PreCommit-GitHub-PR/job/PR-118/1/artifact/yetus-general-check/output/diff-javadoc-javadoc-root.txt
 |
   | scaladoc | 
https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-Connectors-PreCommit-GitHub-PR/job/PR-118/1/artifact/yetus-general-check/output/diff-scaladoc-scaladoc-root.txt
 |
   | unit | 
https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-Connectors-PreCommit-GitHub-PR/job/PR-118/1/artifact/yetus-general-check/output/patch-unit-root.txt
 |
   |  Test Results | 
https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-Connectors-PreCommit-GitHub-PR/job/PR-118/1/testReport/
 |
   | asflicense | 
https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-Connectors-PreCommit-GitHub-PR/job/PR-118/1/artifact/yetus-general-check/output/patch-asflicense-problems.txt
 |
   | Max. process+thread count | 1820 (vs. ulimit of 30000) |
   | modules | C: phoenix5-spark phoenix5-spark3 phoenix5-spark3-it . U: . |
   | Console output | 
https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-Connectors-PreCommit-GitHub-PR/job/PR-118/1/console
 |
   | versions | git=2.7.4 maven=3.3.9 spotbugs=4.1.3 markdownlint=0.22.0 |
   | Powered by | Apache Yetus 0.12.0 https://yetus.apache.org |
   
   
   This message was automatically generated.
   
   




> Spark3 connector tests fail with Spark 3.4.1
> --------------------------------------------
>
>                 Key: PHOENIX-7065
>                 URL: https://issues.apache.org/jira/browse/PHOENIX-7065
>             Project: Phoenix
>          Issue Type: New Feature
>          Components: connectors, spark-connector
>    Affects Versions: connectors-6.0.0
>            Reporter: Istvan Toth
>            Assignee: Istvan Toth
>            Priority: Major
>
> Probably some kind of dependency version conflict with minicluster.
> {noformat}
> [INFO] Running org.apache.phoenix.spark.SaltedTableIT
> [ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.008 
> s <<< FAILURE! - in org.apache.phoenix.spark.SaltedTableIT
> [ERROR] org.apache.phoenix.spark.SaltedTableIT  Time elapsed: 0.002 s  <<< 
> ERROR!
> java.lang.RuntimeException: java.io.IOException: Failed to save in any 
> storage directories while saving namespace.
>     at org.apache.phoenix.query.BaseTest.initMiniCluster(BaseTest.java:549)
>     at org.apache.phoenix.query.BaseTest.setUpTestCluster(BaseTest.java:449)
>     at 
> org.apache.phoenix.query.BaseTest.checkClusterInitialized(BaseTest.java:435)
>     at org.apache.phoenix.query.BaseTest.setUpTestDriver(BaseTest.java:517)
>     at org.apache.phoenix.query.BaseTest.setUpTestDriver(BaseTest.java:512)
>     at 
> org.apache.phoenix.end2end.ParallelStatsDisabledIT.doSetup(ParallelStatsDisabledIT.java:62)
>     at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>     at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>     at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>     at java.lang.reflect.Method.invoke(Method.java:498)
>     at 
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
>     at 
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
>     at 
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
>     at 
> org.junit.internal.runners.statements.RunBefores.invokeMethod(RunBefores.java:33)
>     at 
> org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:24)
>     at 
> org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
>     at org.apache.phoenix.SystemExitRule$1.evaluate(SystemExitRule.java:40)
>     at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:54)
>     at org.junit.rules.RunRules.evaluate(RunRules.java:20)
>     at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
>     at org.junit.runners.ParentRunner.run(ParentRunner.java:413)
>     at org.junit.runners.Suite.runChild(Suite.java:128)
>     at org.junit.runners.Suite.runChild(Suite.java:27)
>     at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331)
>     at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79)
>     at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329)
>     at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66)
>     at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293)
>     at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
>     at org.junit.runners.ParentRunner.run(ParentRunner.java:413)
>     at org.apache.maven.surefire.junitcore.JUnitCore.run(JUnitCore.java:49)
>     at 
> org.apache.maven.surefire.junitcore.JUnitCoreWrapper.createRequestAndRun(JUnitCoreWrapper.java:120)
>     at 
> org.apache.maven.surefire.junitcore.JUnitCoreWrapper.executeLazy(JUnitCoreWrapper.java:105)
>     at 
> org.apache.maven.surefire.junitcore.JUnitCoreWrapper.execute(JUnitCoreWrapper.java:77)
>     at 
> org.apache.maven.surefire.junitcore.JUnitCoreWrapper.execute(JUnitCoreWrapper.java:69)
>     at 
> org.apache.maven.surefire.junitcore.JUnitCoreProvider.invoke(JUnitCoreProvider.java:146)
>     at 
> org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:385)
>     at 
> org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:162)
>     at 
> org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:507)
>     at 
> org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:495)
> Caused by: java.io.IOException: Failed to save in any storage directories 
> while saving namespace.
>     at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.saveFSImageInAllDirs(FSImage.java:1192)
>     at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.saveFSImageInAllDirs(FSImage.java:1149)
>     at org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:175)
>     at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:1138)
>     at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:402)
>     at org.apache.hadoop.hdfs.DFSTestUtil.formatNameNode(DFSTestUtil.java:248)
>     at 
> org.apache.hadoop.hdfs.MiniDFSCluster.configureNameService(MiniDFSCluster.java:1063)
>     at 
> org.apache.hadoop.hdfs.MiniDFSCluster.createNameNodesAndSetConf(MiniDFSCluster.java:949)
>     at 
> org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:881)
>     at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:797)
>     at 
> org.apache.hadoop.hbase.HBaseTestingUtility.startMiniDFSCluster(HBaseTestingUtility.java:673)
>     at 
> org.apache.hadoop.hbase.HBaseTestingUtility.startMiniDFSCluster(HBaseTestingUtility.java:645)
>     at 
> org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:996)
>     at 
> org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:876)
>     at 
> org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:858)
>     at 
> org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:840)
>     at org.apache.phoenix.query.BaseTest.initMiniCluster(BaseTest.java:544)
>     ... 39 more
> {noformat}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

Reply via email to